config.name,config.backend.name,config.backend.version,config.backend._target_,config.backend.task,config.backend.library,config.backend.model_type,config.backend.model,config.backend.processor,config.backend.device,config.backend.device_ids,config.backend.seed,config.backend.inter_op_num_threads,config.backend.intra_op_num_threads,config.backend.model_kwargs.trust_remote_code,config.backend.no_weights,config.backend.device_map,config.backend.torch_dtype,config.backend.eval_mode,config.backend.to_bettertransformer,config.backend.low_cpu_mem_usage,config.backend.attn_implementation,config.backend.cache_implementation,config.backend.autocast_enabled,config.backend.autocast_dtype,config.backend.torch_compile,config.backend.torch_compile_target,config.backend.quantization_scheme,config.backend.quantization_config.bits,config.backend.quantization_config.version,config.backend.quantization_config.exllama_config.version,config.backend.quantization_config.exllama_config.max_input_len,config.backend.quantization_config.exllama_config.max_batch_size,config.backend.deepspeed_inference,config.backend.peft_type,config.scenario.name,config.scenario._target_,config.scenario.iterations,config.scenario.duration,config.scenario.warmup_runs,config.scenario.input_shapes.batch_size,config.scenario.input_shapes.num_choices,config.scenario.input_shapes.sequence_length,config.scenario.new_tokens,config.scenario.memory,config.scenario.latency,config.scenario.energy,config.scenario.generate_kwargs.max_new_tokens,config.scenario.generate_kwargs.min_new_tokens,config.launcher.name,config.launcher._target_,config.launcher.device_isolation,config.launcher.device_isolation_action,config.launcher.numactl,config.launcher.start_method,config.environment.cpu,config.environment.cpu_count,config.environment.cpu_ram_mb,config.environment.system,config.environment.machine,config.environment.platform,config.environment.processor,config.environment.python_version,config.environment.gpu,config.environment.gpu_count,config.environment.gpu_vram_mb,config.environment.optimum_benchmark_version,config.environment.optimum_benchmark_commit,config.environment.transformers_version,config.environment.transformers_commit,config.environment.accelerate_version,config.environment.accelerate_commit,config.environment.diffusers_version,config.environment.diffusers_commit,config.environment.optimum_version,config.environment.optimum_commit,config.environment.timm_version,config.environment.timm_commit,config.environment.peft_version,config.environment.peft_commit,report.traceback,report.load.memory.unit,report.load.memory.max_ram,report.load.memory.max_global_vram,report.load.memory.max_process_vram,report.load.memory.max_reserved,report.load.memory.max_allocated,report.load.latency.unit,report.load.latency.count,report.load.latency.total,report.load.latency.mean,report.load.latency.stdev,report.load.latency.p50,report.load.latency.p90,report.load.latency.p95,report.load.latency.p99,report.load.latency.values,report.load.throughput,report.load.energy.unit,report.load.energy.cpu,report.load.energy.ram,report.load.energy.gpu,report.load.energy.total,report.load.efficiency,report.prefill.memory.unit,report.prefill.memory.max_ram,report.prefill.memory.max_global_vram,report.prefill.memory.max_process_vram,report.prefill.memory.max_reserved,report.prefill.memory.max_allocated,report.prefill.latency.unit,report.prefill.latency.count,report.prefill.latency.total,report.prefill.latency.mean,report.prefill.latency.stdev,report.prefill.latency.p50,report.prefill.latency.p90,report.prefill.latency.p95,report.prefill.latency.p99,report.prefill.latency.values,report.prefill.throughput.unit,report.prefill.throughput.value,report.prefill.energy.unit,report.prefill.energy.cpu,report.prefill.energy.ram,report.prefill.energy.gpu,report.prefill.energy.total,report.prefill.efficiency.unit,report.prefill.efficiency.value,report.decode.memory.unit,report.decode.memory.max_ram,report.decode.memory.max_global_vram,report.decode.memory.max_process_vram,report.decode.memory.max_reserved,report.decode.memory.max_allocated,report.decode.latency.unit,report.decode.latency.count,report.decode.latency.total,report.decode.latency.mean,report.decode.latency.stdev,report.decode.latency.p50,report.decode.latency.p90,report.decode.latency.p95,report.decode.latency.p99,report.decode.latency.values,report.decode.throughput.unit,report.decode.throughput.value,report.decode.energy.unit,report.decode.energy.cpu,report.decode.energy.ram,report.decode.energy.gpu,report.decode.energy.total,report.decode.efficiency.unit,report.decode.efficiency.value,report.per_token.memory,report.per_token.latency.unit,report.per_token.latency.count,report.per_token.latency.total,report.per_token.latency.mean,report.per_token.latency.stdev,report.per_token.latency.p50,report.per_token.latency.p90,report.per_token.latency.p95,report.per_token.latency.p99,report.per_token.latency.values,report.per_token.throughput.unit,report.per_token.throughput.value,report.per_token.energy,report.per_token.efficiency 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4259.401728,5784.928256,0.0,5389.68064,5000.446464,s,1,11.729078125,11.729078125,0.0,11.729078125,11.729078125,11.729078125,11.729078125,[11.729078125],,kWh,0.00013139701470000394,1.4465109525743916e-05,5.4969766197981995e-05,0.00020083189042372987,,MB,1389.924352,5801.705472,0.0,5393.874944,4700.829696,s,10,29.1660693359375,2.9166069335937506,0.003379935020874598,2.9171240234375,2.919841650390625,2.9200343505859374,2.9201885107421877,"[2.908759033203125, 2.913849365234375, 2.914615966796875, 2.915674560546875, 2.916779296875, 2.91746875, 2.919466552734375, 2.919429931640625, 2.92022705078125, 2.919798828125]",tokens/s,87.7732261592634,kWh,8.504647543583359e-05,9.38051017194422e-06,5.6483989631600505e-05,0.00015091097523937833,tokens/kWh,1696364.3604709806,MB,1389.924352,5801.705472,0.0,5393.874944,4877.453824,s,10,16.09183557128906,1.6091835571289064,0.0016500610799483914,1.6091121826171875,1.6106735961914063,1.6117986511230469,1.6126986950683593,"[1.6086622314453125, 1.6089632568359375, 1.6129237060546875, 1.607747314453125, 1.60747509765625, 1.609567626953125, 1.6092611083984374, 1.606819091796875, 1.610423583984375, 1.6099925537109374]",tokens/s,39.150288182414776,kWh,4.717517559125327e-05,5.203773586761981e-06,3.1488108523800014e-05,8.386705770181517e-05,tokens/kWh,751188.8663602952,,s,630,16.088644878387438,0.025537531552995957,0.00032920167288641,0.025559120178222657,0.025860710906982422,0.025961028480529787,0.026728650798797608,"[0.026750688552856446, 0.026063552856445314, 0.025384960174560548, 0.02521897506713867, 0.025051231384277343, 0.025069120407104493, 0.025020864486694334, 0.02503696060180664, 0.025009695053100585, 0.025018688201904296, 0.02498150444030762, 0.025200639724731445, 0.025196544647216795, 0.025241600036621094, 0.025208127975463866, 0.025133760452270507, 0.02517196846008301, 0.025136768341064455, 0.025350048065185548, 0.025332191467285155, 0.025233407974243165, 0.025237472534179687, 0.025191680908203125, 0.0252607364654541, 0.025501792907714843, 0.025448448181152345, 0.02552422332763672, 0.025520128250122072, 0.025542240142822265, 0.025431999206542967, 0.025539039611816406, 0.025579519271850586, 0.025560192108154297, 0.02553945541381836, 0.025569280624389647, 0.025589344024658203, 0.025615999221801758, 0.025739904403686523, 0.02596236801147461, 0.02599679946899414, 0.025923583984375, 0.02587900733947754, 0.025742816925048827, 0.025584480285644532, 0.025551935195922852, 0.025559999465942382, 0.025622528076171876, 0.025818592071533204, 0.025608736038208006, 0.025594976425170897, 0.025612960815429686, 0.027001087188720702, 0.025448223114013672, 0.025607999801635743, 0.025720319747924804, 0.025756479263305664, 0.02578646469116211, 0.02568191909790039, 0.025589759826660157, 0.025675775527954102, 0.025831424713134765, 0.025833471298217774, 0.025769695281982422, 0.026617664337158203, 0.02584988784790039, 0.025318143844604492, 0.025222751617431642, 0.025077823638916016, 0.025186656951904297, 0.025114688873291015, 0.025151424407958985, 0.025049087524414062, 0.025008127212524413, 0.025470624923706053, 0.02510470390319824, 0.025241632461547852, 0.02521625518798828, 0.025117439270019533, 0.025083904266357423, 0.025275968551635743, 0.025254335403442383, 0.025081472396850588, 0.025160064697265627, 0.02521628761291504, 0.02538159942626953, 0.025438207626342774, 0.02543596839904785, 0.02537696075439453, 0.025425151824951173, 0.02531113624572754, 0.025471616744995117, 0.025583839416503905, 0.025645055770874024, 0.025558847427368164, 0.025518112182617188, 0.02555504035949707, 0.025444192886352537, 0.025569503784179687, 0.025605567932128905, 0.025657920837402343, 0.02575155258178711, 0.025961503982543946, 0.02603107261657715, 0.02575529670715332, 0.025649023056030274, 0.025737695693969727, 0.02571468734741211, 0.025640960693359374, 0.02569011116027832, 0.02571059226989746, 0.025665536880493164, 0.025606143951416017, 0.02549964714050293, 0.02555084800720215, 0.025612031936645508, 0.02569036865234375, 0.025825216293334962, 0.025744703292846678, 0.025735488891601564, 0.025817344665527344, 0.025830911636352538, 0.02575017547607422, 0.025959808349609374, 0.026086048126220705, 0.025961824417114258, 0.02584796714782715, 0.026719039916992187, 0.025845760345458983, 0.025427936553955078, 0.025190143585205077, 0.02499203109741211, 0.025038719177246094, 0.025045120239257812, 0.025034751892089844, 0.024982784271240236, 0.02506604766845703, 0.02518239974975586, 0.025184064865112304, 0.025589792251586915, 0.02745903968811035, 0.02511235237121582, 0.02508073616027832, 0.025116479873657227, 0.02537081527709961, 0.02533785629272461, 0.02532761573791504, 0.025270240783691406, 0.02549260711669922, 0.02550876808166504, 0.025413631439208984, 0.025479167938232423, 0.02553446388244629, 0.02552934455871582, 0.025481727600097655, 0.025459199905395507, 0.025444032669067383, 0.025749824523925782, 0.025389055252075195, 0.025472223281860353, 0.025522464752197264, 0.02561235237121582, 0.02563443183898926, 0.025757919311523436, 0.025756256103515625, 0.02585385513305664, 0.025996608734130858, 0.025958976745605468, 0.026241119384765626, 0.025780351638793945, 0.02577401542663574, 0.025626688003540038, 0.025585664749145507, 0.02572697639465332, 0.02572287940979004, 0.02572492790222168, 0.02573833656311035, 0.025840192794799804, 0.025769567489624022, 0.02571955108642578, 0.025734848022460937, 0.025710527420043944, 0.02563929557800293, 0.02568806457519531, 0.025886335372924806, 0.025804767608642577, 0.025786783218383787, 0.025823232650756835, 0.025911296844482422, 0.025960447311401368, 0.02671174430847168, 0.025864511489868163, 0.025480703353881837, 0.025338367462158205, 0.02520841598510742, 0.025155168533325195, 0.025155839920043947, 0.02505945587158203, 0.02508559989929199, 0.02506831932067871, 0.025124864578247072, 0.025091680526733398, 0.025083904266357423, 0.025239967346191407, 0.025245439529418944, 0.025166080474853515, 0.025210655212402344, 0.02534217643737793, 0.025370624542236327, 0.025341951370239257, 0.025403263092041016, 0.025315296173095702, 0.025323680877685547, 0.025343904495239256, 0.02537059211730957, 0.02556105613708496, 0.025718463897705077, 0.025610719680786133, 0.02550783920288086, 0.025444351196289062, 0.025464351654052735, 0.025424352645874025, 0.025513055801391602, 0.025617311477661133, 0.02557481575012207, 0.02543881607055664, 0.025641984939575195, 0.025684991836547853, 0.025676959991455077, 0.025666400909423827, 0.025806272506713867, 0.025827903747558594, 0.025831071853637696, 0.025866592407226562, 0.025841663360595703, 0.02575155258178711, 0.025655296325683592, 0.025647104263305662, 0.025618431091308593, 0.025487360000610353, 0.025449695587158202, 0.025555391311645508, 0.02555939292907715, 0.02561414337158203, 0.025692352294921873, 0.02569808006286621, 0.02568623924255371, 0.02555084800720215, 0.025630016326904297, 0.025625280380249024, 0.02572492790222168, 0.025773887634277345, 0.025905344009399416, 0.026732576370239257, 0.025757183074951173, 0.025468704223632812, 0.025313983917236327, 0.02514739227294922, 0.02501968002319336, 0.02504960060119629, 0.02512816047668457, 0.02519945526123047, 0.025151647567749024, 0.025133056640625, 0.02525814437866211, 0.025229120254516603, 0.02523753547668457, 0.025243648529052733, 0.02512892723083496, 0.025277631759643555, 0.02519331169128418, 0.025362432479858397, 0.025372671127319335, 0.025394752502441408, 0.025407360076904296, 0.025366432189941408, 0.025385631561279296, 0.02543404769897461, 0.025354303359985352, 0.025443679809570314, 0.025495904922485352, 0.025512256622314454, 0.025425920486450194, 0.025407487869262696, 0.02541904067993164, 0.025461471557617188, 0.025386592864990235, 0.02536899185180664, 0.02554265594482422, 0.02561177635192871, 0.025614048004150392, 0.025623327255249025, 0.02570595169067383, 0.02577187156677246, 0.025785024642944337, 0.025802751541137696, 0.025827327728271485, 0.025599807739257813, 0.025501184463500977, 0.025568960189819336, 0.02569113540649414, 0.025761472702026368, 0.02614918327331543, 0.02575542449951172, 0.02568828773498535, 0.025671680450439452, 0.025659391403198242, 0.025612287521362305, 0.025595903396606445, 0.025624576568603515, 0.02572083282470703, 0.02573334312438965, 0.02574006462097168, 0.02574569511413574, 0.02569196891784668, 0.02571356773376465, 0.02693596839904785, 0.02584726333618164, 0.02541638374328613, 0.025191648483276367, 0.02514998435974121, 0.025098623275756834, 0.025034208297729493, 0.025016864776611327, 0.025077632904052734, 0.025194623947143554, 0.025116672515869142, 0.02511788749694824, 0.025137983322143554, 0.025294527053833008, 0.025243967056274415, 0.02517196846008301, 0.02532761573791504, 0.02533955192565918, 0.025343360900878905, 0.025406400680541993, 0.025483295440673827, 0.02546832084655762, 0.025375200271606446, 0.025327392578125, 0.02534163284301758, 0.025402015686035156, 0.025397184371948243, 0.025710655212402345, 0.025689439773559572, 0.02552284812927246, 0.025393056869506835, 0.025451871871948244, 0.02547727966308594, 0.02546339225769043, 0.02546073532104492, 0.025750688552856445, 0.02573967933654785, 0.02572127914428711, 0.025972736358642577, 0.025851903915405275, 0.02579257583618164, 0.025564832687377928, 0.025575391769409178, 0.025589439392089845, 0.025655935287475586, 0.025728416442871094, 0.025723007202148436, 0.025719263076782226, 0.025637983322143554, 0.025651872634887694, 0.025690048217773438, 0.025641088485717774, 0.02562393569946289, 0.025916000366210938, 0.025939584732055664, 0.026108512878417967, 0.02590924835205078, 0.025763391494750976, 0.025778528213500976, 0.025677919387817383, 0.0256407356262207, 0.025682144165039063, 0.025773855209350587, 0.02698716735839844, 0.025975008010864258, 0.025462560653686524, 0.025212928771972655, 0.025130815505981445, 0.025069536209106444, 0.02502668762207031, 0.02511676788330078, 0.02516927909851074, 0.025154048919677735, 0.025067327499389648, 0.025049087524414062, 0.02508012771606445, 0.02509516716003418, 0.02500819206237793, 0.025099199295043947, 0.025243648529052733, 0.025432064056396485, 0.02547439956665039, 0.025442975997924805, 0.025347583770751952, 0.025333824157714843, 0.025229759216308593, 0.025274368286132814, 0.02537676811218262, 0.02536857604980469, 0.0253439998626709, 0.025491104125976563, 0.025590112686157226, 0.025573375701904297, 0.025475072860717773, 0.0255467529296875, 0.025587711334228515, 0.02555084800720215, 0.025571136474609374, 0.02558790397644043, 0.025646207809448242, 0.025705343246459962, 0.025860288619995116, 0.025843488693237306, 0.025780256271362306, 0.02571059226989746, 0.02557535934448242, 0.02556710433959961, 0.02558332824707031, 0.025786848068237306, 0.02591744041442871, 0.025832895278930665, 0.025735744476318358, 0.02573107147216797, 0.025726112365722656, 0.025731935501098632, 0.025651199340820312, 0.0256975040435791, 0.025768415451049805, 0.02571295928955078, 0.025735008239746095, 0.02581315231323242, 0.025810304641723632, 0.02582796859741211, 0.026036224365234374, 0.025896575927734374, 0.02573686408996582, 0.026806079864501953, 0.025878271102905272, 0.02539155197143555, 0.025081920623779296, 0.025094079971313476, 0.025012224197387696, 0.02501356887817383, 0.025004512786865236, 0.025091999053955077, 0.02506755256652832, 0.025137439727783203, 0.024979455947875977, 0.025241119384765625, 0.02605923271179199, 0.02504425621032715, 0.025035327911376953, 0.025249759674072267, 0.025237312316894533, 0.025179744720458985, 0.025215776443481445, 0.025247488021850586, 0.02529100799560547, 0.025212928771972655, 0.025273504257202147, 0.025449312210083008, 0.02545254325866699, 0.025392543792724608, 0.025370527267456054, 0.025365184783935547, 0.025372671127319335, 0.02550783920288086, 0.025643007278442383, 0.025595903396606445, 0.02548476791381836, 0.025351776123046874, 0.02561324882507324, 0.025718751907348632, 0.02563484764099121, 0.02561756706237793, 0.025735904693603515, 0.02572096061706543, 0.025572704315185546, 0.02553910446166992, 0.02555507278442383, 0.02557868766784668, 0.025532480239868163, 0.02555366325378418, 0.025703487396240236, 0.02564806365966797, 0.025655040740966795, 0.025644607543945312, 0.0256375675201416, 0.025552896499633788, 0.025685279846191407, 0.025729759216308594, 0.025788415908813478, 0.025894336700439453, 0.025927488327026366, 0.025920255661010742, 0.02588057518005371, 0.025876415252685546, 0.02581100845336914, 0.02590460777282715, 0.026578208923339845, 0.025716800689697266, 0.025268735885620116, 0.025078144073486328, 0.025097888946533205, 0.025051519393920897, 0.02499305534362793, 0.025010623931884766, 0.02500761604309082, 0.025008895874023437, 0.02494879913330078, 0.0251473274230957, 0.025264127731323242, 0.025266176223754884, 0.025134368896484374, 0.025184864044189452, 0.025229440689086915, 0.025264127731323242, 0.025263200759887694, 0.025332639694213867, 0.025365983963012696, 0.025754079818725586, 0.025368640899658203, 0.025409536361694338, 0.025439264297485352, 0.025485696792602538, 0.025391199111938476, 0.025435871124267578, 0.025455232620239257, 0.02535785675048828, 0.025243295669555663, 0.025445247650146486, 0.025606239318847656, 0.02560985565185547, 0.025705984115600586, 0.0257030086517334, 0.02585219192504883, 0.02592767906188965, 0.025944063186645508, 0.02590105628967285, 0.025915391921997072, 0.02583318328857422, 0.025874271392822265, 0.025798559188842773, 0.025690271377563478, 0.025678207397460937, 0.02566044807434082, 0.025647775650024414, 0.02560767936706543, 0.025929759979248047, 0.02594691276550293, 0.025795583724975587, 0.025758527755737306, 0.02583683204650879, 0.025856927871704103, 0.025843711853027345, 0.02583510398864746, 0.02576630401611328, 0.025774080276489256, 0.025773567199707033, 0.02603670310974121, 0.02605673599243164, 0.02587843132019043, 0.02668339157104492, 0.025769535064697265, 0.025364927291870117, 0.02520659255981445, 0.025089887619018553, 0.025038656234741212, 0.02510406494140625, 0.02523017692565918, 0.025217023849487305, 0.025135103225708007, 0.025216960906982423, 0.02521238327026367, 0.025148000717163086, 0.025162912368774413, 0.025148256301879883, 0.02530508804321289, 0.025237407684326172, 0.025161823272705077, 0.025212608337402343, 0.025214431762695312, 0.025270303726196288, 0.02534662437438965, 0.025311487197875977, 0.025464832305908205, 0.02615910339355469, 0.02532761573791504, 0.025392223358154296, 0.026376672744750976, 0.025403839111328125, 0.02534121513366699, 0.025285343170166015, 0.025444351196289062, 0.025579200744628907, 0.02557779121398926, 0.025577472686767577, 0.025535871505737304, 0.025557056427001953, 0.025817663192749023, 0.02598297691345215, 0.025948160171508788, 0.02581830406188965, 0.025801536560058593, 0.02584899139404297, 0.025705312728881834, 0.02576755142211914, 0.025767616271972656, 0.025683807373046874, 0.025717599868774414, 0.025838943481445314, 0.025702207565307618, 0.02566435241699219, 0.025761791229248047, 0.025768991470336913, 0.025717599868774414, 0.0256943359375, 0.02575526428222656, 0.025768320083618165, 0.02575926399230957, 0.025651679992675782, 0.025634815216064453, 0.02571059226989746, 0.025792032241821288, 0.025803232192993165]",tokens/s,39.15805245016659,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1302.376448,1127.153664,0.0,731.906048,703.86944,s,1,8.321544921875,8.321544921875,0.0,8.321544921875,8.321544921875,8.321544921875,8.321544921875,[8.321544921875],,kWh,3.9562046445826126e-05,4.354408836396406e-06,1.2616121204039876e-05,5.653257648626241e-05,,MB,1333.53472,1406.07488,0.0,998.244352,942.608384,s,10,1.3574564971923828,0.13574564971923828,0.0006022325849746154,0.1357723388671875,0.13615484466552735,0.13659751815795898,0.1369516569519043,"[0.13704019165039064, 0.1360564727783203, 0.13602336120605468, 0.13559088134765626, 0.13500009155273437, 0.1360320281982422, 0.13476524353027344, 0.13590611267089844, 0.13540354919433595, 0.13563856506347657]",tokens/s,1885.8799565914849,kWh,4.1684940085098125e-06,4.5957960022677493e-07,2.7506633898877353e-06,7.3787369986243225e-06,tokens/kWh,34694284.407714784,MB,1351.610368,1414.463488,0.0,1006.63296,942.610944,s,10,12.693910522460937,1.2693910522460938,0.00947761483403399,1.2684224853515624,1.2760179931640625,1.283985498046875,1.290359501953125,"[1.2726170654296876, 1.2716322021484374, 1.2919530029296875, 1.2742159423828125, 1.2652127685546875, 1.263024658203125, 1.2567308349609374, 1.2608265380859376, 1.2634500732421876, 1.2742474365234375]",tokens/s,49.6300961697549,kWh,3.7052500298990594e-05,4.086647627488998e-06,1.592885155371036e-05,5.706799948018996e-05,tokens/kWh,1103946.1795374344,,s,630,12.691761045455934,0.020145652453104658,0.00039547894563328847,0.020049711227416993,0.02049824275970459,0.020685639858245848,0.021514591522216804,"[0.021014944076538086, 0.02051215934753418, 0.02039033508300781, 0.020323808670043946, 0.020648063659667967, 0.020427360534667968, 0.02026851272583008, 0.020166208267211914, 0.020063199996948243, 0.020060159683227538, 0.019875648498535157, 0.01978816032409668, 0.019892032623291016, 0.019791040420532226, 0.01980499267578125, 0.0199169921875, 0.019912511825561523, 0.019830272674560546, 0.02010982322692871, 0.019912704467773438, 0.020035232543945312, 0.019884384155273438, 0.019961311340332032, 0.019886016845703125, 0.020058719635009766, 0.020555776596069338, 0.020473823547363282, 0.020477983474731447, 0.020447231292724608, 0.02038969612121582, 0.020399328231811523, 0.02032089614868164, 0.020771167755126954, 0.02040003204345703, 0.02041152000427246, 0.020441823959350586, 0.02041472053527832, 0.020454431533813478, 0.020433887481689453, 0.02023139190673828, 0.020179744720458984, 0.020033536911010744, 0.020154367446899413, 0.02020102310180664, 0.020150592803955078, 0.020244319915771483, 0.020246816635131837, 0.020213760375976563, 0.020036960601806642, 0.02012550354003906, 0.020101280212402345, 0.019978336334228516, 0.020092927932739257, 0.020095584869384765, 0.020354175567626955, 0.020146720886230467, 0.020072799682617187, 0.020064159393310545, 0.019963552474975586, 0.020052288055419924, 0.02015363121032715, 0.020281183242797853, 0.020307167053222656, 0.021242591857910158, 0.020348928451538087, 0.020164384841918945, 0.020166879653930665, 0.020033023834228517, 0.02022585678100586, 0.020150976181030275, 0.02026038360595703, 0.02031158447265625, 0.019882944107055663, 0.01992684745788574, 0.019947711944580077, 0.0198922233581543, 0.019946592330932617, 0.019897247314453127, 0.019863168716430665, 0.01991516876220703, 0.019944799423217772, 0.02006825637817383, 0.01995030403137207, 0.020011199951171874, 0.019848352432250978, 0.019966623306274415, 0.019793920516967774, 0.01997327995300293, 0.019870559692382814, 0.020010719299316405, 0.020119840621948243, 0.020022623062133788, 0.02010998344421387, 0.020020544052124025, 0.020069055557250977, 0.020125215530395507, 0.02005971145629883, 0.0206529598236084, 0.02009449577331543, 0.02041644859313965, 0.02023232078552246, 0.019976608276367186, 0.01991641616821289, 0.01999679946899414, 0.019942848205566407, 0.020005056381225586, 0.020006816864013673, 0.02012656021118164, 0.02005561637878418, 0.02014761543273926, 0.020411136627197266, 0.020518688201904296, 0.020289535522460937, 0.02019571113586426, 0.020487808227539064, 0.020319807052612306, 0.020322687149047853, 0.020785600662231445, 0.02049827194213867, 0.020321727752685547, 0.020389631271362306, 0.020605503082275392, 0.020679071426391603, 0.020353023529052734, 0.02083635139465332, 0.02066966438293457, 0.02101068878173828, 0.020642879486083985, 0.020489152908325196, 0.020442848205566407, 0.020418079376220703, 0.020417280197143554, 0.02042812728881836, 0.020456096649169923, 0.02060054397583008, 0.020635520935058594, 0.020539552688598632, 0.020436511993408204, 0.020533279418945314, 0.02081407928466797, 0.02062585639953613, 0.024229280471801756, 0.020761184692382813, 0.020666048049926757, 0.020508703231811524, 0.020731647491455077, 0.020415008544921873, 0.02042470359802246, 0.020653120040893556, 0.020720352172851564, 0.02045155143737793, 0.02015350341796875, 0.020220767974853514, 0.020340736389160157, 0.020358335494995116, 0.020455583572387696, 0.020626176834106447, 0.02057206344604492, 0.02039193534851074, 0.02047385597229004, 0.02043903923034668, 0.020427999496459962, 0.020426816940307617, 0.02035785675048828, 0.020129791259765627, 0.020760223388671874, 0.020304351806640624, 0.020166048049926756, 0.020134368896484376, 0.020170751571655272, 0.02013100814819336, 0.020007743835449218, 0.020063776016235352, 0.01998076820373535, 0.02003763198852539, 0.020067840576171874, 0.020617055892944335, 0.02003830337524414, 0.019979583740234376, 0.02005471992492676, 0.020125696182250977, 0.020140031814575195, 0.020178943634033202, 0.021573055267333986, 0.021371456146240236, 0.020505887985229492, 0.020249120712280272, 0.02083420753479004, 0.02080963134765625, 0.02111497688293457, 0.020498239517211914, 0.020428991317749022, 0.020416383743286134, 0.020392223358154295, 0.020440927505493166, 0.02031718444824219, 0.020271968841552735, 0.020690176010131837, 0.02251254463195801, 0.0206167049407959, 0.020451135635375976, 0.02045529556274414, 0.02024892807006836, 0.02038217544555664, 0.020469087600708008, 0.020343008041381835, 0.020382144927978515, 0.02023423957824707, 0.021823232650756835, 0.02080998420715332, 0.020315616607666016, 0.02026905632019043, 0.020255327224731445, 0.02026412773132324, 0.020177663803100584, 0.020266624450683595, 0.020178815841674805, 0.020220415115356445, 0.020123392105102538, 0.020086496353149415, 0.019902111053466797, 0.020061023712158205, 0.020018655776977538, 0.02008323287963867, 0.019963136672973635, 0.019950368881225585, 0.020059200286865236, 0.019687871932983398, 0.019892736434936522, 0.019844192504882813, 0.019958208084106446, 0.020074975967407226, 0.019951168060302733, 0.020090911865234377, 0.019987903594970702, 0.019864543914794922, 0.020000127792358397, 0.019845184326171876, 0.019939903259277345, 0.019893503189086913, 0.01988275146484375, 0.019935232162475586, 0.019842239379882814, 0.019954496383666993, 0.01985273551940918, 0.019718719482421876, 0.019900415420532228, 0.020129119873046875, 0.02003420829772949, 0.02006220817565918, 0.02003558349609375, 0.02011942481994629, 0.02127257537841797, 0.02061123275756836, 0.02033833694458008, 0.020100704193115236, 0.02012015914916992, 0.020002815246582033, 0.019899423599243165, 0.01990297508239746, 0.020083168029785155, 0.020129791259765627, 0.020106592178344727, 0.01988060760498047, 0.019961151123046875, 0.019860160827636718, 0.020074527740478517, 0.019941183090209962, 0.01992207908630371, 0.019876928329467775, 0.019892448425292968, 0.01979363250732422, 0.019850624084472655, 0.0197761287689209, 0.02004991912841797, 0.020048927307128907, 0.020403167724609376, 0.020418560028076172, 0.0203656005859375, 0.02012950325012207, 0.02012371253967285, 0.020059551239013672, 0.020032032012939453, 0.0200949764251709, 0.020059871673583984, 0.019910367965698242, 0.019864128112792968, 0.020107263565063475, 0.01989017677307129, 0.019903968811035157, 0.019936960220336915, 0.01990083122253418, 0.019906400680541992, 0.019847776412963865, 0.019861503601074217, 0.019959808349609375, 0.020002687454223633, 0.02030780792236328, 0.020228384017944336, 0.02018636894226074, 0.02017910385131836, 0.020447839736938478, 0.02070528030395508, 0.020379552841186522, 0.020195423126220705, 0.020129472732543945, 0.020166976928710938, 0.02005526351928711, 0.019952064514160157, 0.020201568603515626, 0.019834463119506835, 0.019863487243652344, 0.01993187141418457, 0.019951456069946288, 0.02000726318359375, 0.02103910446166992, 0.02066022491455078, 0.020802879333496095, 0.020128448486328124, 0.020199424743652345, 0.02007244873046875, 0.02051683235168457, 0.02019945526123047, 0.020395360946655273, 0.020388511657714842, 0.020428064346313477, 0.02040115165710449, 0.020199071884155272, 0.01990662384033203, 0.019912704467773438, 0.01977894401550293, 0.019823232650756837, 0.019842048645019532, 0.01975359916687012, 0.01975859260559082, 0.01985215950012207, 0.019763200759887696, 0.019834880828857423, 0.01993907165527344, 0.01984342384338379, 0.020145856857299804, 0.019850847244262695, 0.0199051513671875, 0.019906143188476562, 0.01992336082458496, 0.019955711364746095, 0.019800064086914062, 0.01998847961425781, 0.01995088005065918, 0.019825376510620118, 0.02006220817565918, 0.019908607482910155, 0.01990870475769043, 0.019762880325317384, 0.01986172866821289, 0.019877887725830077, 0.019942848205566407, 0.019839551925659178, 0.019938432693481445, 0.020003711700439453, 0.019984224319458007, 0.01999068832397461, 0.02007859230041504, 0.019954719543457032, 0.019956703186035155, 0.02002118492126465, 0.020000831604003906, 0.020446367263793945, 0.020009824752807617, 0.020150144577026366, 0.02010451126098633, 0.020036287307739258, 0.020115583419799805, 0.01996985626220703, 0.019994367599487306, 0.02018092727661133, 0.020010784149169923, 0.020023040771484375, 0.020304832458496094, 0.02017001533508301, 0.020011104583740235, 0.0199420166015625, 0.020060159683227538, 0.0199968318939209, 0.019956928253173828, 0.019904928207397463, 0.019978496551513673, 0.020092863082885742, 0.02043280029296875, 0.019975711822509765, 0.01999078369140625, 0.01976563262939453, 0.01986355209350586, 0.019881759643554688, 0.019908767700195312, 0.01981760025024414, 0.01986886405944824, 0.019859136581420897, 0.019873184204101564, 0.019788255691528322, 0.019800256729125977, 0.01976937675476074, 0.019691488265991212, 0.019797344207763672, 0.019794208526611328, 0.019758623123168947, 0.01980624008178711, 0.019929920196533203, 0.019979616165161133, 0.020128416061401366, 0.019957759857177734, 0.020028703689575194, 0.019944160461425782, 0.019998720169067383, 0.019978239059448243, 0.019902463912963866, 0.019882144927978514, 0.019969984054565428, 0.019926015853881835, 0.01982352066040039, 0.01983897590637207, 0.019853311538696287, 0.019941375732421874, 0.019874975204467772, 0.019917631149291994, 0.020088863372802735, 0.019947519302368166, 0.020028928756713867, 0.02004150390625, 0.019964639663696288, 0.019981920242309572, 0.019962272644042968, 0.019797792434692384, 0.0200930233001709, 0.01991641616821289, 0.019908735275268555, 0.019888288497924806, 0.019929311752319337, 0.020080608367919923, 0.020156448364257812, 0.02000486373901367, 0.020390207290649415, 0.02018070411682129, 0.020049503326416016, 0.02004038429260254, 0.019947519302368166, 0.020090879440307616, 0.020098623275756837, 0.01983296012878418, 0.019965696334838866, 0.019865888595581055, 0.019966239929199218, 0.020395519256591797, 0.019920799255371095, 0.020116064071655275, 0.019838144302368164, 0.019913536071777344, 0.01993427276611328, 0.01989913558959961, 0.019926816940307616, 0.01988240051269531, 0.019970048904418947, 0.019797056198120118, 0.019921855926513674, 0.019961856842041017, 0.02007676887512207, 0.019861183166503905, 0.019953760147094726, 0.01993654441833496, 0.01993187141418457, 0.01989017677307129, 0.019968000411987305, 0.019766624450683595, 0.019983007431030275, 0.019936479568481446, 0.01988185691833496, 0.02004470443725586, 0.02004377555847168, 0.020133535385131837, 0.019804288864135742, 0.019953887939453126, 0.02068009567260742, 0.020265567779541017, 0.02010259246826172, 0.020019168853759765, 0.020053951263427735, 0.0199354248046875, 0.020263328552246093, 0.02011974334716797, 0.02008460807800293, 0.020125696182250977, 0.01993471908569336, 0.020007423400878906, 0.020016639709472657, 0.02034239959716797, 0.020102016448974608, 0.019910655975341796, 0.019868896484375, 0.019968799591064453, 0.019937280654907227, 0.020024511337280275, 0.019938112258911133, 0.01980620765686035, 0.020019039154052735, 0.020642911911010742, 0.020136255264282227, 0.020114015579223633, 0.019978239059448243, 0.019929088592529298, 0.02084774398803711, 0.023999359130859373, 0.020039680480957032, 0.020015104293823242, 0.019963903427124022, 0.01984105682373047, 0.01986147117614746, 0.019957056045532228, 0.019749568939208983, 0.01989836883544922, 0.019904512405395508, 0.019826688766479493, 0.019964128494262694, 0.019961055755615235, 0.019825151443481445, 0.019869760513305666, 0.02001919937133789, 0.019916767120361327, 0.0198492488861084, 0.019947519302368166, 0.019861503601074217, 0.019862592697143553, 0.019815359115600586, 0.01985868835449219, 0.019946239471435548, 0.019740575790405272, 0.019882080078125, 0.019843072891235353, 0.0198922233581543, 0.01998847961425781, 0.020019359588623047, 0.01990991973876953, 0.020671039581298827, 0.019895776748657227, 0.019843519210815428, 0.019949535369873046, 0.01987187194824219, 0.01982784080505371, 0.020158527374267578, 0.019861503601074217, 0.020082592010498047, 0.02010767936706543, 0.02006675148010254, 0.01992198371887207, 0.020072736740112306, 0.01993187141418457, 0.01997209548950195, 0.02001817512512207, 0.020224767684936522, 0.019996927261352538, 0.020077951431274416, 0.019937919616699218, 0.01997545623779297, 0.020107999801635742, 0.019998720169067383, 0.01985536003112793, 0.019965919494628906, 0.020161727905273437, 0.02130534362792969, 0.020923551559448243, 0.02159702491760254, 0.020706687927246094, 0.02038438415527344, 0.020373504638671876, 0.020336416244506834, 0.02030204772949219, 0.02012892723083496, 0.020056095123291016, 0.01995859146118164, 0.020099071502685546, 0.02013987159729004, 0.02011494445800781, 0.020169023513793946, 0.020094655990600587, 0.02009766387939453, 0.020081888198852538, 0.020083520889282228, 0.02009280014038086, 0.020236192703247072, 0.020003040313720702, 0.020606271743774413, 0.02029142379760742, 0.020251487731933592, 0.02297654342651367, 0.020439008712768554, 0.02048409652709961, 0.020469343185424805, 0.020453792572021484, 0.020590591430664062, 0.02013987159729004, 0.020205728530883788, 0.020092927932739257, 0.02014364814758301, 0.020128223419189455, 0.02006220817565918, 0.0200086727142334, 0.019810592651367188, 0.019974143981933593, 0.019896480560302736, 0.019863391876220705, 0.020022560119628906, 0.020064159393310545, 0.019976415634155274, 0.020083295822143556, 0.019994623184204103, 0.020023296356201172, 0.020023296356201172, 0.019983903884887695, 0.020220319747924806, 0.020105279922485352, 0.020139968872070313, 0.02004083251953125, 0.01995871925354004, 0.0198922233581543, 0.01992252731323242, 0.01993974494934082, 0.019848640441894532, 0.01986761665344238, 0.020012767791748046, 0.01981939125061035, 0.019933183670043944]",tokens/s,49.63850152422785,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTJForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3798, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ self.model = DeciCoderModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ self.self_attn = DeciCoderAttention(config=config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ self._init_rope() File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1709, in __getattr__ raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2188.980224,2907.635712,0.0,2512.388096,2240.694784,s,1,9.7695615234375,9.7695615234375,0.0,9.7695615234375,9.7695615234375,9.7695615234375,9.7695615234375,[9.7695615234375],,kWh,7.267370287916795e-05,8.008815164698533e-06,2.8941412041988412e-05,0.0001096239300858549,,MB,2220.744704,2922.315776,0.0,2514.485248,2226.413568,s,10,11.320579589843748,1.1320579589843747,0.00220818888242787,1.131412109375,1.135014697265625,1.1358196655273438,1.1364636401367187,"[1.13111181640625, 1.1293829345703126, 1.13000927734375, 1.129998291015625, 1.13131396484375, 1.132047607421875, 1.1337449951171874, 1.13151025390625, 1.1366246337890624, 1.1348358154296876]",tokens/s,226.1368315714774,kWh,3.3219415524163195e-05,3.66364249167958e-06,2.218648997140138e-05,5.906954798724415e-05,tokens/kWh,4333874.3688250715,MB,2223.202304,2922.315776,0.0,2514.485248,2337.090048,s,10,14.727046752929688,1.472704675292969,0.01827515411526005,1.4760736694335939,1.4904201782226563,1.496946661376953,1.5021678479003906,"[1.50347314453125, 1.4839168701171874, 1.4779920654296874, 1.488410888671875, 1.4889698486328125, 1.4606937255859376, 1.4510174560546876, 1.4741552734375, 1.4511300048828124, 1.4472874755859375]",tokens/s,42.77843416737117,kWh,4.2461431748752146e-05,4.683793481167955e-06,2.5790492854596556e-05,7.293571808451665e-05,tokens/kWh,863774.3159942114,,s,630,14.723933340072621,0.023371322762020048,0.0005964451738736854,0.023385295867919924,0.023886754608154297,0.024053939056396485,0.02478540544509888,"[0.024895488739013674, 0.02428108787536621, 0.024162208557128906, 0.02412566375732422, 0.02428505516052246, 0.02400796890258789, 0.02394316864013672, 0.02396342468261719, 0.02384748840332031, 0.023877727508544923, 0.023895711898803712, 0.024006847381591798, 0.024020736694335937, 0.02381279945373535, 0.02369740867614746, 0.023749664306640626, 0.02341142463684082, 0.02344576072692871, 0.023819679260253905, 0.023547584533691407, 0.023505823135375976, 0.02391244888305664, 0.02404351997375488, 0.023994367599487306, 0.02394281578063965, 0.02386569595336914, 0.023822336196899413, 0.023814144134521483, 0.023744512557983398, 0.023799808502197265, 0.023793664932250977, 0.02387353515625, 0.02390220832824707, 0.023851007461547852, 0.023953407287597657, 0.02406982421875, 0.02416774368286133, 0.023966720581054687, 0.023871360778808595, 0.023928960800170897, 0.023993535995483397, 0.0239005126953125, 0.02383830451965332, 0.023816991806030273, 0.023870784759521483, 0.023820703506469726, 0.023885759353637695, 0.02446156883239746, 0.02357676887512207, 0.023729183197021483, 0.023614431381225588, 0.0235696964263916, 0.023460575103759766, 0.023627775192260742, 0.023582719802856447, 0.023625728607177734, 0.023752639770507813, 0.023735712051391602, 0.02366326332092285, 0.023674400329589843, 0.023770784378051756, 0.023790399551391603, 0.02373222351074219, 0.02451420783996582, 0.024062463760375977, 0.02376246452331543, 0.023933248519897463, 0.023619583129882812, 0.02367830467224121, 0.023575199127197265, 0.023662271499633788, 0.023514944076538084, 0.02358527946472168, 0.023488512039184572, 0.023840192794799805, 0.02371027183532715, 0.023816192626953125, 0.0237172794342041, 0.023749216079711914, 0.023424543380737305, 0.02331286430358887, 0.023435264587402343, 0.023625728607177734, 0.023727167129516603, 0.023395263671875, 0.023448671340942383, 0.023449920654296876, 0.02324540710449219, 0.023433216094970705, 0.02329315185546875, 0.02328451156616211, 0.02309939193725586, 0.022988800048828126, 0.02294131278991699, 0.022978879928588866, 0.02282604789733887, 0.022969343185424804, 0.02285276794433594, 0.022995807647705077, 0.022923263549804687, 0.02289459228515625, 0.022841344833374022, 0.02304614448547363, 0.023334144592285156, 0.023698144912719727, 0.023719871520996094, 0.02369955253601074, 0.023662591934204103, 0.02369126319885254, 0.023618783950805664, 0.023712543487548827, 0.023649696350097657, 0.02380041694641113, 0.02372515106201172, 0.02380064010620117, 0.023760128021240234, 0.02386147117614746, 0.02380044746398926, 0.024550975799560545, 0.024170944213867188, 0.02375657653808594, 0.023644384384155274, 0.02377471923828125, 0.023900672912597655, 0.023805824279785157, 0.02377452850341797, 0.02434809684753418, 0.023167552947998046, 0.02290265655517578, 0.02283123207092285, 0.022769632339477538, 0.022892000198364258, 0.022755231857299805, 0.025170591354370116, 0.022978080749511718, 0.02284351921081543, 0.022686048507690428, 0.0227061767578125, 0.02264473533630371, 0.022683647155761717, 0.022648448944091796, 0.02259744071960449, 0.022615840911865234, 0.0226309757232666, 0.022636768341064453, 0.02272051239013672, 0.022700031280517577, 0.022923263549804687, 0.02319273567199707, 0.02330300712585449, 0.023971424102783204, 0.023386112213134767, 0.023476640701293947, 0.023481632232666017, 0.023667072296142577, 0.023775583267211915, 0.0240392951965332, 0.023793792724609374, 0.02370355224609375, 0.023810047149658203, 0.02372403144836426, 0.02384486389160156, 0.02382972717285156, 0.023990528106689453, 0.02390275192260742, 0.02407356834411621, 0.023795743942260743, 0.024021631240844728, 0.023954559326171875, 0.023962495803833007, 0.023861183166503906, 0.023791168212890627, 0.02357913589477539, 0.023570432662963867, 0.023564287185668945, 0.023549951553344727, 0.023489824295043947, 0.02358255958557129, 0.02342710494995117, 0.023516000747680663, 0.023488191604614257, 0.023400768280029297, 0.02328972816467285, 0.023367807388305663, 0.023262367248535157, 0.023530336380004884, 0.026771039962768556, 0.0236200008392334, 0.023451648712158202, 0.02421299171447754, 0.02359129524230957, 0.023473695755004884, 0.023906911849975586, 0.023584768295288085, 0.023644159317016602, 0.02342911911010742, 0.02364825630187988, 0.02343231964111328, 0.02361612892150879, 0.023560447692871092, 0.0239554557800293, 0.02353936004638672, 0.023603551864624022, 0.02343756866455078, 0.023461631774902344, 0.02328371238708496, 0.023529472351074218, 0.023412736892700195, 0.02348793601989746, 0.023394527435302733, 0.023666784286499022, 0.02362563133239746, 0.023658559799194335, 0.023503135681152344, 0.023766815185546877, 0.02355574417114258, 0.023476415634155274, 0.023511423110961913, 0.023359487533569336, 0.02342483139038086, 0.023455455780029298, 0.023490655899047853, 0.023548288345336912, 0.023463935852050782, 0.023566335678100587, 0.02345779228210449, 0.023756799697875978, 0.023435264587402343, 0.023436447143554688, 0.023384479522705077, 0.02353740882873535, 0.02340675163269043, 0.02375734329223633, 0.023526527404785155, 0.02362566375732422, 0.02355414390563965, 0.023628639221191405, 0.023545856475830077, 0.023576576232910155, 0.023479520797729494, 0.023604000091552734, 0.023635456085205078, 0.02378598403930664, 0.023541759490966797, 0.023567647933959962, 0.023485151290893555, 0.02345369529724121, 0.023330816268920897, 0.023353343963623048, 0.024657920837402345, 0.02670182418823242, 0.023580671310424805, 0.024350048065185547, 0.02363216018676758, 0.02364044761657715, 0.023745952606201173, 0.023666624069213868, 0.023709407806396486, 0.02359187126159668, 0.02369740867614746, 0.023576576232910155, 0.023635967254638672, 0.023533567428588868, 0.023644128799438478, 0.02353139114379883, 0.023736480712890626, 0.02367487907409668, 0.023726367950439455, 0.023561952590942382, 0.023644159317016602, 0.023564287185668945, 0.023644159317016602, 0.02361667251586914, 0.02353424072265625, 0.023563552856445312, 0.02366556739807129, 0.023447551727294923, 0.023539327621459962, 0.023521663665771485, 0.02364825630187988, 0.02353705596923828, 0.023725824356079103, 0.02364627265930176, 0.02361356735229492, 0.0234967041015625, 0.023409311294555663, 0.02346182441711426, 0.023506208419799803, 0.023423391342163084, 0.0234520320892334, 0.023394304275512694, 0.023433216094970705, 0.023430688858032227, 0.0236343994140625, 0.02349260711669922, 0.023708927154541017, 0.02379648017883301, 0.023731487274169922, 0.02354243278503418, 0.023822399139404298, 0.023553983688354492, 0.023598880767822267, 0.02355023956298828, 0.02364543914794922, 0.02355027198791504, 0.023818687438964845, 0.02355718421936035, 0.023702463150024413, 0.023637983322143556, 0.02369334411621094, 0.02365235137939453, 0.023788799285888673, 0.02370227241516113, 0.023919967651367186, 0.02394588851928711, 0.024335968017578126, 0.0239069766998291, 0.023363359451293947, 0.023546079635620117, 0.023568384170532225, 0.02364825630187988, 0.02335651206970215, 0.023186431884765626, 0.023230112075805665, 0.023402463912963866, 0.023341344833374023, 0.0236168327331543, 0.023314271926879883, 0.02355081558227539, 0.023432384490966796, 0.023470144271850586, 0.02290902328491211, 0.022977184295654297, 0.02289459228515625, 0.023311519622802736, 0.02307302474975586, 0.023215904235839843, 0.02303878402709961, 0.02293328094482422, 0.0226777286529541, 0.023011327743530274, 0.02276959991455078, 0.02279020881652832, 0.02276300811767578, 0.022876672744750977, 0.02272051239013672, 0.02315660858154297, 0.02312716865539551, 0.02324787139892578, 0.02331443214416504, 0.02334716796875, 0.02372719955444336, 0.023352256774902345, 0.022865055084228515, 0.022922079086303712, 0.022793792724609376, 0.022815168380737303, 0.022820863723754883, 0.023043264389038087, 0.023571264266967772, 0.023027231216430664, 0.023144351959228517, 0.023199327468872072, 0.02287289619445801, 0.023074304580688477, 0.022886528015136718, 0.02299139213562012, 0.02303385543823242, 0.023366880416870118, 0.02309404754638672, 0.023197696685791015, 0.023230464935302734, 0.023203840255737306, 0.02313113594055176, 0.02311833572387695, 0.023120384216308593, 0.02327552032470703, 0.023116928100585937, 0.02428108787536621, 0.02344550323486328, 0.023260543823242188, 0.023128704071044923, 0.02284339141845703, 0.022740320205688478, 0.022678176879882814, 0.022855199813842774, 0.022821344375610352, 0.023007232666015624, 0.023166976928710937, 0.02314035224914551, 0.02273219108581543, 0.022700000762939453, 0.022641279220581054, 0.022730752944946288, 0.022675455093383787, 0.0228351993560791, 0.022740991592407226, 0.022992576599121094, 0.022700063705444334, 0.023000511169433593, 0.023165792465209962, 0.02333807945251465, 0.023315359115600585, 0.023392255783081056, 0.02327552032470703, 0.023339008331298827, 0.023154624938964842, 0.02334316825866699, 0.023181312561035155, 0.02344960021972656, 0.023224319458007812, 0.023271360397338868, 0.023212095260620118, 0.023109632492065428, 0.022874111175537108, 0.02283235168457031, 0.02276793670654297, 0.02290121650695801, 0.022801599502563476, 0.023685184478759766, 0.02448851203918457, 0.023138496398925783, 0.02287775993347168, 0.023214527130126953, 0.022855039596557616, 0.02278054428100586, 0.02348646354675293, 0.02285136032104492, 0.022597696304321287, 0.022798175811767577, 0.022740415573120117, 0.023046943664550783, 0.02296022415161133, 0.023158784866333007, 0.022885408401489258, 0.02273369598388672, 0.02257315254211426, 0.02265212821960449, 0.022739168167114257, 0.02266579246520996, 0.022765567779541016, 0.02425651168823242, 0.02450432014465332, 0.02449564743041992, 0.02472742462158203, 0.031978080749511716, 0.02352742385864258, 0.023504896163940428, 0.0235231990814209, 0.02303603172302246, 0.0233123836517334, 0.02315385627746582, 0.023290687561035157, 0.023224319458007812, 0.022990848541259764, 0.022688831329345703, 0.02283932876586914, 0.022901151657104494, 0.02309328079223633, 0.023037824630737305, 0.022950496673583985, 0.02281881523132324, 0.022844640731811524, 0.022993696212768554, 0.023386112213134767, 0.023117055892944337, 0.023061248779296876, 0.02289664077758789, 0.02284339141845703, 0.02273040008544922, 0.022827360153198244, 0.02270207977294922, 0.02306662368774414, 0.023138208389282225, 0.023126079559326173, 0.023152576446533204, 0.023177312850952148, 0.023721536636352538, 0.023816640853881837, 0.023312128067016602, 0.02344166374206543, 0.023388160705566406, 0.023500799179077148, 0.023814079284667968, 0.024817407608032225, 0.024035263061523437, 0.023375999450683593, 0.023172704696655274, 0.023007360458374024, 0.022837791442871094, 0.023006752014160158, 0.022994815826416017, 0.02303027153015137, 0.02304595184326172, 0.022980831146240235, 0.022666559219360352, 0.022749536514282225, 0.022800384521484376, 0.022736480712890625, 0.022644672393798828, 0.02271321678161621, 0.02267955207824707, 0.023828479766845705, 0.0248090877532959, 0.02422166442871094, 0.02366464042663574, 0.02346188735961914, 0.023438976287841796, 0.023300479888916016, 0.023336416244506837, 0.023220415115356444, 0.023320928573608398, 0.02331443214416504, 0.02323967933654785, 0.02297078323364258, 0.022929407119750975, 0.022931488037109374, 0.023011552810668946, 0.02300713539123535, 0.023153087615966798, 0.022967647552490235, 0.022985023498535158, 0.02282326316833496, 0.022738943099975584, 0.022808000564575194, 0.022851551055908203, 0.022796255111694336, 0.02305878448486328, 0.02298819160461426, 0.02285436820983887, 0.022700191497802735, 0.022769311904907226, 0.02287183952331543, 0.023244543075561522, 0.02318623924255371, 0.023117536544799804, 0.023033472061157228, 0.02312259292602539, 0.02303900718688965, 0.023202144622802734, 0.023108224868774414, 0.023150592803955077, 0.023056192398071287, 0.02309343910217285, 0.022990848541259764, 0.023095071792602537, 0.022884576797485352, 0.022775808334350587, 0.022664543151855468, 0.02287273597717285, 0.023154495239257812, 0.022857311248779297, 0.022849311828613283, 0.02281350326538086, 0.022618112564086915, 0.022773759841918945, 0.023173120498657225, 0.022958080291748048, 0.022960128784179686, 0.022781951904296875, 0.02270207977294922, 0.022761472702026365, 0.02272870445251465, 0.02279529571533203, 0.022872447967529297, 0.023422943115234376, 0.023231103897094728, 0.024118783950805665, 0.023091327667236327, 0.02265331268310547, 0.02267046356201172, 0.02254435157775879, 0.022706432342529295, 0.022643360137939453, 0.02305622482299805, 0.023043455123901366, 0.02325984001159668, 0.023365663528442382, 0.023320640563964844, 0.023164928436279295, 0.023637472152709962, 0.0233907527923584, 0.023330816268920897, 0.023195648193359376, 0.023283327102661133, 0.023136383056640626, 0.023138175964355467, 0.023567808151245116, 0.023251903533935546, 0.022848960876464843, 0.0228603515625, 0.0227891845703125, 0.02301228713989258, 0.023004959106445313, 0.022823007583618164, 0.02274675178527832, 0.022931968688964844, 0.022734848022460938, 0.022849536895751952, 0.02268569564819336, 0.022781951904296875, 0.022841344833374022, 0.022826496124267577, 0.022919679641723634, 0.02288435173034668, 0.02292451286315918, 0.02311382484436035, 0.02309190368652344, 0.023121919631958008, 0.02284339141845703, 0.02285977554321289, 0.022804479598999023, 0.022844863891601563, 0.022685407638549804, 0.02272489547729492, 0.02265350341796875, 0.022752864837646485, 0.02270863914489746, 0.02273823928833008, 0.022633152008056642, 0.022824256896972657, 0.022671968460083007, 0.02291516876220703, 0.02300912094116211, 0.023048351287841797, 0.022845312118530272, 0.023006752014160158, 0.022968927383422853, 0.023063840866088866, 0.02303971290588379]",tokens/s,42.78747977521694,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,7394.459648,9457.041408,0.0,9061.793792,8463.626752,s,1,14.789779296875,14.789779296875,0.0,14.789779296875,14.789779296875,14.789779296875,14.789779296875,[14.789779296875],,kWh,0.00021770074485833292,2.400667013197282e-05,9.690757752599255e-05,0.0003386149925162983,,MB,1728.344064,9469.62432,0.0,9061.793792,7981.246464,s,10,57.42651806640625,5.742651806640625,0.003014339069672028,5.7430356445312505,5.7456560546875,5.74627919921875,5.74677771484375,"[5.736865234375, 5.7391357421875, 5.740529296875, 5.741982421875, 5.7416044921875, 5.7440888671875, 5.745517578125, 5.74690234375, 5.745162109375, 5.74472998046875]",tokens/s,44.578708342367115,kWh,0.00016747126064083507,1.8472623745901746e-05,0.00011132408905920182,0.00029726797344593864,tokens/kWh,861175.8509752695,MB,1733.210112,9469.62432,0.0,9061.793792,8267.75296,s,10,27.192193359375,2.7192193359375,0.0020808683062563346,2.718815185546875,2.7216886962890627,2.7225609008789062,2.7232586645507815,"[2.71697265625, 2.7162060546875, 2.71926171875, 2.718357421875, 2.71836865234375, 2.719740478515625, 2.721494873046875, 2.720678466796875, 2.72343310546875, 2.717679931640625]",tokens/s,23.168414245730425,kWh,7.98462075979137e-05,8.806621371887446e-06,5.331282042799756e-05,0.0001419656493977987,tokens/kWh,443769.3221369991,,s,630,27.18830438232421,0.04315603870210193,0.0004030313345170255,0.04318473625183106,0.0436389347076416,0.043749713706970214,0.04395503551483154,"[0.04369161605834961, 0.0429202880859375, 0.04282505416870117, 0.042560001373291016, 0.04252892684936523, 0.0425082893371582, 0.042571327209472654, 0.042624446868896486, 0.042531841278076174, 0.04251206588745117, 0.04270521545410156, 0.04283571243286133, 0.042911998748779295, 0.042708992004394535, 0.04255718231201172, 0.042590465545654294, 0.04277043151855469, 0.04260262298583985, 0.04261193466186523, 0.043023006439208984, 0.043069438934326174, 0.04301004791259765, 0.04294246292114258, 0.04290764617919922, 0.04274995040893555, 0.04309404754638672, 0.043157089233398435, 0.0430780143737793, 0.042982528686523434, 0.04280575942993164, 0.042822017669677734, 0.0429035530090332, 0.04323638534545898, 0.04375651168823242, 0.04356300735473633, 0.04335823822021485, 0.04316342544555664, 0.04296438217163086, 0.04308252716064453, 0.04298342514038086, 0.043130847930908205, 0.04309987258911133, 0.04354694366455078, 0.04351724624633789, 0.04376851272583008, 0.043385921478271486, 0.04356806564331055, 0.043587390899658206, 0.043561153411865235, 0.04382905578613281, 0.04372304153442383, 0.04327414321899414, 0.04323328018188476, 0.04336649703979492, 0.04346665573120117, 0.04350678253173828, 0.04373331069946289, 0.04368239974975586, 0.04339712142944336, 0.04337583923339844, 0.04360675048828125, 0.04361619186401367, 0.04347865676879883, 0.043917343139648436, 0.042971168518066406, 0.04242230224609375, 0.04248982238769531, 0.0422459831237793, 0.04228316879272461, 0.04256361770629883, 0.0425975341796875, 0.04255007934570312, 0.04246323013305664, 0.042912864685058595, 0.042779743194580076, 0.042687873840332034, 0.04252454376220703, 0.042496063232421874, 0.04240166473388672, 0.04293286514282227, 0.0431512336730957, 0.043516033172607424, 0.04324147033691406, 0.042797054290771484, 0.04277644729614258, 0.04286886215209961, 0.04292403030395508, 0.04289737701416016, 0.04279040145874023, 0.04316009521484375, 0.043156959533691405, 0.04288547134399414, 0.043036865234375, 0.042987518310546875, 0.04306687927246094, 0.043080257415771483, 0.04318611145019531, 0.043322368621826174, 0.0432193603515625, 0.04300588989257813, 0.042799774169921874, 0.0429854736328125, 0.04352159881591797, 0.043544864654541014, 0.04344351959228516, 0.04332406234741211, 0.04335161590576172, 0.04355955123901367, 0.043769889831542966, 0.04353839874267578, 0.04321887969970703, 0.04319830322265625, 0.043469215393066404, 0.04349529647827149, 0.04345996856689453, 0.04325235366821289, 0.04325539016723633, 0.04318032073974609, 0.04323452758789063, 0.043614688873291015, 0.04365760040283203, 0.04363468933105469, 0.043611583709716795, 0.04358816146850586, 0.043753471374511715, 0.043919361114501954, 0.044273662567138675, 0.04301824188232422, 0.0425615348815918, 0.04251836776733398, 0.042737823486328125, 0.04262057495117187, 0.04251475143432617, 0.0426844482421875, 0.04270284652709961, 0.04259955215454102, 0.04263155364990234, 0.042743297576904295, 0.04269120025634766, 0.04265407943725586, 0.043087871551513675, 0.042831871032714845, 0.04300732803344726, 0.0428243522644043, 0.04283801651000976, 0.04261478424072265, 0.04260659027099609, 0.04311040115356445, 0.04317593765258789, 0.04298543930053711, 0.04286377716064453, 0.04280819320678711, 0.042904735565185544, 0.04286515045166016, 0.04320495986938477, 0.04335001754760742, 0.043218944549560545, 0.043194366455078126, 0.043022174835205075, 0.042976863861083986, 0.04295699310302734, 0.043286048889160156, 0.043281246185302734, 0.04334499359130859, 0.04343603134155274, 0.04328291320800781, 0.04306784057617188, 0.04326604843139648, 0.04362854385375976, 0.0436357421875, 0.04356195068359375, 0.043632640838623046, 0.04387343978881836, 0.04359987258911133, 0.043426624298095705, 0.043232864379882815, 0.04355321502685547, 0.04350774383544922, 0.04345033645629883, 0.043718593597412106, 0.043726913452148436, 0.04355180740356445, 0.04342870330810547, 0.04323337554931641, 0.043629665374755856, 0.04351273727416992, 0.04347289657592773, 0.04361769485473633, 0.04359228897094727, 0.04399923324584961, 0.04312473678588867, 0.042788864135742184, 0.04256972885131836, 0.04243820953369141, 0.042445247650146484, 0.042657791137695314, 0.04242764663696289, 0.042211231231689454, 0.042654464721679684, 0.042686561584472656, 0.042780670166015625, 0.04268547058105469, 0.04256252670288086, 0.04270489501953125, 0.04274995040893555, 0.04269670486450195, 0.04251574325561523, 0.043086559295654296, 0.04319612884521484, 0.04305292892456054, 0.04307513427734375, 0.043096160888671874, 0.043425537109375, 0.04299059295654297, 0.04272127914428711, 0.04282572937011719, 0.04318207931518555, 0.04315878295898438, 0.04308044815063477, 0.04278409576416015, 0.04266870498657226, 0.04273152160644531, 0.04335327911376953, 0.043389217376708984, 0.04327494430541992, 0.04321059036254883, 0.04334710311889649, 0.043431968688964845, 0.04322537612915039, 0.04332598495483399, 0.04346060943603516, 0.04328409576416015, 0.04324390411376953, 0.04350054550170898, 0.04344934463500977, 0.043649024963378906, 0.043668704986572264, 0.04346540832519531, 0.04351359939575195, 0.04341372680664062, 0.04336838531494141, 0.043278430938720705, 0.04330876922607422, 0.043513278961181644, 0.0437011833190918, 0.043622398376464845, 0.04362035369873047, 0.04372873687744141, 0.04369171142578125, 0.04356143951416016, 0.04359372711181641, 0.04395212936401367, 0.04394675064086914, 0.04282550430297852, 0.042463294982910155, 0.04242432022094727, 0.042649600982666014, 0.04260659027099609, 0.042610431671142576, 0.042481918334960934, 0.042491584777832034, 0.04248128128051758, 0.042533214569091794, 0.04248806381225586, 0.04263945770263672, 0.04289344024658203, 0.04285427093505859, 0.04270694351196289, 0.042727294921875, 0.04289308929443359, 0.0430263671875, 0.043024799346923825, 0.04297318267822266, 0.04312473678588867, 0.04354662322998047, 0.043450366973876955, 0.043184127807617184, 0.04283596801757812, 0.04284774398803711, 0.04299008178710938, 0.04301824188232422, 0.042821792602539065, 0.04299321746826172, 0.04296102523803711, 0.04294416046142578, 0.04318054580688477, 0.043093441009521484, 0.043264190673828126, 0.043458911895751955, 0.043248863220214845, 0.043400001525878903, 0.043334815979003904, 0.043162464141845706, 0.043476608276367186, 0.043620864868164064, 0.04356083297729492, 0.04349731063842773, 0.04396883010864258, 0.043847518920898436, 0.043592864990234376, 0.04351881790161133, 0.04348700714111328, 0.0433355827331543, 0.04317011260986328, 0.043276287078857424, 0.04325785446166992, 0.04372889709472656, 0.04375868988037109, 0.04344022369384765, 0.04343017578125, 0.04335833740234375, 0.04328694534301758, 0.043617313385009765, 0.04361904144287109, 0.0435810546875, 0.04391788864135742, 0.043101505279541014, 0.042659934997558595, 0.04260310363769531, 0.04245625686645508, 0.04251846313476562, 0.042840576171875, 0.042879360198974606, 0.04279292678833008, 0.04272745513916015, 0.04274176025390625, 0.04294451141357422, 0.043005344390869144, 0.042687007904052734, 0.04254521560668945, 0.04264720153808594, 0.04267164611816406, 0.042834335327148435, 0.04280771255493164, 0.042668033599853515, 0.042733345031738285, 0.043071712493896484, 0.04312473678588867, 0.043053054809570314, 0.04302428817749023, 0.04297942352294922, 0.04295270538330078, 0.04294406509399414, 0.04286243057250977, 0.043172382354736326, 0.043280448913574215, 0.04312473678588867, 0.043071487426757815, 0.04316774368286133, 0.04330627059936523, 0.04339990234375, 0.04319968032836914, 0.043195201873779294, 0.043138240814208986, 0.043168350219726564, 0.0431577262878418, 0.04328230285644531, 0.04349942398071289, 0.04327651214599609, 0.0434666862487793, 0.04369974517822266, 0.043638336181640626, 0.043432319641113284, 0.04348579025268555, 0.04352534484863281, 0.04343068695068359, 0.04329676818847656, 0.043716480255126956, 0.04367577743530274, 0.04356915283203125, 0.04364432144165039, 0.043862625122070314, 0.043843582153320314, 0.043720703125, 0.043433982849121096, 0.043390975952148435, 0.0433704948425293, 0.04383961486816406, 0.044400257110595705, 0.04305286407470703, 0.04248223876953125, 0.042251392364501955, 0.04251123046875, 0.04254719924926758, 0.042641407012939454, 0.042372608184814455, 0.042336769104003906, 0.042964607238769534, 0.042903934478759766, 0.04314726257324219, 0.04318207931518555, 0.04309126281738281, 0.04274422454833984, 0.04263504028320313, 0.04284425735473633, 0.04321116638183594, 0.043020286560058595, 0.0431756477355957, 0.04299750518798828, 0.04285446548461914, 0.043006431579589846, 0.043079681396484375, 0.043185344696044924, 0.04321516799926758, 0.04300032043457031, 0.0429567985534668, 0.04302643203735351, 0.042901599884033206, 0.0429403190612793, 0.0433520622253418, 0.04333363342285156, 0.043243518829345705, 0.043235328674316405, 0.043632640838623046, 0.04347084808349609, 0.04317184066772461, 0.04329657745361328, 0.04327443313598633, 0.04322918319702149, 0.04319027328491211, 0.043409534454345707, 0.043443103790283204, 0.04358652877807617, 0.043597824096679685, 0.043492767333984376, 0.04355910491943359, 0.04363510513305664, 0.04339507293701172, 0.04325580978393555, 0.043386207580566404, 0.04356163024902344, 0.043493377685546876, 0.0434436149597168, 0.04352259063720703, 0.04329068756103516, 0.04347644805908203, 0.04386051177978516, 0.043730430603027344, 0.043614688873291015, 0.04366543960571289, 0.04351795196533203, 0.04399484634399414, 0.04309251022338867, 0.04267567825317383, 0.042358592987060545, 0.042380126953125, 0.04255958557128906, 0.042567584991455076, 0.04247347259521484, 0.042446849822998046, 0.04244275283813476, 0.042774528503417966, 0.0433438720703125, 0.04320870590209961, 0.042802783966064455, 0.0426824951171875, 0.0426662712097168, 0.04263971328735352, 0.04294403076171875, 0.04310847854614258, 0.04301532745361328, 0.042952606201171875, 0.04291696166992188, 0.04291056060791015, 0.04301475143432617, 0.043183937072753906, 0.04319692611694336, 0.0432143669128418, 0.04296761703491211, 0.04282540893554688, 0.04287500762939453, 0.043083969116210936, 0.0429854736328125, 0.043194366455078126, 0.04352614212036133, 0.0433889274597168, 0.04336640167236328, 0.04317340850830078, 0.042969566345214844, 0.043253761291503906, 0.04327423858642578, 0.043255966186523435, 0.043890529632568356, 0.04373503875732422, 0.043585601806640624, 0.043317024230957034, 0.043348129272460935, 0.04341459274291992, 0.04354553604125976, 0.04360396957397461, 0.043577407836914064, 0.043370079040527344, 0.04307759857177734, 0.04326438522338867, 0.04366505432128906, 0.043558750152587894, 0.043833408355712894, 0.04373139190673828, 0.04361747360229492, 0.043579681396484375, 0.04350006484985352, 0.043788288116455076, 0.04398284912109375, 0.043649024963378906, 0.043940513610839844, 0.04284931182861328, 0.0425931510925293, 0.04257187271118164, 0.04259404754638672, 0.04281769561767578, 0.04277993774414063, 0.042633377075195315, 0.04259104156494141, 0.04251193618774414, 0.04263068771362305, 0.04245375823974609, 0.04274176025390625, 0.04276633453369141, 0.043044864654541014, 0.04304076766967774, 0.04310220718383789, 0.04293632125854492, 0.0428851203918457, 0.042883071899414066, 0.04328755187988281, 0.043337791442871094, 0.04331206512451172, 0.04318003082275391, 0.04311782455444336, 0.04272742462158203, 0.042868927001953126, 0.043399742126464844, 0.04320665740966797, 0.04311145782470703, 0.04314006423950195, 0.04292755126953125, 0.042856704711914065, 0.04292025756835938, 0.043423744201660154, 0.0433070068359375, 0.04337631988525391, 0.04336633682250977, 0.043524318695068356, 0.04325987243652344, 0.04343622589111328, 0.04341145706176758, 0.04335411071777344, 0.043390975952148435, 0.0436545295715332, 0.04338137435913086, 0.04340879821777344, 0.043555423736572264, 0.04367792129516602, 0.04374489593505859, 0.04369359970092773, 0.043631103515625, 0.043544158935546876, 0.04374512100219727, 0.04375417709350586, 0.04354374313354492, 0.04363052749633789, 0.04362444686889649, 0.04361305618286133, 0.04395622253417969, 0.04384143829345703, 0.04375971221923828, 0.04361759948730469, 0.043859649658203125, 0.0427770881652832, 0.04242208099365234, 0.042065921783447265, 0.042633216857910154, 0.042796607971191405, 0.04267446517944336, 0.042582176208496095, 0.04299769592285156, 0.042452640533447265, 0.042326431274414066, 0.04268236923217773, 0.04276236724853515, 0.04262041473388672, 0.04250467300415039, 0.04256249618530274, 0.042807392120361325, 0.04272422409057617, 0.04289251327514648, 0.04283488082885742, 0.04274883270263672, 0.04311734390258789, 0.04341980743408203, 0.04336844635009766, 0.04322073745727539, 0.0429304313659668, 0.04306739044189453, 0.04295017623901367, 0.0430536003112793, 0.04320832061767578, 0.043321792602539065, 0.04329619216918945, 0.043137470245361326, 0.0429752311706543, 0.04299980926513672, 0.043353248596191406, 0.04333039855957031, 0.0433287353515625, 0.04326819229125976, 0.04333843231201172, 0.043587646484375, 0.043267135620117185, 0.043431968688964845, 0.04344102478027344, 0.04359369659423828, 0.04346883010864258, 0.04353225708007812, 0.043425792694091796, 0.04338079833984375, 0.04334995269775391, 0.04354252624511719, 0.04344627380371094, 0.0436346549987793, 0.04347292709350586, 0.04337209701538086, 0.04330873489379883, 0.043526241302490234, 0.043664096832275394, 0.0434804801940918, 0.04342217636108398, 0.0433092155456543, 0.0436275520324707, 0.043660160064697265]",tokens/s,23.1717282233157,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,8197.976064,12133.00736,0.0,11737.759744,11171.24352,s,1,16.15705859375,16.15705859375,0.0,16.15705859375,16.15705859375,16.15705859375,16.15705859375,[16.15705859375],,kWh,0.0002674889275624385,2.9477125894187513e-05,0.00011286620140399384,0.00040983225486061986,,MB,3844.460544,12147.687424,0.0,11739.856896,10924.35968,s,10,58.036142578124995,5.8036142578125,0.0021390588196081445,5.8043203125,5.805863720703125,5.805999731445312,5.806108540039062,"[5.79872802734375, 5.8013388671875, 5.802384765625, 5.80354345703125, 5.80465673828125, 5.80398388671875, 5.8046845703125, 5.80485302734375, 5.8061357421875, 5.80583349609375]",tokens/s,44.11044370417747,kWh,0.0001693208911624985,1.8675229001102623e-05,0.00011251667334660209,0.0003005127935102032,tokens/kWh,851877.209651336,MB,3844.460544,12149.784576,0.0,11741.954048,10924.36224,s,10,29.07179711914063,2.907179711914063,0.0021990716385634295,2.9076781005859376,2.909408544921875,2.9098297607421877,2.910166733398438,"[2.908518310546875, 2.908570556640625, 2.908798095703125, 2.9102509765625, 2.90931494140625, 2.906340087890625, 2.906837890625, 2.90321435546875, 2.905914794921875, 2.904037109375]",tokens/s,21.67048694713177,kWh,8.492498021875388e-05,9.36892409923487e-06,5.66316008607959e-05,0.0001509255051787847,tokens/kWh,417424.4765678994,,s,630,29.06872789764404,0.046140837932768325,0.00039546610824790876,0.046134544372558595,0.04650437202453613,0.04662674808502197,0.047907263565063474,"[0.047685951232910154, 0.045671966552734374, 0.04559734344482422, 0.04569497680664063, 0.04538380813598633, 0.04575600051879883, 0.046360286712646484, 0.04605587387084961, 0.04597481536865235, 0.04598255920410156, 0.04580147171020508, 0.04569497680664063, 0.04575641632080078, 0.04568016052246094, 0.046053855895996094, 0.0458419189453125, 0.0457564811706543, 0.04612908935546875, 0.046100990295410156, 0.04559872055053711, 0.045706783294677734, 0.04615011215209961, 0.045891166687011715, 0.045817569732666014, 0.04613584136962891, 0.0471404800415039, 0.04574057769775391, 0.046276607513427735, 0.04626227188110352, 0.04591820907592774, 0.04631455993652344, 0.04622022247314453, 0.04603887939453125, 0.046028926849365236, 0.046370849609375, 0.04622441482543945, 0.045937599182128905, 0.045948158264160155, 0.04641667175292969, 0.046349727630615234, 0.04619027328491211, 0.04615875244140625, 0.04602262496948242, 0.04618399810791016, 0.046508289337158205, 0.04642019271850586, 0.04618239974975586, 0.046053375244140625, 0.046537952423095705, 0.04658051300048828, 0.04630323028564453, 0.0465715217590332, 0.04640332794189453, 0.0461759033203125, 0.04671958541870117, 0.04661862564086914, 0.04628873443603516, 0.04673961639404297, 0.0466874885559082, 0.04631763076782226, 0.046364959716796876, 0.04636671829223633, 0.04633203125, 0.04797027206420899, 0.046301185607910154, 0.04573171234130859, 0.04546572875976562, 0.045362945556640624, 0.045827903747558595, 0.046047679901123045, 0.04549369430541992, 0.04565599822998047, 0.04662134552001953, 0.04610403060913086, 0.04573222351074219, 0.04568691253662109, 0.045830142974853515, 0.04561711883544922, 0.04550454330444336, 0.04608988952636719, 0.04595859146118164, 0.0458309440612793, 0.04575859069824219, 0.045946880340576174, 0.04637286376953125, 0.046432254791259765, 0.04611840057373047, 0.04597593688964844, 0.0459218864440918, 0.04634883117675781, 0.04636659240722656, 0.04606140899658203, 0.0459574089050293, 0.0460052490234375, 0.04615248107910156, 0.04637513732910156, 0.04640563201904297, 0.04602470397949219, 0.04631552124023437, 0.046273887634277346, 0.04638172912597656, 0.046274593353271484, 0.04606768035888672, 0.0461036491394043, 0.04602124786376953, 0.04616563034057617, 0.046342494964599606, 0.04674387359619141, 0.0465797119140625, 0.04618598556518555, 0.04669827270507813, 0.046631168365478516, 0.046206527709960935, 0.04613622283935547, 0.04648278427124024, 0.04599862289428711, 0.046225536346435545, 0.04632166290283203, 0.046114398956298826, 0.04626063919067383, 0.046634719848632815, 0.046346527099609375, 0.04645052719116211, 0.04660860824584961, 0.04639328002929687, 0.04619615936279297, 0.04867356872558594, 0.04640361785888672, 0.04590108871459961, 0.0458001594543457, 0.04580556869506836, 0.046233150482177736, 0.045695423126220706, 0.045537281036376956, 0.0462213134765625, 0.04594438552856445, 0.045528831481933596, 0.04609408187866211, 0.045910655975341795, 0.04579155349731445, 0.045914112091064455, 0.04604662322998047, 0.04559222412109375, 0.04549932861328125, 0.045958751678466796, 0.046125473022460936, 0.045969310760498046, 0.046071903228759765, 0.04612108612060547, 0.045873023986816405, 0.046152801513671876, 0.04615171051025391, 0.045933216094970704, 0.04577264022827148, 0.045790752410888674, 0.046196670532226564, 0.046013118743896485, 0.046104801177978515, 0.046243839263916016, 0.04630508804321289, 0.0459747200012207, 0.04605440139770508, 0.04626998519897461, 0.04605795288085938, 0.04594483184814453, 0.04634422302246094, 0.046458847045898435, 0.046247135162353514, 0.04615843200683594, 0.04650358581542969, 0.04660892868041992, 0.046227455139160156, 0.04694012832641602, 0.04652649688720703, 0.046157470703125, 0.04615974426269531, 0.046328289031982425, 0.046045055389404295, 0.04642364883422852, 0.04652291107177734, 0.04630527877807617, 0.04614348983764648, 0.04655926513671875, 0.046268383026123044, 0.046034942626953124, 0.046497791290283204, 0.04655513763427734, 0.04642566299438477, 0.046285247802734374, 0.048330753326416016, 0.0461578254699707, 0.04567606353759766, 0.045634014129638675, 0.045565185546875, 0.04548479843139648, 0.04542233657836914, 0.0456808967590332, 0.046159870147705076, 0.046276222229003905, 0.04629043197631836, 0.04636966323852539, 0.04600831985473633, 0.045926078796386716, 0.0462174072265625, 0.04598729705810547, 0.04562188720703125, 0.0458158073425293, 0.04619203186035156, 0.04594137573242187, 0.045969345092773437, 0.04642412948608399, 0.04625740814208985, 0.045868991851806644, 0.04589241409301758, 0.046233631134033205, 0.04605948638916016, 0.04572137451171875, 0.045891807556152346, 0.045967201232910156, 0.04570854568481445, 0.045969982147216794, 0.04637868881225586, 0.0461910400390625, 0.04601878356933594, 0.046333023071289066, 0.04635465621948242, 0.045873855590820314, 0.046086143493652344, 0.04635443115234375, 0.046153728485107424, 0.04614963150024414, 0.04690502548217773, 0.046682430267333985, 0.04632371139526367, 0.046635009765625, 0.04655923080444336, 0.0463394889831543, 0.04627475357055664, 0.04621542358398437, 0.04620307159423828, 0.04619260787963867, 0.046473217010498044, 0.04616115188598633, 0.04619545745849609, 0.04665676879882812, 0.04666444778442383, 0.04633190536499023, 0.04621279907226562, 0.046725440979003906, 0.046415870666503906, 0.04640361785888672, 0.0467059211730957, 0.04791497421264648, 0.04597577667236328, 0.04602243041992188, 0.04580988693237305, 0.04555699157714844, 0.04589235305786133, 0.046071807861328126, 0.04588336181640625, 0.04542876815795899, 0.04563558578491211, 0.046208576202392576, 0.04593222427368164, 0.04564249420166016, 0.04581785583496094, 0.04615167999267578, 0.04618035125732422, 0.04585062408447266, 0.04628684616088867, 0.0462437744140625, 0.045822017669677736, 0.046253089904785154, 0.04621001434326172, 0.04604111862182617, 0.04593660736083984, 0.04587519836425781, 0.04573593521118164, 0.04585881423950195, 0.04638003158569336, 0.04607664108276367, 0.045912353515625, 0.04572159957885742, 0.046284801483154295, 0.04611276626586914, 0.04610240173339844, 0.04638864135742188, 0.04638793563842773, 0.04635583877563477, 0.04622403335571289, 0.04661644744873047, 0.046458976745605465, 0.04645238494873047, 0.046333824157714844, 0.04664368057250977, 0.046273887634277346, 0.04640774536132813, 0.046219871520996096, 0.046094337463378904, 0.04614144134521484, 0.04649552154541016, 0.04629731369018555, 0.04615155029296875, 0.04656755065917969, 0.046311134338378905, 0.04595296096801758, 0.04627190399169922, 0.04661446380615234, 0.04634521484375, 0.046186496734619144, 0.046524417877197265, 0.04625411224365234, 0.046215137481689456, 0.04657094573974609, 0.046416446685791014, 0.048107391357421876, 0.04614569473266601, 0.04600035095214844, 0.04551446533203125, 0.04519811248779297, 0.046034175872802736, 0.045975616455078125, 0.04581036758422852, 0.04597350311279297, 0.04611686325073242, 0.04603084945678711, 0.04552499389648437, 0.045520126342773436, 0.04597836685180664, 0.04599529647827148, 0.04567728042602539, 0.04565164947509766, 0.04603110504150391, 0.04588959884643555, 0.04604108810424805, 0.046132926940917966, 0.045951297760009766, 0.04591001510620117, 0.046151103973388674, 0.04623996734619141, 0.04588374328613281, 0.04581990432739258, 0.04585027313232422, 0.046274078369140624, 0.045988670349121095, 0.04624796676635742, 0.04623049545288086, 0.04624281692504883, 0.045709312438964846, 0.04623360061645508, 0.04622335815429687, 0.045956417083740236, 0.046055999755859375, 0.04634339141845703, 0.046137344360351565, 0.04629391860961914, 0.04628275299072265, 0.04603289413452148, 0.04649123382568359, 0.0464101448059082, 0.04637635040283203, 0.046106849670410156, 0.046079486846923826, 0.046330753326416015, 0.04647731018066406, 0.04632758331298828, 0.04665315246582031, 0.0466099853515625, 0.04624246215820312, 0.04616016006469727, 0.046565376281738284, 0.04618035125732422, 0.04613324737548828, 0.046295040130615236, 0.0462213134765625, 0.046137344360351565, 0.04644764709472656, 0.04642505645751953, 0.049053215026855466, 0.046432254791259765, 0.04566860961914063, 0.0454760627746582, 0.04546559906005859, 0.04594073486328125, 0.045675807952880856, 0.04574841690063477, 0.04558287811279297, 0.045708961486816406, 0.045486431121826175, 0.04594879913330078, 0.046159038543701174, 0.04601513671875, 0.046020702362060545, 0.0457852783203125, 0.04549836730957031, 0.045606910705566404, 0.046034942626953124, 0.046053375244140625, 0.045837535858154296, 0.04637366485595703, 0.04620479965209961, 0.046181632995605466, 0.045918846130371095, 0.04615590286254883, 0.04619641494750976, 0.04590636825561523, 0.045590526580810545, 0.04621078491210937, 0.0461888313293457, 0.04592559814453125, 0.04643900680541992, 0.04642969512939453, 0.04614556884765625, 0.04609500885009766, 0.04645852661132813, 0.04611872100830078, 0.04586953735351563, 0.04627872085571289, 0.04634767913818359, 0.046201438903808595, 0.04641791915893555, 0.04671078491210937, 0.046358528137207033, 0.04605865478515625, 0.046433120727539065, 0.04614070510864258, 0.046035678863525394, 0.04593628692626953, 0.04615615844726562, 0.04632995223999024, 0.0464422721862793, 0.04641392135620117, 0.04637900924682617, 0.046247230529785154, 0.04636127853393555, 0.046456863403320316, 0.04613014221191406, 0.04606057739257813, 0.046452705383300784, 0.04644659042358398, 0.046166015625, 0.048121726989746094, 0.04624873733520508, 0.04612092971801758, 0.045854015350341795, 0.04557590484619141, 0.04574079895019531, 0.046020862579345706, 0.045956382751464846, 0.04579507064819336, 0.04604617691040039, 0.04610022354125977, 0.045660446166992184, 0.045733856201171874, 0.046013694763183594, 0.04583673477172852, 0.04552473449707031, 0.0458155517578125, 0.046166656494140625, 0.04593068695068359, 0.045625343322753906, 0.04583958435058594, 0.046082111358642576, 0.04605001449584961, 0.04570320129394531, 0.045711326599121097, 0.045676544189453126, 0.04612710571289062, 0.04579660797119141, 0.04612921524047851, 0.04612575912475586, 0.04596310424804687, 0.04614364624023438, 0.045963264465332034, 0.045707263946533204, 0.04571136093139649, 0.04603084945678711, 0.046207008361816404, 0.045975521087646486, 0.046209022521972655, 0.046266368865966793, 0.04597350311279297, 0.04578303909301758, 0.04622492980957031, 0.04628323364257812, 0.04613891220092774, 0.0464040641784668, 0.046360576629638675, 0.04631049728393555, 0.04601500701904297, 0.045994369506835935, 0.04642172622680664, 0.04637036895751953, 0.04630559921264649, 0.04642844772338867, 0.046127231597900394, 0.04606547164916992, 0.04640582275390625, 0.046513504028320315, 0.046457313537597654, 0.046591358184814455, 0.04631216049194336, 0.045973247528076175, 0.04615321731567383, 0.04788838577270508, 0.04567574310302734, 0.04563228988647461, 0.046018558502197264, 0.045694465637207034, 0.04562700653076172, 0.04554617691040039, 0.04596128082275391, 0.04604735946655274, 0.046018558502197264, 0.04585062408447266, 0.045742080688476565, 0.04583974456787109, 0.04572134399414062, 0.0459293441772461, 0.04584198379516601, 0.04611731338500977, 0.04588339233398438, 0.04560627365112305, 0.04551948928833008, 0.04606771087646484, 0.046129024505615235, 0.046155265808105465, 0.04623628616333008, 0.04619571304321289, 0.045943809509277345, 0.04578303909301758, 0.045840385437011716, 0.046159870147705076, 0.04605542373657227, 0.04620492935180664, 0.046227455139160156, 0.04605747222900391, 0.04641996765136719, 0.046607681274414066, 0.046459583282470705, 0.04619171142578125, 0.046082977294921876, 0.04643017578125, 0.04610460662841797, 0.04610012817382812, 0.046559585571289065, 0.046220672607421874, 0.04596316909790039, 0.04626710510253906, 0.04653875350952148, 0.04618361663818359, 0.0460173454284668, 0.045928001403808594, 0.04597574234008789, 0.04626457595825195, 0.04619440078735351, 0.04606595230102539, 0.0462110710144043, 0.04602182388305664, 0.04639622497558594, 0.04657561492919922, 0.04650393676757812, 0.04650300979614258, 0.04654127883911133, 0.046440895080566404, 0.04642816162109375, 0.04624089431762695, 0.047880001068115234, 0.04609452819824219, 0.04592771148681641, 0.04542867279052734, 0.04546160125732422, 0.04594348907470703, 0.04569251251220703, 0.04527145767211914, 0.04559657669067383, 0.046053470611572264, 0.04602675247192383, 0.04567859268188477, 0.04598732757568359, 0.04583033752441406, 0.04581795120239258, 0.04599625778198242, 0.046102142333984374, 0.045986175537109375, 0.04549337768554688, 0.04565900802612305, 0.04610412979125977, 0.04600467300415039, 0.04611686325073242, 0.04613040161132813, 0.0461561279296875, 0.045682113647460935, 0.04587567901611328, 0.04602729415893555, 0.04573507308959961, 0.04572361755371094, 0.04626931381225586, 0.046258174896240234, 0.045963264465332034, 0.04594825744628906, 0.045892127990722655, 0.04602399826049805, 0.04576544189453125, 0.046346176147460935, 0.04641187286376953, 0.046268192291259766, 0.046305473327636716, 0.04620492935180664, 0.04635036849975586, 0.046014430999755856, 0.046353790283203126, 0.046401409149169924, 0.046026878356933594, 0.04627020645141602, 0.046314334869384764, 0.046215198516845704, 0.04663296127319336, 0.04641097640991211, 0.04611334228515625, 0.04589561462402344, 0.0461962890625, 0.04614012908935547, 0.046020606994628906, 0.04637814331054688, 0.046526622772216794, 0.046641857147216796, 0.04665753555297852, 0.04655654525756836, 0.04652300643920899]",tokens/s,21.67277502539284,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,837.832704,558.825472,0.0,163.577856,152.009216,s,1,7.81829833984375,7.81829833984375,0.0,7.81829833984375,7.81829833984375,7.81829833984375,7.81829833984375,[7.81829833984375],,kWh,2.061342946251443e-05,2.2665893650147675e-06,6.590283049973111e-06,2.947030187750231e-05,,MB,1144.639488,607.059968,0.0,199.22944,184.525824,s,30,0.27252745628356934,0.009084248542785644,0.00012152637885798035,0.009049232006072997,0.009166883373260498,0.009344044923782347,0.009509320259094237,"[0.009488736152648926, 0.00904524803161621, 0.009084927558898925, 0.009018976211547852, 0.009019359588623047, 0.008982655525207519, 0.009010944366455078, 0.008999648094177247, 0.0090098876953125, 0.009012224197387696, 0.00906704044342041, 0.00901411247253418, 0.00906492805480957, 0.00910752010345459, 0.00910636806488037, 0.009166848182678223, 0.00910752010345459, 0.009037407875061035, 0.009167200088500976, 0.00909222412109375, 0.009102399826049806, 0.009023072242736817, 0.008996543884277343, 0.009065407752990723, 0.009096256256103516, 0.009053215980529786, 0.009517727851867675, 0.009028767585754395, 0.009033151626586914, 0.009007136344909667]",tokens/s,28180.646840987767,kWh,3.128453211593004e-07,3.448620549427338e-08,2.0713932255303204e-07,5.544708492066058e-07,tokens/kWh,461701458.9068321,MB,1156.018176,611.254272,0.0,203.423744,184.528384,s,30,9.887180541992185,0.3295726847330729,0.0058644436390758664,0.3279118041992187,0.33756495361328126,0.3415366729736328,0.34511653106689455,"[0.3344143981933594, 0.3261952209472656, 0.32928411865234375, 0.32592599487304685, 0.32275234985351564, 0.3224992980957031, 0.32375885009765626, 0.32371170043945313, 0.3241580505371094, 0.32543087768554685, 0.32709075927734377, 0.32853652954101564, 0.33719573974609374, 0.34088787841796875, 0.34636190795898436, 0.33470394897460937, 0.327072021484375, 0.3322218627929687, 0.3420675048828125, 0.3320158386230469, 0.3324908752441406, 0.32653515625, 0.328955810546875, 0.33036843872070315, 0.3292696838378906, 0.32728707885742186, 0.3310740051269531, 0.3248435974121094, 0.3233717041015625, 0.3266993408203125]",tokens/s,191.1566186106257,kWh,9.326762520640574e-06,1.0285978483634633e-06,3.632299232866651e-06,1.3987659601870692e-05,tokens/kWh,4503970.0559752295,,s,1890,9.87304197835922,0.005223831734581599,0.0001788884848357813,0.005182575941085815,0.005374908828735352,0.005455828928947449,0.005879215574264526,"[0.004988800048828125, 0.005204288005828858, 0.005177023887634277, 0.005194784164428711, 0.005305312156677246, 0.005215328216552734, 0.005198751926422119, 0.005212063789367676, 0.005212480068206787, 0.005111680030822754, 0.005128096103668213, 0.005083136081695557, 0.005173344135284424, 0.005100927829742432, 0.005124639987945556, 0.00509116792678833, 0.005113408088684082, 0.0051513280868530275, 0.005275648117065429, 0.006165664196014404, 0.0067430720329284665, 0.006778687953948975, 0.006813856124877929, 0.0058837437629699705, 0.0052367358207702636, 0.005208064079284668, 0.005130239963531494, 0.005141600131988526, 0.005143040180206298, 0.005113440036773682, 0.005102047920227051, 0.005159232139587402, 0.005126175880432129, 0.005117951869964599, 0.005142528057098389, 0.005119552135467529, 0.005118400096893311, 0.005100607872009277, 0.005180287837982178, 0.005127391815185547, 0.005116159915924072, 0.005226655960083008, 0.0052044157981872554, 0.005121119976043701, 0.005147552013397217, 0.005166624069213867, 0.005138912200927734, 0.005115903854370117, 0.0051241922378540035, 0.005178976058959961, 0.00513267183303833, 0.005117856025695801, 0.005166207790374756, 0.005570687770843506, 0.005112192153930664, 0.0057675838470458984, 0.005724160194396972, 0.005162335872650146, 0.005238944053649902, 0.00515558385848999, 0.0060659837722778324, 0.005370048046112061, 0.005129983901977539, 0.004917247772216797, 0.00511187219619751, 0.005158239841461181, 0.0051320638656616215, 0.005188127994537353, 0.005182752132415772, 0.0051682558059692385, 0.005157919883728027, 0.005120863914489746, 0.005170656204223633, 0.005108255863189698, 0.005107391834259034, 0.005155136108398437, 0.0051773438453674315, 0.005170752048492431, 0.005228991985321045, 0.005181439876556396, 0.005172639846801758, 0.0052351999282836915, 0.005164959907531738, 0.005370048046112061, 0.005203904151916504, 0.0051809921264648435, 0.005196224212646485, 0.0051933760643005375, 0.00516703987121582, 0.005232831954956055, 0.005175903797149658, 0.005163008213043213, 0.005151872158050537, 0.005268032073974609, 0.005207712173461914, 0.0051329278945922855, 0.0051538882255554195, 0.005269983768463135, 0.0051487040519714354, 0.0051736001968383785, 0.005154592037200928, 0.005166175842285156, 0.005163519859313965, 0.005194431781768799, 0.005158656120300293, 0.005154784202575684, 0.005186751842498779, 0.005138463973999023, 0.005149663925170899, 0.00515008020401001, 0.005171648025512695, 0.005150784015655518, 0.00513427209854126, 0.005156032085418701, 0.005155648231506348, 0.005132287979125977, 0.00513427209854126, 0.00521398401260376, 0.005157087802886963, 0.005155104160308838, 0.005166656017303467, 0.005158815860748291, 0.005156703948974609, 0.0051964159011840825, 0.005203487873077392, 0.005261727809906006, 0.004899871826171875, 0.005165919780731201, 0.005156991958618164, 0.0051835842132568356, 0.0051905598640441895, 0.0051968002319335935, 0.0051968002319335935, 0.005169280052185058, 0.00528601598739624, 0.0059461121559143066, 0.005298175811767578, 0.005432896137237549, 0.0055710082054138185, 0.00602294397354126, 0.005326208114624024, 0.005280608177185058, 0.005258336067199707, 0.005222432136535645, 0.005272575855255127, 0.0052358717918395995, 0.00532147216796875, 0.005255136013031006, 0.005222400188446045, 0.0053350720405578615, 0.005337215900421142, 0.005215104103088379, 0.005194623947143554, 0.005219744205474853, 0.005178368091583252, 0.005152448177337647, 0.005195775985717774, 0.0051363840103149415, 0.005158912181854248, 0.005238783836364746, 0.00516096019744873, 0.0051571521759033205, 0.005213151931762696, 0.005165887832641602, 0.005135839939117432, 0.00512662410736084, 0.005107295989990234, 0.005199999809265137, 0.005215871810913086, 0.005138944149017334, 0.00513040018081665, 0.005117248058319092, 0.005141183853149414, 0.005126143932342529, 0.005140768051147461, 0.005103007793426513, 0.005087423801422119, 0.005398655891418457, 0.005119999885559082, 0.005288000106811523, 0.005142303943634033, 0.005095583915710449, 0.005111807823181152, 0.005099487781524658, 0.005142560005187988, 0.005128191947937012, 0.005116159915924072, 0.005100319862365723, 0.005151711940765381, 0.00482473611831665, 0.005212736129760742, 0.005142784118652344, 0.00507689619064331, 0.005098944187164307, 0.005112224102020264, 0.005089439868927002, 0.005074783802032471, 0.005115839958190918, 0.005315775871276855, 0.005878655910491943, 0.006088160037994385, 0.00587014389038086, 0.005425119876861572, 0.0051584959030151365, 0.005160799980163574, 0.005155263900756836, 0.005125760078430176, 0.00514844799041748, 0.005155551910400391, 0.005162015914916992, 0.005112095832824707, 0.005112512111663818, 0.005119999885559082, 0.005099743843078613, 0.005123040199279785, 0.005089248180389404, 0.0051307201385498045, 0.0051055998802185054, 0.00514518404006958, 0.005105311870574951, 0.005095615863800049, 0.005105728149414062, 0.005092576026916504, 0.0051207361221313474, 0.0051363840103149415, 0.0051066880226135255, 0.0050800638198852536, 0.005115007877349854, 0.005105792045593261, 0.005190400123596192, 0.00512224006652832, 0.005123904228210449, 0.005120160102844239, 0.005111040115356446, 0.00510422420501709, 0.00510975980758667, 0.005081408023834229, 0.005125823974609375, 0.00510262393951416, 0.00511075210571289, 0.005089280128479004, 0.005115359783172607, 0.0051324481964111325, 0.005190144062042237, 0.005121920108795166, 0.005136447906494141, 0.005136320114135742, 0.00511353588104248, 0.005183519840240479, 0.005131807804107666, 0.005158783912658692, 0.0051495680809020995, 0.004891776084899903, 0.005127871990203857, 0.005142528057098389, 0.005173247814178467, 0.005109119892120362, 0.005155456066131592, 0.0051036162376403805, 0.005148672103881836, 0.005080128192901611, 0.005145535945892334, 0.005094912052154541, 0.005118207931518555, 0.005079264163970947, 0.005130271911621093, 0.005127840042114258, 0.00509552001953125, 0.00514243221282959, 0.005112160205841065, 0.005221407890319824, 0.005116384029388427, 0.005126431941986084, 0.005132512092590332, 0.005103040218353272, 0.005172095775604248, 0.0051049599647521975, 0.005228127956390381, 0.005118559837341308, 0.0051519680023193355, 0.005108704090118408, 0.005072959899902344, 0.0051506562232971195, 0.0051036162376403805, 0.005130239963531494, 0.005093376159667969, 0.005136256217956543, 0.005126527786254883, 0.005101119995117188, 0.005115263938903809, 0.005111743927001953, 0.00513318395614624, 0.005087456226348877, 0.0051847038269042965, 0.005124703884124756, 0.005105440139770508, 0.005089503765106201, 0.0051344318389892575, 0.005105567932128906, 0.0050973758697509764, 0.005107840061187744, 0.005092959880828857, 0.005104000091552734, 0.0050787200927734375, 0.005120319843292236, 0.005097472190856934, 0.005087232112884522, 0.005126016139984131, 0.005097184181213379, 0.005096960067749024, 0.005095712184906006, 0.005106304168701172, 0.005070847988128662, 0.00510262393951416, 0.005136832237243653, 0.004888031959533691, 0.005095967769622803, 0.005125408172607422, 0.005098400115966797, 0.005095071792602539, 0.005081247806549072, 0.005132287979125977, 0.005119999885559082, 0.005150720119476319, 0.005112160205841065, 0.005139488220214844, 0.005104288101196289, 0.0050869441032409665, 0.005101151943206787, 0.005091487884521484, 0.005091040134429931, 0.005085311889648438, 0.005128479957580566, 0.005106048107147217, 0.0050728960037231445, 0.005113664150238037, 0.005085375785827637, 0.005103519916534424, 0.00524399995803833, 0.005132512092590332, 0.005139071941375733, 0.005095583915710449, 0.005142528057098389, 0.005107711791992187, 0.005150720119476319, 0.005115903854370117, 0.005152768135070801, 0.0051036162376403805, 0.005152768135070801, 0.005137504100799561, 0.005155744075775147, 0.00513263988494873, 0.005092864036560059, 0.005132480144500733, 0.005102719783782959, 0.005139296054840088, 0.005107103824615478, 0.005124671936035156, 0.005110079765319824, 0.005098879814147949, 0.00509772777557373, 0.0050748162269592285, 0.005130559921264649, 0.005086143970489502, 0.005145503997802735, 0.0051456642150878905, 0.005103871822357177, 0.00509600019454956, 0.00511190414428711, 0.005086463928222656, 0.005084000110626221, 0.005126143932342529, 0.005127647876739502, 0.005155360221862793, 0.005097472190856934, 0.005115039825439453, 0.005104159832000733, 0.00509164810180664, 0.004860095977783203, 0.005099616050720215, 0.005144959926605225, 0.005108160018920898, 0.005126016139984131, 0.005107615947723389, 0.0051138558387756345, 0.005095712184906006, 0.005110879898071289, 0.005081888198852539, 0.005095583915710449, 0.005093056201934815, 0.005134335994720459, 0.00509065580368042, 0.0050730562210083005, 0.005105535984039306, 0.005074975967407226, 0.005102208137512207, 0.00511520004272461, 0.005130911827087402, 0.005107327938079834, 0.0051279678344726565, 0.005150400161743164, 0.0051166400909423825, 0.005294271945953369, 0.005134335994720459, 0.005144576072692871, 0.005119999885559082, 0.00521830415725708, 0.005105663776397705, 0.005154880046844482, 0.005128223896026612, 0.005113759994506836, 0.005117951869964599, 0.005093376159667969, 0.005152768135070801, 0.005208000183105469, 0.005125664234161377, 0.005153312206268311, 0.005121312141418457, 0.005142367839813232, 0.005321599960327148, 0.005163008213043213, 0.005160543918609619, 0.005093920230865479, 0.0051504640579223635, 0.005133535861968994, 0.005110688209533691, 0.0051129918098449705, 0.005128608226776123, 0.005132863998413086, 0.005115263938903809, 0.005145055770874023, 0.0051337919235229495, 0.00516377592086792, 0.005147456169128418, 0.005154943943023681, 0.005155712127685547, 0.005275455951690674, 0.00514083194732666, 0.005145792007446289, 0.005157343864440918, 0.005165247917175293, 0.004903520107269287, 0.005132287979125977, 0.005154111862182617, 0.005147264003753662, 0.005092864036560059, 0.0051205759048461915, 0.005124095916748047, 0.005119391918182373, 0.005128448009490967, 0.0051244478225708, 0.00511411190032959, 0.0051212477684020995, 0.005125887870788574, 0.005157663822174073, 0.005156576156616211, 0.0051298561096191405, 0.005132959842681885, 0.005128191947937012, 0.005132287979125977, 0.0051567678451538085, 0.005149951934814453, 0.005127007961273193, 0.005115903854370117, 0.0051363840103149415, 0.005185535907745361, 0.005105663776397705, 0.0051333122253417966, 0.005134880065917969, 0.005117599964141846, 0.005124927997589112, 0.005130239963531494, 0.005139743804931641, 0.00513097620010376, 0.00510969591140747, 0.005138495922088623, 0.005137663841247558, 0.0051138558387756345, 0.005329631805419922, 0.005125152111053467, 0.005100063800811768, 0.005120480060577392, 0.005154399871826172, 0.005130784034729004, 0.005134304046630859, 0.0051567678451538085, 0.0051223039627075195, 0.00516707181930542, 0.005150496006011963, 0.00517091178894043, 0.005151008129119873, 0.005121568202972412, 0.005167359828948974, 0.005158944129943848, 0.005154399871826172, 0.005128640174865723, 0.0051645121574401855, 0.005130815982818604, 0.00509555196762085, 0.005107711791992187, 0.0051138558387756345, 0.005105663776397705, 0.005079040050506592, 0.0050936322212219234, 0.004838687896728516, 0.005122432231903076, 0.005089568138122559, 0.005109824180603027, 0.005066751956939697, 0.005122047901153564, 0.005102687835693359, 0.005096288204193116, 0.005090367794036865, 0.005139455795288086, 0.005082399845123291, 0.005092127799987793, 0.00511081600189209, 0.00507587194442749, 0.005085343837738037, 0.005078879833221436, 0.005105855941772461, 0.005110591888427734, 0.005116256237030029, 0.005134816169738769, 0.005097311973571777, 0.005116511821746826, 0.005090047836303711, 0.005135359764099121, 0.005154208183288574, 0.00514739179611206, 0.005132480144500733, 0.00512992000579834, 0.005139904022216797, 0.005147168159484863, 0.00513801622390747, 0.005111999988555908, 0.005105887889862061, 0.005129759788513184, 0.005144927978515625, 0.005128511905670166, 0.005173056125640869, 0.005133887767791748, 0.005147071838378906, 0.005142528057098389, 0.005142335891723633, 0.00516428804397583, 0.005151167869567871, 0.005110335826873779, 0.005121984004974365, 0.0051212158203125, 0.005149824142456055, 0.005262400150299072, 0.005190271854400635, 0.005414912223815918, 0.005136320114135742, 0.005173312187194824, 0.005260992050170899, 0.005153151988983154, 0.005355807781219482, 0.005158656120300293, 0.005144000053405762, 0.005155072212219238, 0.005196000099182129, 0.005189663887023925, 0.005182655811309814, 0.005185664176940918, 0.005164000034332275, 0.004919328212738037, 0.00517571210861206, 0.005123807907104492, 0.005140128135681152, 0.005148672103881836, 0.005124127864837647, 0.005152736186981201, 0.005163743972778321, 0.005162911891937256, 0.0051138558387756345, 0.00516096019744873, 0.005142848014831543, 0.005125408172607422, 0.005140895843505859, 0.005167103767395019, 0.0051586880683898925, 0.00511846399307251, 0.0051463360786437986, 0.005124095916748047, 0.005484384059906006, 0.005137695789337158, 0.005150591850280762, 0.005153791904449463, 0.00514412784576416, 0.0051719040870666505, 0.005115647792816162, 0.005123136043548584, 0.005124544143676758, 0.005128704071044922, 0.005132287979125977, 0.005105311870574951, 0.005142111778259278, 0.005130176067352295, 0.005109568119049072, 0.005123072147369385, 0.005127327919006348, 0.0051178879737854, 0.005118879795074463, 0.0051157760620117185, 0.00514031982421875, 0.005120287895202637, 0.0051223039627075195, 0.005154560089111328, 0.00516096019744873, 0.0051764798164367675, 0.005145440101623535, 0.005177023887634277, 0.005160831928253174, 0.005359903812408447, 0.005220511913299561, 0.0051951041221618656, 0.005233344078063965, 0.005246751785278321, 0.0052154879570007326, 0.005179647922515869, 0.0052353920936584475, 0.005175295829772949, 0.005178751945495605, 0.005143424034118652, 0.005176352024078369, 0.005152671813964844, 0.005185632228851318, 0.005236832141876221, 0.004978432178497315, 0.005244416236877441, 0.005194752216339111, 0.005207871913909912, 0.005178912162780762, 0.0052269759178161625, 0.005187871932983398, 0.005150496006011963, 0.005185472011566162, 0.005193215847015381, 0.0051717119216918945, 0.005240447998046875, 0.005175680160522461, 0.005222400188446045, 0.005173247814178467, 0.00522873592376709, 0.005160160064697265, 0.005160831928253174, 0.005185440063476563, 0.005159743785858154, 0.0051402878761291505, 0.005134016036987305, 0.005144639968872071, 0.005124544143676758, 0.00510368013381958, 0.0051506562232971195, 0.005140160083770752, 0.0051346559524536135, 0.00512553596496582, 0.0051512961387634275, 0.005151840209960937, 0.005153120040893555, 0.005167712211608886, 0.005140607833862305, 0.005128064155578613, 0.005119999885559082, 0.005171199798583984, 0.005167103767395019, 0.00511353588104248, 0.005173567771911621, 0.005152768135070801, 0.005138432025909424, 0.005165056228637695, 0.005224448204040527, 0.005230048179626465, 0.0051617598533630375, 0.005195519924163818, 0.005172895908355713, 0.005202271938323975, 0.005259263992309571, 0.0052633600234985355, 0.005211264133453369, 0.005337952136993408, 0.005318016052246094, 0.005360191822052002, 0.005364128112792969, 0.005260992050170899, 0.005227551937103271, 0.0051844801902771, 0.0051652159690856935, 0.005242720127105713, 0.005152031898498535, 0.0051914558410644535, 0.004919072151184082, 0.005292543888092041, 0.005234399795532227, 0.005251071929931641, 0.005201920032501221, 0.005154816150665284, 0.005175583839416504, 0.005165952205657959, 0.0052145919799804685, 0.005243360042572021, 0.005204256057739258, 0.005226111888885498, 0.005228640079498291, 0.005248511791229248, 0.005259744167327881, 0.0051940159797668455, 0.005228288173675537, 0.005185535907745361, 0.005244224071502685, 0.005196479797363282, 0.005166592121124268, 0.005199423789978027, 0.005196735858917236, 0.005353407859802246, 0.005232704162597656, 0.005250688076019287, 0.005212543964385986, 0.005210112094879151, 0.00517523193359375, 0.005169216156005859, 0.0051807999610900875, 0.005214144229888916, 0.005216159820556641, 0.005184512138366699, 0.00519491195678711, 0.00519868803024292, 0.005148352146148682, 0.005150815963745117, 0.0052080960273742675, 0.005175392150878907, 0.005152703762054444, 0.005197760105133057, 0.005148575782775879, 0.005174367904663086, 0.005223264217376709, 0.0052287039756774905, 0.005261312007904053, 0.005217631816864013, 0.005285791873931885, 0.005235455989837647, 0.005203968048095703, 0.00522649621963501, 0.005238783836364746, 0.005222400188446045, 0.0052425918579101565, 0.0051981120109558104, 0.005203968048095703, 0.005186944007873535, 0.005162816047668457, 0.005221248149871826, 0.005232575893402099, 0.0052787518501281735, 0.00521724796295166, 0.004969056129455566, 0.005335135936737061, 0.005207968235015869, 0.005227935791015625, 0.005193696022033691, 0.00524780797958374, 0.00529798412322998, 0.005328192234039307, 0.005284543991088867, 0.005341184139251709, 0.005252831935882568, 0.00528764820098877, 0.005288512229919434, 0.005305984020233154, 0.005302656173706055, 0.0052984957695007325, 0.005382016181945801, 0.005364927768707275, 0.005353248119354248, 0.005372767925262451, 0.005351424217224121, 0.005380415916442871, 0.005293983936309815, 0.005346591949462891, 0.005306015968322754, 0.005303135871887207, 0.005310143947601318, 0.005376319885253906, 0.005334047794342041, 0.005381120204925537, 0.005367775917053222, 0.005318848133087159, 0.005364799976348877, 0.005361760139465332, 0.0053275198936462405, 0.0053591041564941405, 0.005376512050628662, 0.005380095958709717, 0.00537772798538208, 0.005426623821258545, 0.005471168041229248, 0.005396255970001221, 0.005443136215209961, 0.005412864208221436, 0.005391039848327637, 0.005420959949493408, 0.005389440059661865, 0.005413856029510498, 0.00545577621459961, 0.005390336036682129, 0.005404672145843506, 0.005391551971435547, 0.005425504207611084, 0.005466591835021972, 0.005414912223815918, 0.00539024019241333, 0.005393727779388428, 0.005374752044677735, 0.0053203201293945315, 0.005312511920928955, 0.005322656154632568, 0.005271743774414062, 0.005337376117706299, 0.005039391994476318, 0.005370175838470459, 0.005334688186645508, 0.005325439929962158, 0.005273600101470947, 0.005267583847045899, 0.005281792163848877, 0.00531660795211792, 0.0053344001770019535, 0.005315199851989746, 0.005337056159973145, 0.005318687915802002, 0.005327136039733887, 0.005347199916839599, 0.005422944068908691, 0.005347328186035156, 0.005361120223999024, 0.0053294401168823245, 0.005302239894866943, 0.005339168071746826, 0.005484543800354004, 0.005539072036743164, 0.00532147216796875, 0.005312511920928955, 0.005303487777709961, 0.005303199768066406, 0.0053820481300354, 0.0053608322143554685, 0.005419839859008789, 0.0054579200744628905, 0.00542310380935669, 0.005486176013946533, 0.005448319911956787, 0.0054754562377929685, 0.005499551773071289, 0.005543039798736572, 0.005516128063201904, 0.005474336147308349, 0.005441855907440185, 0.005460864067077637, 0.005430079936981201, 0.0054488320350646975, 0.005526624202728272, 0.005465888023376465, 0.005550079822540284, 0.005503039836883545, 0.005526527881622314, 0.005581759929656983, 0.005484543800354004, 0.005480447769165039, 0.005464096069335938, 0.005479648113250733, 0.005426176071166992, 0.005428991794586182, 0.0053853759765625, 0.005391200065612793, 0.005373407840728759, 0.005414912223815918, 0.005418879985809326, 0.005452383995056152, 0.005433407783508301, 0.0054206719398498535, 0.005355807781219482, 0.005096640110015869, 0.005421216011047363, 0.00542521619796753, 0.005372767925262451, 0.005386079788208008, 0.005431168079376221, 0.005424191951751709, 0.005400991916656494, 0.00543123197555542, 0.005456511974334717, 0.005371551990509033, 0.0054860482215881344, 0.005505375862121582, 0.005476352214813233, 0.0055541439056396484, 0.005634431838989258, 0.006486207962036133, 0.0063569917678833006, 0.006600287914276123, 0.005706143856048584, 0.005570047855377197, 0.005565248012542725, 0.005732031822204589, 0.005527552127838135, 0.005423200130462646, 0.005476255893707275, 0.005471456050872803, 0.005499680042266846, 0.005464032173156739, 0.005480768203735352, 0.005488096237182617, 0.005418399810791016, 0.005413407802581787, 0.005404352188110351, 0.005390399932861328, 0.0053534722328186036, 0.005384768009185791, 0.005342527866363526, 0.005296192169189453, 0.005327775955200196, 0.005361408233642578, 0.005670080184936524, 0.005365920066833496, 0.005916768074035644, 0.005345823764801025, 0.00532700777053833, 0.005330783843994141, 0.005804351806640625, 0.005291007995605469, 0.005302080154418945, 0.005351871967315674, 0.005326687812805176, 0.005341792106628418, 0.005378079891204834, 0.005373631954193115, 0.005371327877044678, 0.0053556480407714845, 0.005388480186462402, 0.0054646081924438475, 0.005390336036682129, 0.005413055896759033, 0.005486400127410889, 0.005433343887329102, 0.005168416023254394, 0.005481376171112061, 0.005430272102355957, 0.005416895866394043, 0.005276544094085693, 0.0052674560546875, 0.005401599884033203, 0.005324863910675049, 0.005297056198120117, 0.005334752082824707, 0.005312992095947266, 0.005302271842956543, 0.005381120204925537, 0.005356383800506592, 0.005316319942474365, 0.005328800201416016, 0.005257599830627442, 0.0052772159576416015, 0.005337567806243897, 0.0053678078651428224, 0.005308191776275634, 0.005342527866363526, 0.005363999843597412, 0.005319295883178711, 0.005358880043029785, 0.005389023780822754, 0.005363071918487549, 0.0053582401275634765, 0.005343200206756592, 0.005349311828613281, 0.0053589119911193845, 0.005310624122619629, 0.0053108158111572264, 0.005294335842132568, 0.005275648117065429, 0.005248159885406494, 0.005310688018798828, 0.005290880203247071, 0.005242623805999756, 0.005289504051208496, 0.005281472206115722, 0.005274335861206055, 0.0053125758171081545, 0.005306367874145508, 0.00528553581237793, 0.005323103904724121, 0.00532700777053833, 0.0052713918685913085, 0.005287871837615967, 0.005273663997650146, 0.0052674560546875, 0.005267104148864746, 0.0052247681617736816, 0.00522979211807251, 0.005177631855010987, 0.005220895767211914, 0.0052074241638183595, 0.005315455913543701, 0.0052624959945678715, 0.00524348783493042, 0.005234687805175781, 0.005287936210632324, 0.005277152061462402, 0.005014592170715332, 0.005278528213500977, 0.005172351837158203, 0.005221343994140625, 0.005145055770874023, 0.005261856079101563, 0.0052032961845397945, 0.005210336208343506, 0.005199776172637939, 0.005190048217773437, 0.005169151782989502, 0.005209919929504395, 0.00522873592376709, 0.005197824001312256, 0.005152768135070801, 0.005162112236022949, 0.00514140796661377, 0.005170720100402832, 0.005171648025512695, 0.005168191909790039, 0.005176256179809571, 0.005185215950012207, 0.005140096187591553, 0.005309120178222656, 0.005139616012573242, 0.005177408218383789, 0.005194528102874756, 0.005174943923950195, 0.005168767929077148, 0.00516809606552124, 0.005183231830596924, 0.005146624088287354, 0.005193503856658935, 0.00518393611907959, 0.0052856640815734865, 0.005202047824859619, 0.005177216053009033, 0.00521830415725708, 0.005228096008300781, 0.005173696041107178, 0.0052219839096069335, 0.005148159980773926, 0.0051905598640441895, 0.005189631938934326, 0.005195295810699463, 0.005155519962310791, 0.005154240131378174, 0.005163104057312012, 0.005330495834350586, 0.005157567977905273, 0.005224480152130127, 0.005137792110443115, 0.00518179178237915, 0.005148416042327881, 0.0051914558410644535, 0.005159647941589356, 0.005146304130554199, 0.005212480068206787, 0.0051580162048339845, 0.005155712127685547, 0.0051651840209960935, 0.005174176216125488, 0.005160064220428467, 0.004895391941070556, 0.005168960094451904, 0.005132224082946777, 0.005187744140625, 0.005169439792633057, 0.005131423950195313, 0.005122752189636231, 0.005177087783813476, 0.005183743953704834, 0.00513427209854126, 0.00517033576965332, 0.0053190398216247554, 0.0051943359375, 0.0051875200271606445, 0.005187583923339844, 0.0051998720169067385, 0.005170207977294922, 0.005202911853790283, 0.005142848014831543, 0.005238463878631592, 0.005181632041931152, 0.005178912162780762, 0.005195487976074218, 0.005126272201538086, 0.005633823871612549, 0.005163360118865967, 0.005185855865478516, 0.005265408039093018, 0.005244927883148193, 0.0053095040321350095, 0.005280704021453858, 0.0052509760856628415, 0.005228640079498291, 0.005294079780578613, 0.005307648181915283, 0.0052432317733764645, 0.005257631778717041, 0.0053125758171081545, 0.005238080024719238, 0.005307007789611817, 0.005296000003814697, 0.005320543766021728, 0.005304800033569336, 0.005269248008728028, 0.00527513599395752, 0.005280320167541504, 0.005351424217224121, 0.005355519771575928, 0.005392384052276611, 0.00535920000076294, 0.005359360218048096, 0.0054289278984069825, 0.005382783889770508, 0.005396448135375977, 0.005433279991149903, 0.005396927833557129, 0.005392191886901855, 0.005360896110534668, 0.005366720199584961, 0.005337215900421142, 0.005365344047546386, 0.005398848056793213, 0.005361631870269775, 0.005116799831390381, 0.005351232051849365, 0.005370048046112061, 0.005410143852233887, 0.006420127868652344, 0.005540863990783691, 0.005390336036682129, 0.005463263988494873, 0.005409440040588379, 0.005365888118743897, 0.005390336036682129, 0.00536575984954834, 0.0052941122055053715, 0.0053731842041015625, 0.005416959762573242, 0.005518208026885987, 0.005485439777374267, 0.005409056186676025, 0.005438144207000732, 0.005431295871734619, 0.005407999992370605, 0.005448448181152344, 0.005451776027679443, 0.005439551830291748, 0.005475999832153321, 0.005433631896972656, 0.005455872058868408, 0.0054139518737792965, 0.005384352207183838, 0.005427487850189209, 0.005374495983123779, 0.005404416084289551, 0.0053311681747436526, 0.0053002238273620605, 0.005268928050994873, 0.005287487983703613, 0.005241856098175048, 0.005207168102264404, 0.005225664138793946, 0.0052397122383117676, 0.005190783977508545, 0.005174943923950195, 0.005224736213684082, 0.005229695796966553, 0.00519868803024292, 0.0053309440612792965, 0.005246784210205078, 0.005221920013427734, 0.005237152099609375, 0.005201695919036865, 0.005232863903045655, 0.005328320026397705, 0.0063136000633239745, 0.008728863716125489, 0.005476255893707275, 0.005293983936309815, 0.005258080005645752, 0.005302624225616455, 0.005238431930541992, 0.005231999874114991, 0.005274240016937256, 0.0052288961410522465, 0.0052486081123352055, 0.004963456153869629, 0.00524563217163086, 0.005227488040924072, 0.005312032222747802, 0.005252863883972168, 0.00522108793258667, 0.005315872192382812, 0.005257184028625488, 0.005221183776855469, 0.005252895832061767, 0.005253536224365234, 0.0052193598747253414, 0.005229279994964599, 0.005236320018768311, 0.005204383850097656, 0.005203711986541748, 0.005259520053863526, 0.005213632106781006, 0.005220928192138672, 0.005251071929931641, 0.0052280001640319824, 0.00521884822845459, 0.005238624095916748, 0.005236320018768311, 0.005241631984710693, 0.005265183925628662, 0.005234272003173828, 0.005267072200775147, 0.00527235221862793, 0.005283040046691894, 0.005268447875976563, 0.005254687786102295, 0.005300511837005615, 0.0053014721870422365, 0.005264448165893555, 0.005303775787353516, 0.00528166389465332, 0.0052780799865722655, 0.005304448127746582, 0.005270431995391846, 0.0052717761993408205, 0.0052846078872680665, 0.005267263889312744, 0.00521235179901123, 0.005289919853210449, 0.005236000061035156, 0.005235487937927246, 0.005316351890563965, 0.005241087913513184, 0.0052674560546875, 0.005246975898742676, 0.005230591773986816, 0.0052791681289672855, 0.0053043198585510255, 0.005344096183776855, 0.005316319942474365, 0.0053613119125366215, 0.005327199935913086, 0.005321760177612305, 0.005304351806640625, 0.0053441920280456546, 0.005324416160583496, 0.005285247802734375, 0.00495420789718628, 0.005272607803344726, 0.005307295799255371, 0.0053108158111572264, 0.0053075838088989255, 0.005519328117370606, 0.005347487926483155, 0.005324128150939941, 0.005309599876403809, 0.00529807996749878, 0.005273536205291748, 0.005324128150939941, 0.005253791809082031, 0.005265408039093018, 0.00526694393157959, 0.005274112224578857, 0.005265408039093018, 0.005334527969360352, 0.005274112224578857, 0.005237823963165284, 0.0052581758499145505, 0.005278719902038574, 0.005248032093048096, 0.005351583957672119, 0.005286943912506103, 0.005263423919677735, 0.005298367977142334, 0.0052936959266662595, 0.005285024166107178, 0.0053010878562927246, 0.005514143943786621, 0.005330016136169433, 0.005792128086090088, 0.0053144960403442385, 0.005286272048950195, 0.005287807941436767, 0.005648736000061035, 0.005291232109069824, 0.005230751991271973, 0.0052302079200744625, 0.005230944156646728, 0.005253280162811279, 0.00526796817779541, 0.005207935810089112, 0.0052633600234985355, 0.005238463878631592, 0.005226848125457764, 0.005210207939147949, 0.005216256141662597, 0.005165056228637695, 0.005160448074340821, 0.005184000015258789, 0.005171520233154297, 0.005157792091369629, 0.005184351921081543, 0.005128223896026612, 0.0051437759399414066, 0.005130943775177002, 0.005210112094879151, 0.0051456642150878905, 0.005157760143280029, 0.005205183982849121, 0.005243775844573975, 0.004900383949279785, 0.0052368960380554195, 0.00521017599105835, 0.005234687805175781, 0.005236320018768311, 0.00522819185256958, 0.005212672233581543, 0.005193920135498047, 0.005196864128112793, 0.005172416210174561, 0.005160319805145264, 0.005191264152526856, 0.005174047946929931, 0.005205376148223877, 0.005239200115203857, 0.005189280033111573, 0.005195615768432617, 0.005235424041748047, 0.005196991920471191, 0.005198624134063721, 0.005175327777862549, 0.005212224006652832, 0.005216191768646241, 0.005181727886199951, 0.0052211518287658696, 0.0051943678855896, 0.0051346559524536135, 0.005193727970123291, 0.0051404800415039064, 0.00516870403289795, 0.0051792640686035155, 0.0051875200271606445, 0.005207712173461914, 0.00516809606552124, 0.0052135357856750485, 0.00518006420135498, 0.005158847808837891, 0.005167263984680176, 0.00515231990814209, 0.005162335872650146, 0.00517632007598877, 0.005177023887634277, 0.005153088092803955, 0.005162816047668457, 0.005225728034973145, 0.00514899206161499, 0.005155456066131592, 0.005122208118438721, 0.005175136089324951, 0.005136320114135742, 0.005099743843078613, 0.00514031982421875, 0.005107615947723389, 0.005154911994934082, 0.005161983966827393, 0.0052070398330688475, 0.005166399955749512, 0.0051430721282958985, 0.0051874880790710445, 0.005175551891326904, 0.005158559799194336, 0.0051838397979736325, 0.00514243221282959, 0.004873983860015869, 0.005161215782165527, 0.005197824001312256, 0.005212160110473632, 0.00517299222946167, 0.005158175945281982, 0.005153600215911865, 0.005152959823608399, 0.00516707181930542, 0.005154943943023681, 0.005132160186767578, 0.005166111946105957, 0.005128672122955322, 0.005169760227203369, 0.00513424015045166, 0.00513801622390747, 0.0051677761077880855, 0.005136127948760986, 0.005132512092590332, 0.005140031814575195, 0.00517139196395874, 0.005124127864837647, 0.005131423950195313, 0.005135136127471924, 0.005146368026733399, 0.005226816177368164, 0.005177279949188232, 0.005183263778686524, 0.005177631855010987, 0.005208352088928223, 0.005192416191101074, 0.005184768199920654, 0.005235583782196045, 0.0052089600563049315, 0.005182496070861816, 0.005227488040924072, 0.005273952007293701, 0.005592351913452149, 0.005508768081665039, 0.005796639919281006, 0.005468224048614502, 0.0057998719215393065, 0.005203904151916504, 0.005233888149261475, 0.005161087989807129, 0.005220191955566406, 0.00518230390548706, 0.005213568210601807, 0.005191936016082764, 0.005191391944885254, 0.005296768188476563, 0.005203392028808594, 0.005177631855010987, 0.005222815990447998, 0.005155712127685547, 0.005190656185150146, 0.005219840049743653, 0.005226208209991455, 0.005208032131195068, 0.005182144165039063, 0.0052286720275878905, 0.005177663803100586, 0.005203648090362549, 0.004898528099060059, 0.005188127994537353, 0.005572800159454346, 0.0051998720169067385, 0.005212160110473632, 0.005167103767395019, 0.005209695816040039, 0.005212639808654785, 0.005181375980377197, 0.005246975898742676, 0.005206143856048584, 0.0052202239036560055, 0.005277344226837158, 0.005203519821166992, 0.005185887813568115, 0.005201791763305664, 0.00522211217880249, 0.005242976188659668, 0.005190400123596192, 0.005201920032501221, 0.005232639789581299, 0.005234848022460938, 0.00528380823135376, 0.005252927780151367, 0.005238495826721192, 0.005273952007293701, 0.005238783836364746, 0.005305920124053955, 0.005238495826721192, 0.005227231979370117, 0.005224448204040527, 0.0051998720169067385, 0.005230591773986816, 0.005226816177368164, 0.00521292781829834, 0.005256383895874023, 0.005240575790405273, 0.005242015838623047, 0.005247039794921875, 0.0052106881141662596, 0.005265183925628662, 0.005259712219238281, 0.005244927883148193, 0.005276832103729248, 0.005251167774200439, 0.005287744045257568, 0.005266528129577636, 0.005253024101257324, 0.005274975776672363, 0.005363296031951905, 0.005231711864471435, 0.005263519763946533, 0.005305632114410401, 0.005261792182922363, 0.0053012800216674805, 0.005190432071685791, 0.00517958402633667, 0.005212063789367676, 0.0051981120109558104, 0.005220128059387207, 0.0052080960273742675, 0.005276991844177246, 0.0052353920936584475, 0.0049963841438293456, 0.0052623038291931155, 0.005205728054046631, 0.0052304000854492185, 0.00522876787185669, 0.005253344058990478, 0.005216191768646241, 0.005229663848876953, 0.005216351985931397, 0.005259967803955078, 0.005247039794921875, 0.0053144640922546385, 0.005248703956604004, 0.005250688076019287, 0.005218080043792725, 0.00521833610534668, 0.005253600120544434, 0.005232160091400147, 0.005187647819519043, 0.00519865608215332, 0.005194752216339111, 0.005198847770690918, 0.005220608234405517, 0.005170623779296875, 0.005183807849884033, 0.005163328170776367, 0.005189311981201172, 0.005154176235198975, 0.005185152053833008, 0.0059955201148986816, 0.005222655773162842, 0.005168896198272705, 0.005203711986541748, 0.005187679767608642, 0.005183296203613281, 0.005207551956176758, 0.005180255889892578, 0.005167103767395019, 0.0051645441055297855, 0.0052202239036560055, 0.005201536178588867, 0.005173727989196777, 0.005216991901397705, 0.005259071826934815, 0.005265408039093018, 0.005339104175567627, 0.005233952045440674, 0.005179200172424316, 0.005224575996398926, 0.005165952205657959, 0.005178463935852051, 0.005151584148406983, 0.0051998720169067385, 0.005180960178375244, 0.005155295848846436, 0.00517523193359375, 0.005174943923950195, 0.005177919864654541, 0.005201759815216064, 0.005224703788757324, 0.0053452482223510745, 0.005148575782775879, 0.005205728054046631, 0.005071680068969727, 0.006368544101715088, 0.005179967880249024, 0.005208415985107422, 0.005132287979125977, 0.0051354880332946775, 0.0051454720497131345, 0.005187839984893799, 0.005137856006622314, 0.005128511905670166, 0.0051571521759033205, 0.005157887935638428, 0.005153120040893555, 0.005134719848632813, 0.005198048114776611, 0.005168352127075195, 0.005138336181640625, 0.005173759937286377, 0.0051528639793396, 0.0051179838180541995, 0.0050833277702331545, 0.005119840145111084, 0.005111807823181152, 0.005128384113311768, 0.00514140796661377, 0.005135263919830322, 0.005139808177947998, 0.005124767780303955, 0.005142528057098389, 0.005133664131164551, 0.0051271038055419925, 0.00516870403289795, 0.005127999782562256, 0.005126336097717285, 0.005114367961883545, 0.005180480003356934, 0.00512662410736084, 0.005175519943237305, 0.005246880054473877, 0.005171199798583984, 0.005140672206878662, 0.005127647876739502, 0.005163360118865967, 0.005121119976043701, 0.005134528160095215, 0.0051535038948059084, 0.005142528057098389, 0.005209695816040039, 0.005201344013214111, 0.005254111766815186, 0.00524399995803833, 0.005190656185150146, 0.0052174081802368166, 0.0052334399223327634, 0.005187583923339844, 0.005240543842315674, 0.005220895767211914, 0.005186751842498779, 0.005196352005004883, 0.00519379186630249, 0.005208000183105469, 0.005185823917388916, 0.005561439990997314, 0.00509503984451294, 0.00523967981338501, 0.005183487892150879, 0.005215871810913086, 0.00515231990814209, 0.005192160129547119, 0.005150176048278808, 0.005218880176544189, 0.005203519821166992, 0.005202688217163086, 0.005197824001312256, 0.0051363840103149415, 0.005156256198883056, 0.005197408199310302, 0.0052009282112121585, 0.005189760208129883, 0.005152512073516846, 0.005187679767608642, 0.005183775901794433, 0.005176383972167969, 0.005170944213867187, 0.00516380786895752, 0.005167295932769775, 0.005162911891937256, 0.005572800159454346, 0.005828447818756103, 0.0052631359100341795, 0.005767615795135498, 0.005229695796966553, 0.005311135768890381, 0.005228352069854736, 0.005204031944274903, 0.005660927772521972, 0.0052202239036560055, 0.005207647800445556, 0.005239200115203857, 0.005197824001312256, 0.005183775901794433, 0.00522160005569458, 0.005226687908172607, 0.00524457597732544, 0.005216479778289795, 0.005257120132446289, 0.005224991798400879, 0.005208064079284668, 0.005242879867553711, 0.005295584201812744, 0.0054494719505310055, 0.005304096221923828, 0.005291007995605469, 0.005244160175323486, 0.005354239940643311, 0.005268832206726074, 0.005288479804992675, 0.0053060798645019535, 0.0052494401931762694, 0.005238783836364746, 0.005197663784027099, 0.005199039936065674, 0.00515990400314331, 0.005173247814178467, 0.005172927856445312, 0.005171520233154297, 0.004961984157562256, 0.005313920021057129, 0.0051923198699951175, 0.005285759925842285, 0.005164639949798584, 0.005171199798583984, 0.005171455860137939, 0.00522211217880249, 0.005232480049133301, 0.005157663822174073, 0.005185472011566162, 0.005195551872253418, 0.00517142391204834, 0.005184607982635498, 0.005155807971954346, 0.0051608958244323735, 0.005123551845550537, 0.005139135837554932, 0.005154272079467773, 0.005098176002502441, 0.005213568210601807, 0.005105343818664551, 0.005139071941375733, 0.005124095916748047, 0.005121280193328857, 0.005118720054626465, 0.005097472190856934, 0.005100671768188477, 0.005126880168914795, 0.005111968040466309, 0.005105663776397705, 0.005135903835296631, 0.005110400199890137, 0.005111392021179199, 0.005195551872253418, 0.00516048002243042, 0.005224671840667725, 0.005181951999664307, 0.00513804817199707, 0.005106272220611573, 0.0052633600234985355, 0.005107615947723389, 0.005112959861755371, 0.005104640007019043, 0.005101759910583496, 0.005170783996582031, 0.005123487949371338, 0.005098112106323242, 0.0051212158203125, 0.0051660480499267575, 0.005097472190856934, 0.005109407901763916, 0.005153120040893555, 0.005156576156616211, 0.0051857600212097165, 0.005127583980560303, 0.005173920154571533, 0.005203968048095703, 0.005122047901153564, 0.005130239963531494, 0.005137504100799561, 0.005112287998199463, 0.005112256050109863, 0.004868480205535888, 0.005117536067962647, 0.005118271827697754, 0.005170944213867187, 0.005097951889038086, 0.005107232093811035, 0.005114175796508789, 0.005120223999023438, 0.005110079765319824, 0.005215936183929444, 0.005101823806762695, 0.005100319862365723, 0.005108704090118408, 0.0051138558387756345, 0.0051036162376403805, 0.005123744010925293, 0.005122047901153564, 0.005107168197631836, 0.005095680236816406, 0.005091231822967529, 0.0050993280410766605, 0.005122047901153564, 0.005104544162750244, 0.005079040050506592, 0.005130239963531494, 0.005119808197021484, 0.0051099519729614256, 0.005135359764099121, 0.005132768154144287, 0.0051307520866394046, 0.005124127864837647, 0.005162879943847656, 0.005163136005401611, 0.0051468157768249515, 0.005207871913909912, 0.005192927837371826, 0.0051691198348999025, 0.005110239982604981, 0.005185887813568115, 0.0051708478927612305, 0.005163392066955566, 0.005151008129119873, 0.005161920070648193, 0.005119935989379883, 0.00509830379486084, 0.005122047901153564, 0.0051335678100585935, 0.005112768173217774, 0.005197184085845947, 0.005130688190460205, 0.005126016139984131, 0.0051159682273864745, 0.005138783931732178, 0.0051290240287780765, 0.0051608958244323735, 0.0051765117645263675, 0.005138207912445068, 0.005111264228820801, 0.005089824199676514, 0.005111807823181152, 0.00510975980758667, 0.005092991828918457, 0.005128575801849365, 0.004975872039794922, 0.005141248226165771, 0.005150015830993652, 0.0051528959274292, 0.005130623817443848, 0.005139711856842041, 0.005159872055053711, 0.005148672103881836, 0.0051734399795532226, 0.005199135780334473, 0.005153632164001465, 0.005117631912231445, 0.005660672187805176, 0.005142528057098389, 0.005149759769439697, 0.005143008232116699, 0.005105184078216553, 0.005110720157623291, 0.005115231990814209, 0.005135007858276367, 0.005109920024871826, 0.005101408004760742, 0.005166528224945069, 0.005195839881896973, 0.005122560024261475, 0.005137919902801514, 0.005135007858276367, 0.005125984191894532, 0.005212160110473632, 0.005123456001281738, 0.005134975910186768, 0.0051211838722229, 0.005112095832824707, 0.0051504321098327635, 0.005126336097717285, 0.005149343967437744, 0.005139776229858398, 0.005117824077606201, 0.005112639904022217, 0.0051660799980163576, 0.005131264209747314, 0.005243135929107666, 0.005103360176086426, 0.006318079948425293, 0.005168223857879638, 0.005143455982208252, 0.005154816150665284, 0.005201888084411621, 0.005109792232513428, 0.005128191947937012, 0.005119999885559082, 0.0051212477684020995, 0.005102591991424561, 0.005137184143066406, 0.005139455795288086, 0.005093376159667969, 0.005162015914916992, 0.005659264087677002, 0.005111648082733154, 0.005169760227203369, 0.005127168178558349, 0.005194655895233154, 0.00539631986618042]",tokens/s,191.43036200420318,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: DeciLMForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3128.410112,4602.134528,0.0,4206.886912,4087.771648,s,1,11.1307451171875,11.1307451171875,0.0,11.1307451171875,11.1307451171875,11.1307451171875,11.1307451171875,[11.1307451171875],,kWh,0.00011638537738329735,1.2828940313779464e-05,4.410392417200648e-05,0.0001733182418690833,,MB,3141.451776,4774.100992,0.0,4366.270464,4273.703936,s,10,14.535161621093751,1.4535161621093748,0.0024993441056309243,1.4539345092773437,1.456260693359375,1.45683505859375,1.45729455078125,"[1.4518798828125, 1.4514942626953125, 1.4502095947265625, 1.4495482177734376, 1.454238037109375, 1.4536309814453126, 1.456133056640625, 1.4554744873046874, 1.4551436767578125, 1.457409423828125]",tokens/s,176.1246325795835,kWh,4.26348872075035e-05,4.7005923584355155e-06,2.844610609019915e-05,7.578158565613817e-05,tokens/kWh,3378129.367226621,MB,3141.451776,4776.198144,0.0,4368.367616,4274.361856,s,10,21.31460498046875,2.131460498046875,0.01040247824753476,2.1338648681640624,2.13996044921875,2.14533935546875,2.14964248046875,"[2.134578369140625, 2.116716796875, 2.113841552734375, 2.13179150390625, 2.15071826171875, 2.13876513671875, 2.1374931640625, 2.13352880859375, 2.134200927734375, 2.122970458984375]",tokens/s,29.55719801409827,kWh,6.293695610124964e-05,6.94365392428867e-06,3.6816446119800065e-05,0.00010669705614533838,tokens/kWh,590456.7780594056,,s,630,21.31210797500611,0.03382874281747,0.0006934402196118866,0.03361788940429687,0.03453700675964356,0.034875053215026854,0.03590497573852539,"[0.03524480056762695, 0.03443711853027344, 0.03434636688232422, 0.03448486328125, 0.03485887908935547, 0.03460905456542969, 0.03467190551757812, 0.03458553695678711, 0.034907520294189455, 0.03875084686279297, 0.034457599639892575, 0.034369537353515625, 0.03438796615600586, 0.03382812881469727, 0.03398294448852539, 0.03461084747314453, 0.03358486557006836, 0.03357990264892578, 0.033707839965820316, 0.033955230712890624, 0.03351836776733398, 0.0337999038696289, 0.03338787078857422, 0.03337311935424805, 0.03334143829345703, 0.033355777740478515, 0.033468414306640625, 0.033535999298095705, 0.033341598510742185, 0.033810272216796874, 0.033699840545654294, 0.03349420928955078, 0.033534782409667965, 0.03382675170898437, 0.03360979080200195, 0.0336097297668457, 0.0346295051574707, 0.03384332656860352, 0.03359539031982422, 0.033404800415039064, 0.03345625686645508, 0.03340284729003906, 0.0334541130065918, 0.03343270492553711, 0.03348889541625977, 0.033409408569335934, 0.03338905715942383, 0.0331673583984375, 0.033296382904052735, 0.033237152099609375, 0.033365856170654296, 0.033552257537841794, 0.03340531158447266, 0.03328351974487305, 0.033497535705566406, 0.03331651306152344, 0.03358451080322265, 0.0349315185546875, 0.034019359588623045, 0.03354009628295898, 0.03351347351074219, 0.03359139251708984, 0.03346601486206055, 0.03485081481933594, 0.0339046401977539, 0.03373446273803711, 0.03361587142944336, 0.03446393585205078, 0.034231647491455075, 0.033909408569335935, 0.03373875045776367, 0.03372851181030274, 0.033585151672363284, 0.03336806488037109, 0.03338240051269531, 0.03338649749755859, 0.03320627212524414, 0.03334143829345703, 0.033265056610107424, 0.03338915252685547, 0.033445888519287106, 0.033345535278320314, 0.03319804763793945, 0.03346799850463867, 0.03348320007324219, 0.03338560104370117, 0.03343244934082031, 0.033296382904052735, 0.03321596908569336, 0.03344643020629883, 0.03333324813842774, 0.03336294555664063, 0.03417190551757812, 0.03458364868164063, 0.03377388763427734, 0.03366764831542969, 0.03356470489501953, 0.03335987091064453, 0.03318796920776367, 0.03315206527709961, 0.03334428787231445, 0.03331878280639648, 0.03361772918701172, 0.03335120010375976, 0.03369043350219727, 0.03432470321655273, 0.034840545654296874, 0.03355148696899414, 0.03350291061401367, 0.03386617660522461, 0.033546558380126955, 0.03343795013427735, 0.03408662414550781, 0.0336091194152832, 0.0336446418762207, 0.0336596794128418, 0.03345388793945313, 0.033553791046142575, 0.03346720123291016, 0.03337830352783203, 0.03336521530151367, 0.03340073776245117, 0.033260414123535156, 0.03333052825927734, 0.033374881744384764, 0.03351513671875, 0.03515475082397461, 0.03435721588134766, 0.03401859283447266, 0.03362284851074219, 0.03367724609375, 0.03343564987182617, 0.0334516487121582, 0.03359945678710938, 0.03329679870605469, 0.0336732177734375, 0.033255424499511715, 0.03328585433959961, 0.0332720947265625, 0.03329974365234375, 0.03322496032714844, 0.03444070434570313, 0.03347964859008789, 0.033757217407226564, 0.033767391204833984, 0.033468318939208985, 0.03341494369506836, 0.033421630859375, 0.033537406921386716, 0.033321567535400394, 0.033140384674072265, 0.03339209747314453, 0.033197216033935543, 0.0332490234375, 0.033259521484375, 0.03355033493041992, 0.03332499313354492, 0.03354425430297851, 0.03320627212524414, 0.03343769454956055, 0.03337638473510742, 0.033736766815185545, 0.03374675369262695, 0.033519615173339845, 0.03327385711669922, 0.03383091354370117, 0.033492992401123044, 0.03391078567504883, 0.033729663848876955, 0.03352665710449219, 0.03325337600708008, 0.03357491302490234, 0.033388031005859374, 0.03348112106323242, 0.03328435134887695, 0.03335532760620117, 0.033315006256103515, 0.03334537506103516, 0.034027328491210936, 0.03389689636230469, 0.03348611068725586, 0.03356950378417969, 0.033459999084472655, 0.03376668930053711, 0.03325024032592774, 0.03342937469482422, 0.03349843215942383, 0.03356073760986328, 0.033955711364746094, 0.03519667053222656, 0.03428579330444336, 0.03436057662963867, 0.03408972930908203, 0.034457599639892575, 0.03432467269897461, 0.034062145233154296, 0.03401897430419922, 0.034043617248535156, 0.03413033676147461, 0.033884033203125, 0.033998367309570315, 0.03347743988037109, 0.03369574356079102, 0.03390067291259766, 0.03480131149291992, 0.03403571319580078, 0.0338741455078125, 0.03381452941894531, 0.03448585510253906, 0.03376988983154297, 0.03363225555419922, 0.03365801620483398, 0.034005599975585936, 0.03415884780883789, 0.03388415908813477, 0.03377062225341797, 0.03370800018310547, 0.033719200134277344, 0.03371731185913086, 0.033587329864501955, 0.033479103088378905, 0.03345446395874024, 0.03349651336669922, 0.033476158142089846, 0.03359641647338867, 0.035590145111083986, 0.033871295928955075, 0.03372499084472656, 0.03383295822143555, 0.03360153579711914, 0.03356467056274414, 0.033508575439453125, 0.03358595275878906, 0.03347030258178711, 0.033536128997802735, 0.03361407852172851, 0.03351116943359375, 0.03387334442138672, 0.03387241744995117, 0.033443264007568356, 0.03351030349731445, 0.03365788650512695, 0.03361417770385742, 0.033485313415527344, 0.03354403305053711, 0.033594528198242185, 0.033618785858154296, 0.033562625885009766, 0.033709598541259767, 0.03350147247314453, 0.03360713577270508, 0.03350191879272461, 0.03519452667236328, 0.034626911163330075, 0.03562188720703125, 0.035020801544189455, 0.03497539138793945, 0.03465439987182617, 0.034654369354248045, 0.035818912506103515, 0.036297534942626955, 0.03439574432373047, 0.03470684814453125, 0.03447040176391602, 0.03426947021484375, 0.03389379119873047, 0.03353596878051758, 0.033414878845214845, 0.033436416625976566, 0.033678592681884764, 0.033211135864257814, 0.03340227127075195, 0.033985279083251954, 0.03504435348510742, 0.034374046325683597, 0.03395340728759766, 0.03429475021362305, 0.03396953582763672, 0.03391897583007813, 0.035074302673339844, 0.034111873626708984, 0.03402454376220703, 0.043103134155273434, 0.034207454681396486, 0.034044193267822265, 0.034252799987792966, 0.033756832122802734, 0.033638591766357424, 0.03366057586669922, 0.03353036880493164, 0.03339033508300781, 0.03342361450195312, 0.03408617782592773, 0.03377961730957031, 0.03401612854003906, 0.03394118499755859, 0.03378163146972656, 0.033433982849121094, 0.03334761428833008, 0.033512641906738284, 0.03325763320922852, 0.033277664184570316, 0.03357583999633789, 0.03331404876708984, 0.033640865325927735, 0.033333087921142576, 0.03326822280883789, 0.0334964485168457, 0.03339123153686523, 0.03335372924804687, 0.03343360137939453, 0.03351718521118164, 0.03352630233764648, 0.033640289306640626, 0.03350899124145508, 0.03567241668701172, 0.03443526458740234, 0.034310142517089845, 0.0341399040222168, 0.03424895858764648, 0.033976318359375, 0.03405353546142578, 0.03402812957763672, 0.03418521499633789, 0.03467411041259766, 0.03457430267333984, 0.03488828659057617, 0.034590015411376955, 0.03450697708129883, 0.034762527465820314, 0.034667198181152346, 0.03503308868408203, 0.03483145523071289, 0.03480611038208008, 0.03467337417602539, 0.03451478576660156, 0.03427471923828125, 0.03432422256469726, 0.03459302520751953, 0.033995361328125, 0.03363840103149414, 0.0336814079284668, 0.033428577423095705, 0.03355126571655274, 0.03370598220825195, 0.033503231048583985, 0.03344300842285156, 0.03348748779296875, 0.033414783477783205, 0.033304222106933595, 0.03352054214477539, 0.033306625366210936, 0.03334918212890625, 0.034517120361328125, 0.03396163177490234, 0.03332777786254883, 0.033486846923828126, 0.03319561767578125, 0.033673633575439454, 0.034062335968017575, 0.0337894401550293, 0.0344725456237793, 0.03496656036376953, 0.033364543914794924, 0.033498847961425784, 0.033462879180908206, 0.03334348678588867, 0.03330867385864258, 0.033500511169433596, 0.03355491256713867, 0.033425216674804685, 0.03336435317993164, 0.033223934173583984, 0.03330889511108399, 0.03326211166381836, 0.03336191940307617, 0.03342335891723633, 0.03356854248046875, 0.03532489776611328, 0.03397359848022461, 0.03332163238525391, 0.03356441497802734, 0.03347798538208008, 0.0334672966003418, 0.033480449676513674, 0.03344547271728516, 0.03321513748168945, 0.03361587142944336, 0.033331199645996096, 0.03361587142944336, 0.03362547302246094, 0.037157440185546876, 0.03342063903808594, 0.03365126419067383, 0.03354336166381836, 0.03373878479003906, 0.03394831848144531, 0.0378309440612793, 0.03430569458007812, 0.03421820831298828, 0.03462908935546875, 0.03406915283203125, 0.03359436798095703, 0.03363849639892578, 0.03364361572265625, 0.033973888397216795, 0.034140350341796875, 0.034045951843261715, 0.03380633544921875, 0.03440639877319336, 0.033533119201660154, 0.03370064163208008, 0.033656734466552735, 0.033584671020507814, 0.03417747116088867, 0.034688320159912106, 0.034143070220947265, 0.03384524917602539, 0.033503009796142576, 0.03350889587402344, 0.03351587295532227, 0.03362627029418945, 0.03337235260009765, 0.03372639846801758, 0.03336198425292969, 0.03382259368896484, 0.03361804962158203, 0.03357299041748047, 0.033372032165527345, 0.03464601516723633, 0.033498462677001954, 0.03359417724609375, 0.033867008209228514, 0.03352579116821289, 0.0334378547668457, 0.033745086669921875, 0.0355002555847168, 0.03367942428588867, 0.03346015930175781, 0.03379609680175781, 0.03493798446655273, 0.03535238265991211, 0.03461542510986328, 0.034722049713134764, 0.034531326293945314, 0.035940128326416014, 0.03454521560668945, 0.034536094665527345, 0.03452108764648437, 0.03430604934692383, 0.03421184158325195, 0.03421593475341797, 0.03409907150268555, 0.03390476989746094, 0.03423385620117188, 0.03418281555175781, 0.0343724479675293, 0.03405590438842773, 0.03394384002685547, 0.033818496704101565, 0.03385152053833008, 0.033742847442626955, 0.033615200042724606, 0.033669696807861325, 0.03353571319580078, 0.03399103927612305, 0.03382681655883789, 0.033688671112060545, 0.033643199920654294, 0.03345363235473633, 0.03342403030395508, 0.03346799850463867, 0.03359580612182617, 0.033527969360351566, 0.03344163131713867, 0.03361324691772461, 0.03358723068237305, 0.03348534393310547, 0.03341721725463867, 0.03341696166992188, 0.03342361450195312, 0.03343360137939453, 0.033576961517333984, 0.03361929702758789, 0.0338704948425293, 0.033406688690185544, 0.03354447937011719, 0.033808383941650394, 0.0335294075012207, 0.03382931137084961, 0.03359539031982422, 0.033439456939697264, 0.03348019027709961, 0.03346249771118164, 0.0335241584777832, 0.03338252639770508, 0.0333675537109375, 0.03339519882202149, 0.03337625503540039, 0.033621726989746095, 0.03382710266113281, 0.03395139312744141, 0.034310497283935544, 0.034370815277099606, 0.035200927734375, 0.034167648315429684, 0.03440214538574219, 0.03393552017211914, 0.033744384765625, 0.03366390228271485, 0.03356835174560547, 0.033879615783691405, 0.033793758392333985, 0.033534687042236326, 0.033476608276367184, 0.03380428695678711, 0.03364659118652344, 0.03426303863525391, 0.03371408081054687, 0.033693790435791016, 0.03501670455932617, 0.03383295822143555, 0.03359280014038086, 0.0335590705871582, 0.033716129302978515, 0.03426108932495117, 0.03395580673217773, 0.03371820831298828, 0.03434710311889649, 0.035796798706054685, 0.033819999694824215, 0.03363516616821289, 0.033546241760253906, 0.03364803314208984, 0.033743038177490234, 0.03375718307495117, 0.03385795211791992, 0.0335843505859375, 0.03339753723144531, 0.03366899108886719, 0.033659008026123045, 0.03422822570800781, 0.033726463317871096, 0.03361177444458008, 0.03351968002319336, 0.033587039947509764, 0.03350742340087891, 0.03378176116943359, 0.033964031219482424, 0.03376451110839844, 0.03370684814453125, 0.03419558334350586, 0.03351055908203125, 0.037013374328613284, 0.033812511444091795, 0.03367814254760742, 0.033642494201660156, 0.03361574554443359, 0.03371446228027344, 0.03426816177368164, 0.03370070266723633, 0.03353148651123047, 0.03346883010864258, 0.033527774810791014, 0.03359052658081055, 0.03331951904296875, 0.03335596847534179, 0.03531779098510742, 0.034265056610107425, 0.034236415863037106, 0.0341401596069336, 0.033976318359375, 0.03384899139404297, 0.03372067260742188, 0.03388415908813477, 0.0335912971496582, 0.034086910247802735, 0.03349881744384765, 0.03358547210693359, 0.03350502395629883, 0.033360126495361325, 0.03336806488037109, 0.03455292892456055, 0.033549217224121096, 0.03352576065063476, 0.03344384002685547, 0.033288158416748044, 0.03356243133544922, 0.03329158401489258, 0.033391521453857424, 0.033304576873779294, 0.03337798309326172, 0.033242496490478515, 0.03350214385986328, 0.03357049560546875, 0.03400492858886719, 0.03366563034057617, 0.03351472091674805, 0.03350175857543945, 0.03333859252929688, 0.033408798217773435, 0.03334652709960938, 0.03337830352783203, 0.03473392105102539, 0.03347475051879883, 0.0335687370300293, 0.03403369522094726, 0.03417702484130859, 0.034205440521240235, 0.0335546875, 0.03395708847045899, 0.03379894256591797, 0.03371212768554688, 0.0336445426940918, 0.03366092681884766, 0.033557952880859374, 0.03378438568115234, 0.033538047790527346, 0.033630176544189455, 0.033427486419677736, 0.03349094390869141, 0.03359743881225586, 0.033694847106933594, 0.03433110427856445, 0.03344015884399414, 0.033498977661132814, 0.0333067855834961, 0.03361382293701172, 0.03342089462280273, 0.03371987152099609]",tokens/s,29.560661044831235,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1028.931584,965.67296,0.0,570.425344,525.840896,s,1,8.756587890625,8.756587890625,0.0,8.756587890625,8.756587890625,8.756587890625,8.756587890625,[8.756587890625],,kWh,3.2744425933303015e-05,3.60483427928465e-06,1.2091120784013532e-05,4.8440380996601195e-05,,MB,1234.219008,1005.518848,0.0,597.68832,584.940544,s,10,1.2835491333007811,0.12835491333007812,0.0008062953252037128,0.1281603240966797,0.12885750885009767,0.1297009635925293,0.13037572738647463,"[0.13054441833496094, 0.12826966857910158, 0.12867007446289064, 0.1277258529663086, 0.12850579833984374, 0.12770556640625, 0.12805097961425782, 0.12850090026855468, 0.12774406433105467, 0.12783180999755858]",tokens/s,1994.4698131007199,kWh,3.962032842454707e-06,4.369392790356769e-07,2.6291199711619972e-06,7.028092092652381e-06,tokens/kWh,36425248.36401032,MB,1245.32736,1020.198912,0.0,612.368384,597.290496,s,10,10.94657263183594,1.0946572631835938,0.017085611733700476,1.0881235961914062,1.1167506591796874,1.118696008300781,1.1202522875976562,"[1.0778565673828124, 1.084336181640625, 1.0810809326171875, 1.069919189453125, 1.0909920654296874, 1.120641357421875, 1.107659912109375, 1.085255126953125, 1.112512939453125, 1.116318359375]",tokens/s,57.55226052835659,kWh,3.2114559489632315e-05,3.541848400897781e-06,1.332450390283957e-05,4.8980911793369663e-05,tokens/kWh,1286215.3376354263,,s,630,10.940668748855604,0.01736614087119935,0.0004322146221876801,0.017292176246643066,0.017899991035461426,0.018038656044006347,0.018869532661437993,"[0.018772064208984376, 0.017807743072509766, 0.017474048614501952, 0.017560800552368163, 0.01753718376159668, 0.01734662437438965, 0.01703318405151367, 0.01691913604736328, 0.016983264923095702, 0.016830432891845704, 0.01706505584716797, 0.01690595245361328, 0.0170732479095459, 0.016969728469848632, 0.01701718330383301, 0.016942720413208007, 0.01697273635864258, 0.017477184295654296, 0.016902591705322264, 0.016916479110717773, 0.016887519836425783, 0.016920671463012696, 0.016910528182983397, 0.016906240463256835, 0.017026399612426756, 0.016926816940307617, 0.017056447982788086, 0.01707379150390625, 0.01711948776245117, 0.017551616668701173, 0.01711692810058594, 0.017020896911621095, 0.01710492706298828, 0.016979455947875977, 0.01698975944519043, 0.017187776565551757, 0.017101055145263673, 0.016858112335205077, 0.016898815155029296, 0.017084384918212892, 0.016897727966308593, 0.017330528259277344, 0.017354751586914064, 0.016965408325195313, 0.016926464080810548, 0.017324512481689452, 0.017031295776367188, 0.01694643211364746, 0.016988800048828124, 0.017163360595703125, 0.01715702438354492, 0.017075872421264647, 0.017508703231811522, 0.017302719116210938, 0.016974655151367188, 0.017153568267822265, 0.016898368835449217, 0.016957504272460937, 0.016775264739990234, 0.016887359619140625, 0.016833984375, 0.016809024810791016, 0.016869312286376954, 0.017879072189331054, 0.01745792007446289, 0.01732211112976074, 0.01715622329711914, 0.017129344940185545, 0.017690624237060547, 0.017164287567138673, 0.01714566421508789, 0.01712761688232422, 0.017008352279663085, 0.017328384399414063, 0.01723776054382324, 0.017232160568237304, 0.017143808364868163, 0.017442592620849608, 0.017550655364990234, 0.01750912094116211, 0.017464736938476562, 0.017349376678466796, 0.01721507263183594, 0.01713203239440918, 0.01712019157409668, 0.017122079849243164, 0.017231136322021483, 0.017343391418457033, 0.017139711380004884, 0.017286527633666993, 0.01752239990234375, 0.017259424209594726, 0.017510400772094727, 0.017201152801513672, 0.0172258243560791, 0.018036895751953125, 0.019379711151123045, 0.017354848861694337, 0.017074335098266603, 0.01702911949157715, 0.016959487915039064, 0.01699430465698242, 0.01702911949157715, 0.017099775314331055, 0.01692527961730957, 0.0169946231842041, 0.016945247650146485, 0.017070207595825195, 0.01708457565307617, 0.017002208709716797, 0.01676288032531738, 0.016788639068603516, 0.016802047729492186, 0.016798240661621094, 0.01682975959777832, 0.01676691246032715, 0.01677395248413086, 0.017487871170043946, 0.017001663208007813, 0.016910688400268555, 0.017002975463867188, 0.016985183715820314, 0.01712748718261719, 0.016986976623535155, 0.01702092742919922, 0.017102848052978514, 0.01782595252990723, 0.017573728561401367, 0.017944576263427735, 0.017744064331054688, 0.017585248947143556, 0.01753766441345215, 0.017342016220092772, 0.017195199966430662, 0.017181024551391602, 0.017506303787231444, 0.017722944259643554, 0.0174268798828125, 0.01730352020263672, 0.017301536560058593, 0.01757209587097168, 0.017348352432250976, 0.017290592193603516, 0.017869024276733397, 0.01711532783508301, 0.017000703811645507, 0.016893951416015626, 0.01678335952758789, 0.0168407039642334, 0.016704864501953125, 0.016765600204467775, 0.01682636833190918, 0.016855039596557618, 0.01680384063720703, 0.016887231826782225, 0.016863807678222657, 0.0168853759765625, 0.01698649597167969, 0.016959455490112303, 0.01697372817993164, 0.017092735290527343, 0.017160192489624023, 0.016990272521972657, 0.016919872283935548, 0.01695414352416992, 0.017079967498779297, 0.016949440002441408, 0.0169881591796875, 0.017074176788330078, 0.017092607498168946, 0.01692982482910156, 0.016859935760498046, 0.01684217643737793, 0.01679020881652832, 0.016826271057128906, 0.016758975982666017, 0.016852960586547852, 0.01723910331726074, 0.016927679061889647, 0.016893951416015626, 0.016863231658935548, 0.016969728469848632, 0.016865280151367186, 0.01696767997741699, 0.01768387222290039, 0.01769059181213379, 0.01784281539916992, 0.017375423431396485, 0.01763270378112793, 0.01795359992980957, 0.01754710388183594, 0.017178815841674806, 0.016924671173095703, 0.01700364875793457, 0.01722867202758789, 0.016946687698364257, 0.01665184020996094, 0.01674950408935547, 0.016928319931030274, 0.0168288631439209, 0.016948448181152344, 0.016691999435424806, 0.016689151763916017, 0.016754016876220704, 0.016611391067504883, 0.016715967178344726, 0.016697759628295897, 0.016760671615600586, 0.01675507164001465, 0.016764320373535157, 0.016648576736450194, 0.016778783798217775, 0.01677359962463379, 0.016740352630615234, 0.016693119049072266, 0.016709760665893556, 0.016742399215698242, 0.016804031372070313, 0.01677622413635254, 0.016820415496826172, 0.01676348876953125, 0.016850719451904295, 0.016762592315673827, 0.017019392013549805, 0.016865440368652344, 0.016957504272460937, 0.0168056640625, 0.01700454330444336, 0.016943103790283204, 0.016990207672119142, 0.01683456039428711, 0.01784832000732422, 0.01676288032531738, 0.016773120880126953, 0.01683430480957031, 0.017082624435424805, 0.0170150089263916, 0.01698588752746582, 0.01712758445739746, 0.01734009552001953, 0.017123487472534178, 0.017184768676757813, 0.017149696350097655, 0.017108415603637694, 0.01700044822692871, 0.017079103469848634, 0.01720854377746582, 0.017179487228393554, 0.017067968368530275, 0.017104896545410156, 0.017121248245239258, 0.018423103332519532, 0.01776710319519043, 0.017552608489990233, 0.01732246398925781, 0.01736355209350586, 0.017381504058837892, 0.01720307159423828, 0.017149791717529297, 0.017040672302246093, 0.017154272079467774, 0.01702895927429199, 0.017271200180053712, 0.017066015243530273, 0.017106847763061525, 0.017041727066040038, 0.016997568130493163, 0.0169418888092041, 0.017010528564453124, 0.017116607666015624, 0.017333120346069337, 0.01710905647277832, 0.01691231918334961, 0.016975616455078123, 0.017249984741210936, 0.01732211112976074, 0.017506399154663087, 0.017563711166381835, 0.01760678482055664, 0.017380767822265625, 0.01730415916442871, 0.017225791931152343, 0.01717219161987305, 0.017070304870605468, 0.016977855682373047, 0.016990272521972657, 0.017041568756103517, 0.016999776840209962, 0.016970239639282226, 0.01699542427062988, 0.017167488098144532, 0.017272607803344726, 0.01744895935058594, 0.01751862335205078, 0.017579999923706055, 0.01752272033691406, 0.01766953659057617, 0.017660480499267578, 0.01768435287475586, 0.017669504165649413, 0.01760233688354492, 0.01765475273132324, 0.01756483268737793, 0.017555456161499023, 0.01745337677001953, 0.017458911895751952, 0.017498943328857423, 0.017551328659057615, 0.017624223709106445, 0.01763603210449219, 0.01730384063720703, 0.017260448455810547, 0.017254400253295898, 0.017268735885620116, 0.017290815353393555, 0.01960140800476074, 0.018509599685668947, 0.01923708724975586, 0.017690624237060547, 0.017824960708618165, 0.017793983459472657, 0.017545087814331055, 0.017520511627197265, 0.017740991592407225, 0.01761401557922363, 0.017678144454956055, 0.017704896926879883, 0.017663999557495116, 0.01890934371948242, 0.01765155220031738, 0.01733568000793457, 0.017320512771606445, 0.0174715518951416, 0.017545087814331055, 0.017518047332763673, 0.01763372802734375, 0.017612127304077147, 0.017748863220214843, 0.017711103439331053, 0.017650976181030273, 0.018195167541503906, 0.019247295379638672, 0.017751615524291994, 0.017660160064697266, 0.01780531120300293, 0.017565248489379882, 0.01741804885864258, 0.01753971290588379, 0.0173253116607666, 0.017305696487426758, 0.017427104949951172, 0.017334175109863282, 0.017255840301513673, 0.01759916877746582, 0.01724415969848633, 0.017137664794921875, 0.018577056884765623, 0.01753327941894531, 0.017342527389526366, 0.017602495193481445, 0.017776416778564452, 0.017725568771362304, 0.017680479049682618, 0.01779088020324707, 0.01783203125, 0.017776607513427734, 0.01793846321105957, 0.01777663993835449, 0.01799577522277832, 0.017954944610595703, 0.017840000152587892, 0.01790755271911621, 0.017875328063964843, 0.017835840225219727, 0.017800447463989257, 0.017814239501953124, 0.01777199935913086, 0.01778332710266113, 0.018114559173583983, 0.018169055938720702, 0.0179965763092041, 0.01805721664428711, 0.017905664443969727, 0.01791958427429199, 0.017937984466552735, 0.01792255973815918, 0.01901398468017578, 0.017779935836791994, 0.017920896530151366, 0.01776639938354492, 0.017704864501953126, 0.0177457275390625, 0.017764575958251955, 0.017880960464477538, 0.01777027130126953, 0.01782329559326172, 0.017690975189208983, 0.017807775497436524, 0.01780531120300293, 0.01794047927856445, 0.017889280319213868, 0.017924095153808595, 0.01807151985168457, 0.01792207908630371, 0.0176312313079834, 0.01764249610900879, 0.017693056106567382, 0.01759846305847168, 0.017477439880371093, 0.01753481674194336, 0.017538015365600585, 0.01751785659790039, 0.017446943283081055, 0.01720185661315918, 0.017182336807250977, 0.017227743148803713, 0.0173121280670166, 0.017338399887084962, 0.01722153663635254, 0.017295455932617186, 0.017275903701782228, 0.01736729621887207, 0.017308416366577147, 0.01722163200378418, 0.01720319938659668, 0.017129791259765624, 0.017179647445678712, 0.017127712249755858, 0.017135040283203125, 0.017099391937255858, 0.0172109432220459, 0.01721014404296875, 0.0171824951171875, 0.017160415649414062, 0.017708032608032227, 0.017711135864257814, 0.017167327880859375, 0.01718396759033203, 0.01714796829223633, 0.017084415435791016, 0.017194911956787108, 0.017934303283691406, 0.017719072341918947, 0.017641056060791017, 0.017436288833618165, 0.01728371238708496, 0.017110687255859375, 0.017174560546875, 0.017179136276245118, 0.017282976150512695, 0.017090272903442384, 0.01749788856506348, 0.017620832443237304, 0.01710995292663574, 0.01706825637817383, 0.01699523162841797, 0.016999263763427735, 0.017039392471313478, 0.01715814399719238, 0.0171909122467041, 0.017029024124145507, 0.017931552886962892, 0.01716307258605957, 0.017102848052978514, 0.016963584899902344, 0.017102848052978514, 0.017104671478271483, 0.016972000122070313, 0.01710201644897461, 0.01694803237915039, 0.016917823791503906, 0.017154943466186522, 0.016873472213745116, 0.017004352569580078, 0.016854976654052733, 0.01700592041015625, 0.01696339225769043, 0.017245088577270508, 0.017124383926391602, 0.017218528747558595, 0.017360895156860352, 0.017328128814697266, 0.01717987251281738, 0.017406911849975587, 0.01714364814758301, 0.017268863677978516, 0.017542015075683592, 0.01713808059692383, 0.01712393569946289, 0.017196319580078126, 0.017203424453735353, 0.017281536102294923, 0.01728102493286133, 0.017294399261474608, 0.017445152282714843, 0.01739228820800781, 0.017294784545898438, 0.017304128646850585, 0.017297088623046877, 0.017326400756835936, 0.017289567947387695, 0.017267744064331056, 0.01708095932006836, 0.016977567672729493, 0.018184192657470705, 0.017681888580322266, 0.01768294334411621, 0.01757801628112793, 0.01729100799560547, 0.017215744018554687, 0.0174583683013916, 0.017213823318481446, 0.017293344497680663, 0.01733468818664551, 0.01723187255859375, 0.01778179168701172, 0.01748908805847168, 0.01763100814819336, 0.01762940788269043, 0.01784566307067871, 0.017927648544311524, 0.017695680618286132, 0.01788105583190918, 0.01780531120300293, 0.017735679626464843, 0.017711103439331053, 0.017809600830078126, 0.017871904373168945, 0.017750944137573242, 0.01789529609680176, 0.018027807235717775, 0.017879776000976563, 0.017961984634399415, 0.017937408447265626, 0.01782579231262207, 0.01784560012817383, 0.01783875274658203, 0.017571584701538086, 0.017464607238769532, 0.01746019172668457, 0.017477632522583008, 0.017707008361816406, 0.017743871688842772, 0.01760256004333496, 0.017739776611328126, 0.017623136520385742, 0.01747052764892578, 0.017537471771240234, 0.017604192733764647, 0.017474176406860352, 0.01811631965637207, 0.017688959121704102, 0.01759651184082031, 0.017383424758911133, 0.017262592315673828, 0.01737673568725586, 0.017537343978881837, 0.017514720916748047, 0.01731724739074707, 0.017656448364257813, 0.01764761543273926, 0.017872831344604493, 0.01801628875732422, 0.017947872161865233, 0.01758195114135742, 0.0175216007232666, 0.01746112060546875, 0.01786591911315918, 0.01752556800842285, 0.017302848815917968, 0.017377983093261717, 0.017204416275024413, 0.017103328704833984, 0.017148256301879883, 0.01730544090270996, 0.01722368049621582, 0.017295520782470705, 0.01747884750366211, 0.017360992431640625, 0.017152063369750975, 0.017457311630249023, 0.01727743911743164, 0.017260032653808592, 0.01744060707092285, 0.01750886344909668, 0.017462656021118163, 0.01746614456176758, 0.01734623908996582, 0.01717030334472656, 0.017383264541625976, 0.017097312927246092, 0.01720921516418457, 0.017190528869628907, 0.017250240325927733, 0.017579872131347655, 0.017611743927001954, 0.017573888778686524, 0.017759775161743162, 0.017616159439086915, 0.017879999160766602, 0.01814463996887207, 0.017924928665161134, 0.01804284858703613, 0.01831065559387207, 0.017912160873413085, 0.017930240631103517, 0.019090431213378906, 0.018377216339111328, 0.018013952255249023, 0.018182912826538087, 0.01789936065673828, 0.01803455924987793, 0.018040096282958985, 0.017884159088134767, 0.017834175109863282, 0.017833791732788085, 0.017831680297851562, 0.018064895629882814, 0.018292543411254882, 0.018266431808471678, 0.018264703750610352, 0.018222240447998046, 0.018083999633789063, 0.01801900863647461, 0.018149375915527344, 0.017991680145263672, 0.018231296539306642, 0.017768447875976562, 0.01770086479187012, 0.017483903884887696]",tokens/s,57.58331729638546,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1514.401792,1823.408128,0.0,1428.160512,1322.516992,s,1,8.8410439453125,8.8410439453125,0.0,8.8410439453125,8.8410439453125,8.8410439453125,8.8410439453125,[8.8410439453125],,kWh,4.956789914999717e-05,5.458563006720272e-06,1.883390395601703e-05,7.386036611273448e-05,,MB,1584.095232,1842.282496,0.0,1434.451968,1322.072064,s,10,5.343417846679688,0.5343417846679688,0.002258050791299455,0.5337992248535156,0.5352724182128906,0.5380207305908203,0.540219380493164,"[0.54076904296875, 0.5322550659179688, 0.5334747924804687, 0.5346616821289063, 0.5326629028320312, 0.5337119140625, 0.5343487548828125, 0.5333185424804687, 0.53432861328125, 0.5338865356445313]",tokens/s,479.0941067037724,kWh,1.5952725873246397e-05,1.759311121475994e-06,1.056739149484027e-05,2.827942848956266e-05,tokens/kWh,9052516.747093534,MB,1588.801536,1842.282496,0.0,1434.451968,1374.923264,s,10,11.133669555664062,1.113366955566406,0.009659703262118879,1.115445068359375,1.123347900390625,1.1244218139648436,1.1252809448242187,"[1.0983587646484374, 1.0961551513671874, 1.107653564453125, 1.1213216552734375, 1.110700439453125, 1.1127877197265625, 1.1231092529296876, 1.1254957275390625, 1.1181024169921876, 1.11998486328125]",tokens/s,56.58511749879432,kWh,3.164441814133704e-05,3.4902849014094375e-06,1.7112519537961604e-05,5.224722258070807e-05,tokens/kWh,1205805.7230254058,,s,630,11.13072278022767,0.017667813936869302,0.0004447551390950777,0.0175817928314209,0.01808372802734375,0.0183081262588501,0.018845265083312992,"[0.01878835105895996, 0.018011871337890624, 0.017658143997192382, 0.017508352279663086, 0.01740390396118164, 0.017478687286376953, 0.017476512908935548, 0.01736300849914551, 0.01733827209472656, 0.017284543991088867, 0.01727964782714844, 0.01719705581665039, 0.017356800079345702, 0.01731551933288574, 0.017320255279541015, 0.01738675117492676, 0.01744883155822754, 0.017396608352661134, 0.017487295150756837, 0.017453632354736327, 0.01738924789428711, 0.01744095993041992, 0.017496192932128906, 0.017438720703125, 0.01759846305847168, 0.017569696426391602, 0.01753878402709961, 0.01751641654968262, 0.017850879669189454, 0.017489919662475584, 0.017528480529785156, 0.017543519973754883, 0.01735820770263672, 0.017386112213134765, 0.017367040634155274, 0.017360895156860352, 0.01726464080810547, 0.017388576507568358, 0.017408992767333983, 0.01743052864074707, 0.01743667221069336, 0.018354175567626953, 0.017368864059448243, 0.017237279891967775, 0.017392000198364257, 0.017271360397338866, 0.017372991561889647, 0.01723516845703125, 0.01722835159301758, 0.01728553581237793, 0.017372831344604493, 0.01723129653930664, 0.017238784790039062, 0.017256128311157228, 0.017270463943481446, 0.017281824111938477, 0.017333856582641603, 0.017297536849975585, 0.017273120880126953, 0.017268287658691407, 0.017277376174926758, 0.017251840591430666, 0.01727743911743164, 0.018479103088378905, 0.01768409538269043, 0.017545600891113282, 0.01746028709411621, 0.01734342384338379, 0.01726438331604004, 0.017314048767089845, 0.01726246452331543, 0.01737238311767578, 0.01722460746765137, 0.01724006462097168, 0.017205120086669922, 0.01735487937927246, 0.017204927444458007, 0.017463167190551757, 0.01727123260498047, 0.017340192794799803, 0.017248031616210937, 0.017273279190063478, 0.017263999938964845, 0.017384063720703124, 0.017928064346313475, 0.017496192932128906, 0.017375232696533204, 0.017340127944946288, 0.017223968505859374, 0.017357919692993166, 0.017245088577270508, 0.017311744689941407, 0.017242111206054688, 0.01739753532409668, 0.017287391662597656, 0.017348608016967772, 0.01738956832885742, 0.017291263580322267, 0.017354751586914064, 0.017391616821289063, 0.017348512649536133, 0.0179117431640625, 0.017449119567871093, 0.01741209602355957, 0.017257728576660157, 0.017251007080078123, 0.017319616317749024, 0.01732851219177246, 0.017274303436279295, 0.017182464599609374, 0.017234432220458985, 0.01723814392089844, 0.017471328735351562, 0.01735043144226074, 0.017407840728759765, 0.01730384063720703, 0.017289663314819338, 0.017358848571777344, 0.017246208190917968, 0.0172728328704834, 0.01739366340637207, 0.017325567245483398, 0.017320255279541015, 0.017379295349121093, 0.017854175567626952, 0.01852262306213379, 0.01846067237854004, 0.017780096054077148, 0.017719327926635744, 0.017445472717285155, 0.01813020706176758, 0.017293472290039063, 0.017359424591064453, 0.017294719696044923, 0.017451648712158204, 0.01764556884765625, 0.01762719917297363, 0.017901472091674805, 0.018013216018676757, 0.018297855377197265, 0.01790764808654785, 0.017739839553833008, 0.01785651206970215, 0.017682432174682617, 0.01775119972229004, 0.01767100715637207, 0.017647199630737305, 0.017738048553466796, 0.017746015548706053, 0.017686527252197267, 0.01781350326538086, 0.017849760055541994, 0.01775062370300293, 0.01756889533996582, 0.017517152786254882, 0.01746767997741699, 0.01745289611816406, 0.01735491180419922, 0.017451007843017577, 0.01743833541870117, 0.017527135848999023, 0.01754115104675293, 0.01743657684326172, 0.01744495964050293, 0.01739776039123535, 0.01731292724609375, 0.017490943908691405, 0.018607967376708983, 0.018112512588500978, 0.01761075210571289, 0.01758992004394531, 0.017305919647216797, 0.017324064254760744, 0.017305599212646485, 0.017260032653808592, 0.01715782356262207, 0.017199935913085936, 0.017257503509521484, 0.01747862434387207, 0.017529983520507813, 0.017502208709716797, 0.01738630485534668, 0.017372991561889647, 0.017401504516601562, 0.017349216461181642, 0.01726166343688965, 0.017301631927490235, 0.017257247924804688, 0.017192960739135742, 0.018319616317749022, 0.01763711929321289, 0.017663551330566407, 0.017481279373168946, 0.017628000259399413, 0.017666080474853515, 0.017532928466796875, 0.0175861759185791, 0.017669567108154295, 0.01751897621154785, 0.017483232498168945, 0.0175416316986084, 0.01743840026855469, 0.017336864471435547, 0.017366111755371092, 0.01745193672180176, 0.017456832885742186, 0.01747283172607422, 0.017497087478637697, 0.017481695175170897, 0.01756982421875, 0.017584127426147463, 0.017555456161499023, 0.017497152328491212, 0.01817081642150879, 0.01864089584350586, 0.020850143432617187, 0.017783136367797853, 0.01768467140197754, 0.017584127426147463, 0.018764863967895506, 0.017503040313720703, 0.01766531181335449, 0.018856800079345704, 0.017934335708618163, 0.017881088256835938, 0.017524736404418945, 0.017452831268310546, 0.0174553279876709, 0.017356800079345702, 0.01728233528137207, 0.017386207580566405, 0.017666048049926757, 0.017600351333618165, 0.017668256759643553, 0.017541120529174805, 0.017459199905395507, 0.02064793586730957, 0.017683616638183595, 0.017791551589965822, 0.019484960556030273, 0.01758950424194336, 0.017545984268188475, 0.01757151985168457, 0.01751481628417969, 0.017497343063354494, 0.01753487968444824, 0.01750307273864746, 0.018792448043823243, 0.017485824584960938, 0.017532928466796875, 0.017721343994140625, 0.018030303955078125, 0.018448383331298827, 0.017710271835327147, 0.017660736083984375, 0.017528736114501953, 0.017354848861694337, 0.017262592315673828, 0.017338016510009765, 0.017297119140625, 0.017534751892089844, 0.017613279342651368, 0.017352575302124022, 0.01740595245361328, 0.017347072601318358, 0.017285120010375975, 0.017303455352783204, 0.01724835205078125, 0.01724412727355957, 0.01740598487854004, 0.017554943084716796, 0.017432512283325194, 0.017281600952148438, 0.017285120010375975, 0.01724608039855957, 0.017298559188842773, 0.017337343215942384, 0.017334272384643554, 0.017321887969970702, 0.017345792770385744, 0.017347423553466797, 0.01737222480773926, 0.01734137535095215, 0.01724006462097168, 0.01739776039123535, 0.017500160217285156, 0.01752604866027832, 0.017347295761108397, 0.017377279281616212, 0.017552543640136718, 0.01761142349243164, 0.018053119659423827, 0.017789119720458983, 0.01780531120300293, 0.018266111373901366, 0.018056575775146484, 0.017899871826171875, 0.01824083137512207, 0.017849311828613282, 0.017983488082885742, 0.01794767951965332, 0.01800422477722168, 0.018086624145507813, 0.018144895553588867, 0.0179116153717041, 0.018297407150268556, 0.018163711547851562, 0.017925792694091797, 0.018288991928100587, 0.017884544372558594, 0.017617536544799806, 0.017489919662475584, 0.01762099266052246, 0.017763711929321288, 0.01792473602294922, 0.018488128662109374, 0.018112512588500978, 0.017947872161865233, 0.017898271560668946, 0.017725439071655275, 0.017702911376953127, 0.017612127304077147, 0.01764009666442871, 0.0177040958404541, 0.01771811294555664, 0.017843551635742187, 0.017795167922973632, 0.017840703964233397, 0.017887231826782226, 0.01778611183166504, 0.017732351303100587, 0.017545055389404297, 0.01807375907897949, 0.01761017608642578, 0.018229408264160155, 0.017787071228027345, 0.017682111740112305, 0.01746384048461914, 0.017559551239013673, 0.017468767166137697, 0.017369760513305663, 0.01756979179382324, 0.017386911392211914, 0.018088544845581055, 0.01760051155090332, 0.017864416122436524, 0.017747936248779298, 0.017686368942260743, 0.017504735946655272, 0.01747148895263672, 0.017350656509399414, 0.01737932777404785, 0.017550399780273437, 0.017398719787597657, 0.01740390396118164, 0.017383424758911133, 0.017567743301391603, 0.018093088150024413, 0.01761769676208496, 0.01778054428100586, 0.017404287338256837, 0.017557504653930665, 0.0175817928314209, 0.01752092742919922, 0.017563648223876953, 0.01745715141296387, 0.017508352279663086, 0.01750592041015625, 0.017463264465332032, 0.01768396759033203, 0.017499040603637696, 0.017467391967773437, 0.01754515266418457, 0.017553184509277345, 0.01748524856567383, 0.017666112899780272, 0.017656608581542967, 0.017680351257324218, 0.018354976654052734, 0.01796665573120117, 0.01778118324279785, 0.01756979179382324, 0.01752057647705078, 0.017440832138061524, 0.01745305633544922, 0.017625087738037108, 0.01765376091003418, 0.018054431915283203, 0.017691360473632813, 0.017704959869384765, 0.017547264099121093, 0.01755292892456055, 0.017459680557250976, 0.01759231948852539, 0.017672191619873046, 0.017604608535766602, 0.018309120178222657, 0.017714847564697267, 0.01765782356262207, 0.017674623489379884, 0.01762073516845703, 0.017657472610473634, 0.017641311645507814, 0.017652511596679688, 0.01766508865356445, 0.01765839958190918, 0.017953184127807616, 0.01773155212402344, 0.017827871322631837, 0.017678016662597655, 0.017645503997802733, 0.017580415725708006, 0.017601791381835936, 0.0178920955657959, 0.017743871688842772, 0.017808639526367187, 0.017773311614990236, 0.01821286392211914, 0.017768096923828126, 0.01769487953186035, 0.01770515251159668, 0.018274303436279296, 0.0178606071472168, 0.01769267272949219, 0.017743871688842772, 0.017774463653564453, 0.017909151077270508, 0.017969535827636718, 0.01808345603942871, 0.01809891128540039, 0.01825382423400879, 0.0180633602142334, 0.018050912857055665, 0.01811471939086914, 0.01815724754333496, 0.018306911468505858, 0.018057056427001953, 0.017998464584350587, 0.018089216232299806, 0.018002399444580076, 0.0180861759185791, 0.018462751388549806, 0.017849727630615233, 0.018198495864868165, 0.017670080184936522, 0.017670175552368165, 0.017594783782958985, 0.017602752685546875, 0.0175568962097168, 0.01762067222595215, 0.017494655609130858, 0.017457408905029295, 0.01767843246459961, 0.017706911087036134, 0.017698240280151368, 0.017660575866699217, 0.01763532829284668, 0.017604576110839844, 0.01756163215637207, 0.017677503585815428, 0.01777542304992676, 0.017929311752319335, 0.01795574378967285, 0.017917503356933595, 0.017956544876098633, 0.018002687454223634, 0.018032352447509767, 0.018022207260131835, 0.01821232032775879, 0.01860710334777832, 0.018094079971313477, 0.01802444839477539, 0.017911808013916015, 0.017995519638061522, 0.017916160583496092, 0.017958911895751953, 0.01789030456542969, 0.017856576919555663, 0.017808319091796875, 0.01784828758239746, 0.017761760711669922, 0.0177956485748291, 0.017921760559082033, 0.0180164794921875, 0.018110111236572267, 0.01795465660095215, 0.01784454345703125, 0.017815584182739257, 0.017692256927490234, 0.017793664932250975, 0.017946624755859376, 0.01790771293640137, 0.017797056198120116, 0.01765996742248535, 0.017721343994140625, 0.017756160736083985, 0.017762304306030274, 0.0180380802154541, 0.01780191993713379, 0.017746015548706053, 0.017812959671020506, 0.018376895904541016, 0.01800217628479004, 0.017993728637695314, 0.01851913642883301, 0.017879968643188478, 0.017750015258789064, 0.017699871063232422, 0.017827871322631837, 0.017736448287963866, 0.017768640518188477, 0.01783193588256836, 0.017923360824584962, 0.01783260726928711, 0.01781907272338867, 0.017756799697875976, 0.018020351409912108, 0.01787676811218262, 0.017932191848754882, 0.0178176326751709, 0.019167520523071288, 0.01787494468688965, 0.018378528594970703, 0.018028799057006835, 0.017749439239501952, 0.01768707275390625, 0.017761503219604492, 0.01768732833862305, 0.017618335723876954, 0.017492576599121092, 0.017451007843017577, 0.01741983985900879, 0.01737740707397461, 0.017360416412353516, 0.017488672256469728, 0.0175817928314209, 0.017584064483642577, 0.017512800216674805, 0.017532928466796875, 0.017618175506591796, 0.017543935775756837, 0.01742438316345215, 0.017477632522583008, 0.017518592834472657, 0.01762409591674805, 0.01743734359741211, 0.01747385597229004, 0.01749740791320801, 0.017547967910766602, 0.01744486427307129, 0.01756368064880371, 0.01817385673522949, 0.017845504760742186, 0.017701696395874024, 0.018321407318115233, 0.01776639938354492, 0.017822912216186523, 0.01771718406677246, 0.017783679962158204, 0.01779916763305664, 0.01782374382019043, 0.01785241508483887, 0.017707008361816406, 0.017736991882324218, 0.017598688125610353, 0.017643552780151367, 0.01762761688232422, 0.018331647872924805, 0.017616159439086915, 0.017582815170288087, 0.017725439071655275, 0.017768447875976562, 0.017489152908325194, 0.017525503158569336, 0.01740764808654785, 0.017512800216674805, 0.01742972755432129, 0.01750614356994629, 0.017427391052246093, 0.01866035270690918, 0.018094783782958986, 0.017454496383666994, 0.017554304122924805, 0.01755958366394043, 0.01742582321166992, 0.017400415420532226, 0.017375232696533204, 0.017354751586914064, 0.017303071975708007, 0.017222112655639648, 0.017169792175292967, 0.017320735931396485, 0.017275999069213867, 0.017332992553710937, 0.017327199935913085, 0.017292192459106445, 0.017926111221313476, 0.022798368453979492, 0.0210882568359375, 0.017911359786987303, 0.0176746883392334, 0.017688255310058593, 0.01761497688293457, 0.017561792373657226, 0.017772544860839845, 0.017733631134033204, 0.017687936782836915, 0.017433216094970703, 0.018817024230957033, 0.017462656021118163, 0.017437311172485353, 0.01744691276550293, 0.017559551239013673, 0.017459199905395507, 0.017408000946044923, 0.017384864807128905, 0.017406560897827147, 0.01743052864074707, 0.01741619110107422, 0.017481727600097655, 0.017356800079345702, 0.018652671813964843, 0.018743167877197264, 0.017645856857299805, 0.017674591064453123, 0.017723392486572266, 0.01787494468688965, 0.01783193588256836, 0.01817804718017578, 0.017898687362670897]",tokens/s,56.600097984572606,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,6625.062912,9171.828736,0.0,8776.58112,8188.314112,s,1,13.9212919921875,13.9212919921875,0.0,13.9212919921875,13.9212919921875,13.9212919921875,13.9212919921875,[13.9212919921875],,kWh,0.0002018349085958486,2.224601848706114e-05,8.983729409201002e-05,0.00031391822117491974,,MB,1475.559424,9184.411648,0.0,8776.58112,7654.291456,s,10,51.66322021484375,5.166322021484374,0.003286428396186504,5.1669511718750005,5.17017724609375,5.1707001953125005,5.1711185546875,"[5.16087841796875, 5.16267724609375, 5.16403466796875, 5.1631455078125, 5.16671630859375, 5.16718603515625, 5.168517578125, 5.1687802734375, 5.17006103515625, 5.17122314453125]",tokens/s,49.551692468145205,kWh,0.00015074363526833317,1.6627404322531275e-05,0.0001002143301713998,0.00026758536976226424,tokens/kWh,956704.0239436213,MB,1479.76192,9184.411648,0.0,8776.58112,7908.775424,s,10,25.089062011718752,2.5089062011718752,0.003301189497220273,2.50890869140625,2.5116713867187497,2.5143839111328123,2.5165539306640623,"[2.506116455078125, 2.50870654296875, 2.511068603515625, 2.509246337890625, 2.504135498046875, 2.506937255859375, 2.509302734375, 2.517096435546875, 2.50911083984375, 2.50734130859375]",tokens/s,25.110544176810425,kWh,7.337681714708651e-05,8.094045056716978e-06,4.901428921140206e-05,0.00013048515141520552,tokens/kWh,482813.556306749,,s,630,25.08557906723024,0.039818379471794015,0.0003776241935428,0.039851808547973634,0.04026335830688477,0.040357406044006346,0.040607508697509766,"[0.040671230316162106, 0.03988601684570312, 0.03957843017578125, 0.03957705688476563, 0.03949622344970703, 0.0393175048828125, 0.03927449417114258, 0.039602176666259765, 0.03947283172607422, 0.03934444808959961, 0.03931039810180664, 0.03926726531982422, 0.03919462585449219, 0.03933161544799805, 0.03931343841552734, 0.039393184661865234, 0.03930755233764648, 0.039337982177734376, 0.03934003067016602, 0.039577598571777346, 0.03958169555664062, 0.039795936584472655, 0.03973795318603516, 0.03953398513793945, 0.03983849716186524, 0.03976806259155274, 0.03961600112915039, 0.03984371185302735, 0.039841793060302735, 0.03979942321777344, 0.03969638442993164, 0.03966886520385742, 0.03958992004394531, 0.03992867279052734, 0.03998044967651367, 0.039817825317382816, 0.03962611389160156, 0.03969292831420899, 0.039787616729736325, 0.03965020751953125, 0.039769569396972654, 0.039693984985351566, 0.040059680938720706, 0.04009379196166992, 0.04002406311035156, 0.03999948883056641, 0.04016707229614258, 0.04007766342163086, 0.04003184127807617, 0.04007513427734375, 0.04004240036010742, 0.04011894226074219, 0.04015472030639648, 0.04020848083496094, 0.040048736572265625, 0.039960769653320315, 0.03995443344116211, 0.04005177688598633, 0.04013561630249023, 0.040052734375, 0.04020633697509766, 0.04041625595092774, 0.040093952178955075, 0.0407894401550293, 0.0397993278503418, 0.03931545639038086, 0.03915974426269531, 0.03899193572998047, 0.0389815673828125, 0.03900831985473633, 0.039065601348876954, 0.03930112075805664, 0.03945062255859375, 0.03938703918457031, 0.0393891830444336, 0.03932918548583984, 0.03942575836181641, 0.03977238464355469, 0.039844608306884764, 0.039602176666259765, 0.03974515151977539, 0.039944576263427733, 0.0397844467163086, 0.03951385498046875, 0.039507518768310546, 0.039856830596923826, 0.03986431884765625, 0.039739391326904294, 0.03983564758300781, 0.04012851333618164, 0.04006256103515625, 0.039860641479492184, 0.03990249633789063, 0.03976470565795898, 0.03991686248779297, 0.03984764862060547, 0.039752094268798825, 0.03966214370727539, 0.03971072006225586, 0.03991551971435547, 0.04003014373779297, 0.03985987091064453, 0.03964076614379883, 0.03966230392456055, 0.03985391998291016, 0.0400425910949707, 0.04000473785400391, 0.039939006805419924, 0.040191009521484376, 0.04027695846557617, 0.040101886749267575, 0.039901023864746095, 0.040050846099853515, 0.04044134521484375, 0.04020684814453125, 0.040123519897460935, 0.03999337768554687, 0.04030550384521484, 0.04020560073852539, 0.04013312149047851, 0.04006115341186523, 0.04006911849975586, 0.04022886276245117, 0.040081409454345705, 0.04003587341308594, 0.04002819061279297, 0.04058723068237305, 0.03967388916015625, 0.03924582290649414, 0.03915161514282227, 0.03903286361694336, 0.03905503845214844, 0.03923791885375977, 0.03926015853881836, 0.03947689437866211, 0.03952060699462891, 0.039610366821289066, 0.03917824172973633, 0.03915161514282227, 0.03927027130126953, 0.03930928039550781, 0.03920016098022461, 0.039260929107666015, 0.03935574340820312, 0.03950864028930664, 0.03951520156860352, 0.039494590759277345, 0.03984780883789062, 0.03985120010375977, 0.03985049438476562, 0.039839649200439455, 0.0397410888671875, 0.03989184188842773, 0.03975167846679688, 0.039684097290039064, 0.039694271087646484, 0.03968928146362305, 0.040474624633789064, 0.04032614517211914, 0.03995647811889649, 0.04012851333618164, 0.04007673645019531, 0.04006278228759766, 0.03997126388549805, 0.04018207931518555, 0.040101886749267575, 0.040022014617919925, 0.039985088348388674, 0.03983699035644531, 0.03982617568969726, 0.040097793579101565, 0.040263294219970706, 0.040126846313476565, 0.040121631622314455, 0.040051422119140624, 0.03993804931640625, 0.03986191940307617, 0.0402940788269043, 0.04039132690429687, 0.040237056732177735, 0.040196094512939456, 0.04017484664916992, 0.040051456451416015, 0.04042339324951172, 0.04048073577880859, 0.04040915298461914, 0.04025139236450195, 0.04077977752685547, 0.04069375991821289, 0.040468544006347654, 0.03946316909790039, 0.039139934539794925, 0.03903193664550781, 0.03917708969116211, 0.03936822509765625, 0.03922991943359375, 0.03916912078857422, 0.039347103118896484, 0.03926800155639648, 0.03919036865234375, 0.03927248001098633, 0.039428577423095704, 0.03943219375610352, 0.039396991729736326, 0.03934419250488281, 0.03946937561035156, 0.039734432220458984, 0.039741695404052736, 0.039561824798583986, 0.03935372924804687, 0.03933996963500976, 0.03932640075683594, 0.03951375961303711, 0.03963324737548828, 0.039723007202148435, 0.03964518356323242, 0.039569408416748046, 0.03964313507080078, 0.039809024810791016, 0.03983359909057617, 0.0399911994934082, 0.04002825546264648, 0.040005630493164065, 0.03998847961425781, 0.040069889068603516, 0.039962623596191404, 0.039989246368408206, 0.0398106575012207, 0.03980710220336914, 0.03981545639038086, 0.040236095428466796, 0.04026809692382813, 0.04016550445556641, 0.04007372665405273, 0.040111167907714844, 0.04016624069213867, 0.040198078155517576, 0.040183967590332034, 0.04021014404296875, 0.04025724792480469, 0.04029292678833008, 0.040350784301757814, 0.040285118103027345, 0.04028726577758789, 0.040162273406982425, 0.04017356872558594, 0.040513473510742186, 0.040538177490234376, 0.040398750305175785, 0.04030169677734375, 0.04034659194946289, 0.04026572799682617, 0.040588127136230466, 0.03966300964355469, 0.03942015838623047, 0.03930928039550781, 0.03922982406616211, 0.039163902282714845, 0.03908377456665039, 0.03921065521240234, 0.03928115081787109, 0.039395423889160154, 0.03934207916259766, 0.039223297119140625, 0.039200736999511716, 0.039041057586669925, 0.03915161514282227, 0.03952947235107422, 0.03945510482788086, 0.039434879302978516, 0.03944038391113281, 0.03934822463989258, 0.0395365104675293, 0.03951001739501953, 0.03982553482055664, 0.03981107330322266, 0.039806976318359374, 0.039686145782470705, 0.0398900146484375, 0.040237983703613284, 0.04008262252807617, 0.03971974563598633, 0.039608318328857424, 0.03956844711303711, 0.039701438903808596, 0.039825408935546876, 0.03998432159423828, 0.039983936309814457, 0.03994819259643555, 0.03975177764892578, 0.0398131217956543, 0.03967747116088867, 0.039690719604492185, 0.03976732635498047, 0.03960198211669922, 0.039979263305664064, 0.0400022087097168, 0.03997081756591797, 0.03999110412597656, 0.04000787353515625, 0.03988246536254883, 0.03988918304443359, 0.04005068969726563, 0.0402303352355957, 0.040087329864501954, 0.0399529914855957, 0.03978464126586914, 0.04002540969848633, 0.04009632110595703, 0.04011840057373047, 0.03999440002441406, 0.04006111907958984, 0.04035868835449219, 0.04050675201416016, 0.04028889465332031, 0.040471649169921874, 0.03966191864013672, 0.03924028778076172, 0.03903894424438477, 0.038927486419677734, 0.039048065185546876, 0.039221248626708984, 0.03926985549926758, 0.03936924743652344, 0.03928678512573242, 0.03941785430908203, 0.039325374603271485, 0.03931987380981445, 0.03924582290649414, 0.03922739028930664, 0.03953782272338867, 0.03944739151000977, 0.03946223831176758, 0.039611038208007814, 0.03951001739501953, 0.039329792022705076, 0.03950796890258789, 0.0394911994934082, 0.03965990447998047, 0.03970457458496094, 0.03995238494873047, 0.0397817268371582, 0.039629470825195315, 0.039495681762695314, 0.03952640151977539, 0.039939136505126954, 0.03991033554077148, 0.039767200469970704, 0.0398524169921875, 0.039704383850097655, 0.03969247817993164, 0.04002838516235351, 0.04007731246948242, 0.039858272552490234, 0.04023721694946289, 0.0401797103881836, 0.040019359588623044, 0.040034912109375, 0.040005630493164065, 0.03992742538452149, 0.04009811019897461, 0.04029241561889649, 0.04030636978149414, 0.04011449432373047, 0.040134654998779294, 0.04017123031616211, 0.0401124153137207, 0.0401162223815918, 0.04011212921142578, 0.04008345413208008, 0.0405667839050293, 0.040502880096435545, 0.04047423934936523, 0.04016563034057617, 0.040042945861816406, 0.040011680603027344, 0.04018339157104492, 0.04018582534790039, 0.040546302795410154, 0.03985408020019531, 0.03958367919921875, 0.04019574356079102, 0.039250335693359374, 0.03991737747192383, 0.03984812927246094, 0.03968000030517578, 0.039489246368408205, 0.03955331039428711, 0.03920809555053711, 0.039218017578125, 0.03918812942504883, 0.03919286346435547, 0.039407646179199216, 0.03940969467163086, 0.03935232162475586, 0.039554080963134765, 0.03938812637329102, 0.039400577545166016, 0.03957440185546875, 0.039524192810058596, 0.039473312377929684, 0.0397720947265625, 0.039798175811767575, 0.03995238494873047, 0.04012099075317383, 0.039790592193603515, 0.039583518981933595, 0.039628543853759766, 0.039561695098876956, 0.03990937423706055, 0.040032257080078126, 0.04001177597045898, 0.03995852661132813, 0.04042502212524414, 0.040042945861816406, 0.039857887268066404, 0.03988508987426758, 0.03987865447998047, 0.03984384155273438, 0.03985408020019531, 0.039809024810791016, 0.039792640686035156, 0.03976764678955078, 0.03979510498046875, 0.03989503860473633, 0.040202239990234374, 0.04024729537963867, 0.040119873046875, 0.03990367889404297, 0.04016332626342774, 0.04007017517089844, 0.039930721282958985, 0.04001804733276367, 0.04024524688720703, 0.03999334335327148, 0.03998012924194336, 0.04023183822631836, 0.04035583877563476, 0.04017059326171875, 0.04017654418945313, 0.04029644775390625, 0.041139774322509766, 0.04004012680053711, 0.039780288696289065, 0.03935315322875976, 0.0390305290222168, 0.03908633422851562, 0.03926835250854492, 0.03936771011352539, 0.03939632034301758, 0.03954687881469727, 0.03957551956176758, 0.039583263397216795, 0.03950643157958984, 0.03966912078857422, 0.03980505752563476, 0.03986073684692383, 0.03967155075073242, 0.03990758514404297, 0.03966515350341797, 0.03971516799926758, 0.039690399169921876, 0.0399788818359375, 0.03975721740722656, 0.03992035293579101, 0.039875968933105466, 0.0398682861328125, 0.04026393508911133, 0.04022937774658203, 0.039822399139404295, 0.03995743942260742, 0.03987225723266601, 0.0398315200805664, 0.03981955337524414, 0.039922752380371095, 0.03981571197509766, 0.04020681762695313, 0.040040382385253905, 0.04007526397705078, 0.0401797103881836, 0.04033331298828125, 0.04033846282958985, 0.04018684768676758, 0.04016864013671875, 0.04027065658569336, 0.040080894470214845, 0.040290817260742184, 0.04031488037109375, 0.04013391876220703, 0.04028079986572266, 0.0401855697631836, 0.04012675094604492, 0.04028211212158203, 0.04016128158569336, 0.04023462295532226, 0.04014937591552734, 0.04016332626342774, 0.040013824462890625, 0.04007731246948242, 0.04040604782104492, 0.040166366577148435, 0.04000153732299805, 0.03996876907348633, 0.04024483108520508, 0.04061542510986328, 0.03952076721191406, 0.03927040100097656, 0.03989465713500977, 0.03923187255859375, 0.039163902282714845, 0.03902012634277344, 0.03904758453369141, 0.039196670532226564, 0.03918822479248047, 0.039205120086669924, 0.03942195129394531, 0.039396991729736326, 0.03942140960693359, 0.03976204681396484, 0.0397627182006836, 0.03954278564453125, 0.03960566329956055, 0.040185985565185545, 0.040046241760253905, 0.0397217903137207, 0.03947520065307617, 0.039388671875, 0.0397334098815918, 0.03983302307128906, 0.03986115264892578, 0.039792640686035156, 0.03984912109375, 0.0397196159362793, 0.03953065490722656, 0.039362560272216796, 0.039632801055908204, 0.039716064453125, 0.0398889274597168, 0.039887710571289064, 0.039814910888671874, 0.03962905502319336, 0.039782176971435546, 0.04019222259521484, 0.04023849487304688, 0.040039009094238284, 0.04001116943359375, 0.04006947326660156, 0.04032742309570313, 0.04012441635131836, 0.039962623596191404, 0.03984384155273438, 0.04016742324829101, 0.04017334365844726, 0.04012432098388672, 0.04013907241821289, 0.04021440124511719, 0.040402080535888674, 0.04027900695800781, 0.03989904022216797, 0.03986441421508789, 0.040089599609375, 0.04020627212524414, 0.04020844650268555, 0.0402083854675293, 0.04020147323608399, 0.04017023849487305, 0.04048486328125, 0.04073708724975586, 0.0398675537109375, 0.039295841217041015, 0.039229438781738284, 0.039213054656982424, 0.03918643188476562, 0.0393175048828125, 0.039329792022705076, 0.039419902801513675, 0.03922534561157227, 0.03913312149047852, 0.039147552490234376, 0.03956943893432617, 0.0395601921081543, 0.03951308822631836, 0.03950102233886719, 0.03955315017700195, 0.03951001739501953, 0.039440673828125, 0.03932355117797852, 0.03925449752807617, 0.039363838195800784, 0.03939609527587891, 0.039798782348632815, 0.040089599609375, 0.040103008270263675, 0.039842273712158205, 0.03966582489013672, 0.039450912475585936, 0.03949961471557617, 0.039550880432128906, 0.03981881713867187, 0.03986435317993164, 0.039790943145751954, 0.039721057891845706, 0.03976812744140625, 0.039790145874023436, 0.040149089813232425, 0.04009011077880859, 0.039994720458984376, 0.03995212936401367, 0.039981983184814454, 0.03995443344116211, 0.039948287963867186, 0.0402204818725586, 0.04033145523071289, 0.04005683135986328, 0.04019814300537109, 0.04032716751098633, 0.04055654525756836, 0.04022415924072266, 0.040227424621582034, 0.040081409454345705, 0.04009695816040039, 0.04006585693359375, 0.04009267044067383, 0.040010753631591796, 0.04000972747802734, 0.03996876907348633, 0.039927616119384765, 0.04011228942871094, 0.04027395248413086, 0.04027391815185547]",tokens/s,25.114030587517153,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,11116.38016,14586.6752,0.0,14191.427584,13325.783552,s,1,18.065916015625,18.065916015625,0.0,18.065916015625,18.065916015625,18.065916015625,18.065916015625,[18.065916015625],,kWh,0.0003254964083291725,3.589740193542263e-05,0.00014863761890999971,0.0005100314291745948,,MB,2042.126336,14599.258112,0.0,14191.427584,12582.029312,s,10,93.31914257812501,9.331914257812501,0.0027802590925932497,9.33111669921875,9.3355583984375,9.3362474609375,9.336798710937499,"[9.3316044921875, 9.3291533203125, 9.328056640625, 9.3303154296875, 9.33062890625, 9.3295869140625, 9.332990234375, 9.33446484375, 9.3354052734375, 9.3369365234375]",tokens/s,27.432742407130633,kWh,0.000272282078387085,3.003405638682655e-05,0.0001809637003264003,0.0004832798351003118,tokens/kWh,529713.8043155959,MB,2046.562304,14599.258112,0.0,14191.427584,12935.916032,s,10,41.51020703125,4.151020703125,0.0021570119003759226,4.150668701171875,4.153448681640625,4.154599829101562,4.155520747070312,"[4.1493603515625, 4.14903662109375, 4.1501279296875, 4.14786572265625, 4.15031494140625, 4.15220068359375, 4.15133447265625, 4.1557509765625, 4.15319287109375, 4.1510224609375]",tokens/s,15.176990072001304,kWh,0.00012134205874583093,1.3384976105635782e-05,8.099598146339845e-05,0.00021572301631486506,tokens/kWh,292041.16035558505,,s,630,41.505993812561016,0.06588252986120799,0.0005906783672297738,0.06589093017578125,0.06665264968872071,0.06687368431091309,0.06720562927246095,"[0.06626099395751953, 0.06524118041992187, 0.06492034912109375, 0.064827392578125, 0.06484786987304687, 0.06484092712402344, 0.06486438751220704, 0.06557571411132812, 0.06565261077880859, 0.0653803482055664, 0.06544515228271484, 0.06521520233154297, 0.06514073944091797, 0.06492105865478516, 0.06538006591796874, 0.0658372802734375, 0.06559180450439453, 0.06537757110595703, 0.06551020812988281, 0.06589641571044921, 0.06540496063232422, 0.06528419494628906, 0.06634639739990235, 0.06582073974609375, 0.06552150726318359, 0.06543830108642579, 0.06618729400634765, 0.06565475463867188, 0.06527999877929687, 0.06510352325439453, 0.06604576110839844, 0.0660684814453125, 0.06600758361816406, 0.06570515441894531, 0.06550406646728515, 0.06623951721191407, 0.06597277069091798, 0.06570409393310547, 0.06588444519042969, 0.06593052673339844, 0.06609379577636719, 0.06683760070800782, 0.06615337371826172, 0.06606752014160157, 0.06584416198730468, 0.06595708465576172, 0.06618927764892578, 0.06643695831298828, 0.06687615966796875, 0.06616633605957031, 0.06595228576660156, 0.06606655883789063, 0.06641868591308593, 0.06641769409179688, 0.06650355529785157, 0.06634915161132812, 0.06702799987792969, 0.06659078216552734, 0.06666336059570313, 0.06713929748535157, 0.06664335632324218, 0.06630287933349609, 0.06644115447998047, 0.06629222106933594, 0.06507231903076172, 0.06442066955566406, 0.06441280364990234, 0.06548953247070312, 0.06542082977294922, 0.06538041687011718, 0.06509980773925782, 0.06530937957763672, 0.06500717163085938, 0.06462838745117187, 0.0649447021484375, 0.06558060455322266, 0.06562876892089843, 0.06589234924316406, 0.06544793701171875, 0.06586473846435546, 0.06593170928955078, 0.06569833374023437, 0.06545817565917969, 0.06566242980957031, 0.06602403259277344, 0.06578377532958984, 0.06558249664306641, 0.06558367919921874, 0.06575103759765626, 0.06524723052978515, 0.06554768371582032, 0.0661673583984375, 0.06617407989501953, 0.06578883361816407, 0.06566297912597656, 0.0661770248413086, 0.06585689544677735, 0.06562060546875, 0.06565888214111328, 0.0655093765258789, 0.06647196960449218, 0.06593328094482422, 0.06592431640625, 0.06606864166259765, 0.0658458251953125, 0.0657347183227539, 0.06591487884521484, 0.06645145416259765, 0.06640156555175782, 0.06629792022705078, 0.06694297790527344, 0.06663382720947265, 0.06608892822265625, 0.06586339569091797, 0.0663315200805664, 0.0658001937866211, 0.0660664291381836, 0.06624463653564452, 0.0665640640258789, 0.06621718597412109, 0.06602428436279296, 0.06647801971435546, 0.06642073822021484, 0.06665113830566406, 0.06729558563232421, 0.06724787139892578, 0.06553238677978515, 0.06487657928466797, 0.06485542297363281, 0.06471334075927734, 0.06574015808105468, 0.06549565124511719, 0.06512374114990234, 0.06503689575195312, 0.06509954833984374, 0.06488070678710937, 0.06492790222167968, 0.06593897247314454, 0.06559539031982421, 0.06515711975097656, 0.06546685028076171, 0.0652779541015625, 0.06520140838623047, 0.0649324493408203, 0.06576557159423828, 0.06619132995605469, 0.06565888214111328, 0.06532710266113281, 0.06588591766357423, 0.06575337219238281, 0.06548905944824218, 0.06545801544189453, 0.06632653045654296, 0.0660921630859375, 0.06580461120605469, 0.06548127746582032, 0.0662193603515625, 0.06586214447021485, 0.0655013427734375, 0.06512786865234375, 0.06604447937011719, 0.06645670318603515, 0.06587811279296875, 0.06585424041748048, 0.06643228912353516, 0.06586646270751953, 0.06557491302490234, 0.06600498962402344, 0.06623436737060547, 0.06620159912109375, 0.06648627471923828, 0.06696550750732422, 0.06643014526367187, 0.06612665557861327, 0.06580413055419922, 0.06656425476074218, 0.06663168334960938, 0.06616473388671874, 0.066123779296875, 0.06614527893066406, 0.06655446624755859, 0.0663371810913086, 0.066700927734375, 0.0667795181274414, 0.0664815673828125, 0.06647459411621094, 0.06726860809326173, 0.06671155548095703, 0.06664575958251953, 0.06609497833251952, 0.06508803558349609, 0.06462870025634766, 0.06390496063232422, 0.06466854095458985, 0.064901123046875, 0.06564246368408203, 0.06497996520996094, 0.06475682830810547, 0.06573868560791016, 0.06549811553955077, 0.06557571411132812, 0.0658658905029297, 0.06543702697753906, 0.06525727844238281, 0.06545702362060547, 0.06522662353515625, 0.06529427337646485, 0.06475794982910156, 0.06517167663574219, 0.06517727661132812, 0.06561138916015626, 0.06542384338378907, 0.0656214370727539, 0.06545401763916016, 0.06592105865478516, 0.06636399841308593, 0.06572441864013671, 0.06566902160644532, 0.06583273315429687, 0.0656530532836914, 0.06568323516845703, 0.0655383071899414, 0.06576691436767577, 0.06586601257324219, 0.06592111968994141, 0.06610717010498048, 0.06587423706054688, 0.06589190673828126, 0.06572895812988282, 0.06637481689453124, 0.06676156616210938, 0.06644534301757812, 0.06635110473632813, 0.0664060821533203, 0.06620333099365235, 0.06616124725341797, 0.06618019104003907, 0.06607965087890624, 0.06591283416748046, 0.0661053466796875, 0.06583296203613281, 0.0666992645263672, 0.06638784027099609, 0.0661812515258789, 0.06648627471923828, 0.06707539367675781, 0.06690633392333985, 0.06724861145019531, 0.06665001678466796, 0.0665355224609375, 0.06711897277832031, 0.06665158081054688, 0.06610915374755859, 0.06554898834228516, 0.06490940856933594, 0.06470041656494141, 0.06445260620117188, 0.06464300537109376, 0.06483910369873047, 0.06538098907470703, 0.06500057220458984, 0.06501200103759766, 0.06483618927001954, 0.06601519775390625, 0.06618425750732422, 0.06602236938476562, 0.06553190612792968, 0.0659882583618164, 0.06559574127197265, 0.06539254760742187, 0.06500543975830078, 0.06539081573486329, 0.06481078338623047, 0.06480303955078125, 0.06545203399658203, 0.06559673309326172, 0.06553657531738281, 0.06552384185791016, 0.06597122955322265, 0.06633277130126954, 0.06618816375732423, 0.06612892913818359, 0.06629270172119141, 0.06640137481689454, 0.0658658905029297, 0.06587439727783204, 0.06595209503173828, 0.06571987152099609, 0.0655630111694336, 0.06584524536132813, 0.06587187194824219, 0.06576127624511718, 0.06588944244384766, 0.06610755157470703, 0.06677369689941406, 0.06623232269287109, 0.06665010833740234, 0.06692249298095704, 0.06625279998779297, 0.06590156555175782, 0.06626204681396484, 0.06638368225097656, 0.06623452758789063, 0.06584102630615235, 0.06643315124511719, 0.06618313598632812, 0.06618118286132812, 0.06619315338134765, 0.06646505737304688, 0.0668287353515625, 0.0671278076171875, 0.06718643188476563, 0.06650086212158203, 0.0666398696899414, 0.06671737670898438, 0.06626509094238281, 0.06515507507324218, 0.06491136169433594, 0.0649175033569336, 0.06528614044189453, 0.06595958709716797, 0.06577903747558594, 0.06584217834472657, 0.06535987091064453, 0.06508665466308594, 0.06498796844482421, 0.06522163391113281, 0.06521753692626953, 0.06495846557617188, 0.06546841430664062, 0.06557414245605468, 0.06559616088867187, 0.0652759017944336, 0.06517263793945313, 0.06620355224609376, 0.06595683288574218, 0.06603158569335937, 0.06593497467041015, 0.06601705932617187, 0.06547516632080078, 0.06515711975097656, 0.06521651458740234, 0.06587599945068359, 0.0658205795288086, 0.0657900161743164, 0.06575308990478515, 0.06596198272705078, 0.06596141052246093, 0.06579055786132812, 0.06635926055908203, 0.06631423950195313, 0.0662874526977539, 0.06598451232910156, 0.06677107238769531, 0.066236572265625, 0.06599871826171876, 0.06577766418457032, 0.06570751953125, 0.06599027252197266, 0.06556953430175781, 0.06606156921386719, 0.06619225311279296, 0.06609836578369141, 0.06609388732910157, 0.0663531494140625, 0.06694627380371093, 0.0663210220336914, 0.06721347045898438, 0.06667263793945312, 0.06664390563964843, 0.06640847778320312, 0.06645356750488281, 0.06678115081787109, 0.06640640258789063, 0.06641788482666015, 0.06628224182128906, 0.06607174682617188, 0.06641545867919922, 0.06640406036376953, 0.06514431762695312, 0.06491593933105469, 0.06480928039550782, 0.065185791015625, 0.06517286682128906, 0.0650083236694336, 0.06504166412353515, 0.06505545806884766, 0.06513436889648437, 0.06514412689208984, 0.06595059204101562, 0.06606438446044922, 0.0655440673828125, 0.06563001251220703, 0.06611385345458984, 0.065544189453125, 0.0651071014404297, 0.06483177947998046, 0.06518431854248047, 0.06530867004394532, 0.06538444519042969, 0.06523897552490235, 0.06586579132080078, 0.0659372787475586, 0.0659574432373047, 0.06568402862548828, 0.06574038696289063, 0.06609305572509766, 0.06609712219238281, 0.0659296646118164, 0.06590054321289063, 0.0656337890625, 0.06546482849121094, 0.06580429077148438, 0.0657223663330078, 0.0656009292602539, 0.06568816375732422, 0.06589030456542969, 0.06663782501220702, 0.06629580688476562, 0.06596518707275391, 0.06617996978759766, 0.06676070404052735, 0.06638387298583984, 0.06687065887451171, 0.06631897735595703, 0.06636665344238281, 0.06613478088378906, 0.0665109100341797, 0.06606230163574218, 0.06546025848388672, 0.06594290924072266, 0.06595590209960937, 0.06691836547851562, 0.06640402984619141, 0.06689017486572266, 0.06670912170410156, 0.06650975799560546, 0.06688668823242187, 0.06685785675048828, 0.06688563537597657, 0.06697164916992188, 0.06624665832519532, 0.06557843017578124, 0.06500614166259766, 0.06494316864013672, 0.06473564910888673, 0.06482179260253906, 0.06462611389160157, 0.06548470306396484, 0.06520694732666016, 0.06509363555908203, 0.06559334564208984, 0.06609101104736329, 0.06557494354248047, 0.06537216186523437, 0.0652135009765625, 0.06570054626464844, 0.06569596862792969, 0.06543740844726563, 0.06618669128417969, 0.06582972717285156, 0.06541311645507812, 0.06521036529541016, 0.06518169403076172, 0.06590025329589844, 0.06547078704833985, 0.06586969757080079, 0.06622118377685547, 0.06624457550048828, 0.06588028717041015, 0.06547065734863282, 0.065325439453125, 0.0656509780883789, 0.06563145446777344, 0.0658392333984375, 0.06612847900390625, 0.06589155578613282, 0.0658644790649414, 0.06595289611816406, 0.06659980773925782, 0.06636540985107423, 0.06623849487304688, 0.06664150238037109, 0.06630448150634766, 0.06620873260498047, 0.06615519714355468, 0.06625411224365234, 0.06647910308837891, 0.06599884796142579, 0.06629513549804687, 0.06681436920166016, 0.06626534271240235, 0.06607622528076172, 0.0670578842163086, 0.06675888061523437, 0.06628352355957032, 0.06639401245117188, 0.06706524658203125, 0.06669376373291015, 0.06676032257080078, 0.06732230377197265, 0.06697357177734375, 0.06680793762207031, 0.06678221130371094, 0.06632012939453125, 0.06544857788085938, 0.06523881530761719, 0.06496505737304688, 0.06505209350585937, 0.06550105285644531, 0.06508386993408204, 0.0657223663330078, 0.06530220794677734, 0.06527235412597657, 0.06477948760986328, 0.0651965103149414, 0.06550054168701172, 0.06551388549804688, 0.06523875427246094, 0.06561465454101563, 0.06575513458251953, 0.06559916687011719, 0.06523526763916015, 0.065830078125, 0.06573321533203125, 0.06583113861083985, 0.06542700958251953, 0.06547452545166016, 0.06575945281982422, 0.06594172668457031, 0.06554825592041015, 0.06578185272216797, 0.06594556427001953, 0.06572201538085938, 0.0655354232788086, 0.06576016235351563, 0.06627449798583984, 0.066233154296875, 0.06574697875976562, 0.06625651550292969, 0.06573296356201172, 0.06588992309570313, 0.06661772918701171, 0.06610739135742187, 0.06583622741699219, 0.06583094024658204, 0.06600953674316407, 0.06625929260253906, 0.06608086395263672, 0.0665077133178711, 0.06622415924072265, 0.06646578979492188, 0.06612220764160157, 0.06590306854248047, 0.06688153839111328, 0.06623232269287109, 0.06624665832519532, 0.06712899017333984, 0.06666226959228516, 0.06636796569824219, 0.06637126159667969, 0.06645791625976563, 0.06668284606933594, 0.06677916717529297, 0.06640636444091796, 0.06643714904785156, 0.0672799072265625, 0.06628553771972656, 0.06522652435302734, 0.0651204833984375, 0.06464508819580078, 0.0646421127319336, 0.06505366516113281, 0.06527385711669922, 0.06524518585205077, 0.06531890869140625, 0.06494617462158203, 0.06493097686767578, 0.06578672027587891, 0.06555648040771485, 0.06525746917724609, 0.0660859832763672, 0.06566595458984376, 0.06545536041259765, 0.06541577911376953, 0.06604345703125, 0.0657086410522461, 0.06542918395996093, 0.06523945617675782, 0.06537734222412109, 0.06566304016113281, 0.06552649688720703, 0.06531072235107421, 0.06576751708984375, 0.06611759948730468, 0.06609705352783203, 0.06594569396972656, 0.06618931579589844, 0.06594102478027344, 0.06593583679199219, 0.06607039642333984, 0.0665145263671875, 0.06590108489990235, 0.0653803482055664, 0.06601478576660157, 0.06617890930175781, 0.06621231842041016, 0.0657570571899414, 0.06587417602539063, 0.0662908172607422, 0.06630694580078125, 0.0666767349243164, 0.06631014251708985, 0.06598614501953125, 0.06570845031738282, 0.06675628662109374, 0.06618924713134766, 0.06589478302001953, 0.06573014068603515, 0.06681231689453125, 0.06645350646972656, 0.0661376953125, 0.06576783752441406, 0.06647193908691407, 0.06702035522460938, 0.06713311767578126, 0.06673075103759765, 0.06670921325683593, 0.066553955078125, 0.06688162994384765]",tokens/s,15.178530668246328,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.747328,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3543.61344,5170.46272,0.0,4775.215104,4427.072512,s,1,11.0327451171875,11.0327451171875,0.0,11.0327451171875,11.0327451171875,11.0327451171875,11.0327451171875,[11.0327451171875],,kWh,0.00011698165616666547,1.288786926369206e-05,4.9650317498000154e-05,0.0001795198429283577,,MB,1464.664064,5201.92,0.0,4794.089472,4101.02016,s,10,24.840545654296875,2.4840545654296875,0.005153085087852347,2.4845689697265625,2.489054345703125,2.4893999755859375,2.4896764794921875,"[2.470856201171875, 2.481852783203125, 2.48210888671875, 2.482912841796875, 2.485000732421875, 2.48413720703125, 2.486998291015625, 2.48795556640625, 2.4889775390625, 2.48974560546875]",tokens/s,103.05731748518075,kWh,7.23527563062501e-05,7.980278552744877e-06,4.8081927354400034e-05,0.00012841496221339504,tokens/kWh,1993537.167223466,MB,1468.98944,5210.308608,0.0,4802.47808,4101.02272,s,10,16.609656982421875,1.6609656982421874,0.009591729412490228,1.659974853515625,1.6736848510742188,1.676056414794922,1.6779536657714844,"[1.66152978515625, 1.645281982421875, 1.6731578369140625, 1.6579595947265624, 1.6652694091796876, 1.648097900390625, 1.657560302734375, 1.658419921875, 1.6639522705078125, 1.678427978515625]",tokens/s,37.92974175605997,kWh,4.8826849376667114e-05,5.3859862495886415e-06,3.219610909019986e-05,8.640894471645554e-05,tokens/kWh,729091.1861813585,,s,630,16.607083940505998,0.026360450699215842,0.00044283413471687657,0.026275232315063478,0.0267891996383667,0.027089630413055416,0.02799394283294679,"[0.027376960754394532, 0.026368703842163086, 0.02675712013244629, 0.026236640930175782, 0.02658070373535156, 0.02652764892578125, 0.026981023788452147, 0.026679296493530274, 0.026828800201416016, 0.02613817596435547, 0.026312639236450195, 0.02611203193664551, 0.026216928482055663, 0.026341087341308595, 0.026441152572631837, 0.026542943954467775, 0.026488832473754883, 0.026295488357543945, 0.026131263732910158, 0.02632908821105957, 0.02611609649658203, 0.02612633514404297, 0.026101760864257813, 0.026068992614746093, 0.026284032821655274, 0.026497119903564452, 0.026433439254760743, 0.02652067184448242, 0.026581632614135743, 0.026333471298217774, 0.026310752868652344, 0.02627779197692871, 0.026443199157714845, 0.026215200424194337, 0.026232608795166014, 0.02735443115234375, 0.026434175491333006, 0.026458112716674805, 0.026445535659790038, 0.026218399047851563, 0.026506975173950197, 0.02614691162109375, 0.02625190353393555, 0.026293855667114258, 0.026276159286499023, 0.026189311981201172, 0.026765663146972655, 0.026099967956542968, 0.026248512268066407, 0.02612633514404297, 0.026204864501953126, 0.026270912170410155, 0.026274623870849608, 0.02674483108520508, 0.02627993583679199, 0.02613862419128418, 0.026365280151367188, 0.02610633659362793, 0.02603759956359863, 0.02607916831970215, 0.02628291130065918, 0.026199520111083983, 0.02625152015686035, 0.02738444709777832, 0.02653545570373535, 0.02613897514343262, 0.026787967681884767, 0.026032127380371094, 0.026220544815063477, 0.026028032302856444, 0.02615500831604004, 0.02614886474609375, 0.02609152030944824, 0.0261345272064209, 0.026647808074951172, 0.026542783737182617, 0.02639263916015625, 0.026472415924072266, 0.026093055725097656, 0.02595075225830078, 0.025968639373779297, 0.026064895629882814, 0.0258919677734375, 0.02594825553894043, 0.025881343841552735, 0.025950239181518554, 0.025852928161621092, 0.02591862487792969, 0.025816415786743162, 0.025860767364501953, 0.026040159225463867, 0.02589823913574219, 0.025890975952148437, 0.02703011131286621, 0.02599091148376465, 0.025952735900878907, 0.025980703353881834, 0.0262891845703125, 0.026416095733642578, 0.025862144470214843, 0.02593712043762207, 0.025867040634155274, 0.026091167449951172, 0.02597104072570801, 0.026527679443359375, 0.02595382308959961, 0.025868608474731446, 0.02622060775756836, 0.025979040145874023, 0.026107616424560547, 0.02598940849304199, 0.02591948890686035, 0.025837535858154296, 0.026324159622192384, 0.02601055908203125, 0.026016735076904298, 0.02591632080078125, 0.02599907112121582, 0.025913663864135742, 0.02591542434692383, 0.025829343795776366, 0.0258571834564209, 0.025954687118530273, 0.026548704147338866, 0.02603759956359863, 0.026161823272705078, 0.027317888259887697, 0.026281824111938478, 0.02615350341796875, 0.026267648696899414, 0.026218399047851563, 0.028845375061035156, 0.026424224853515626, 0.026544160842895508, 0.026750816345214843, 0.026462207794189452, 0.026352895736694335, 0.02632150459289551, 0.02627302360534668, 0.026359743118286132, 0.026470975875854494, 0.026880416870117187, 0.026326623916625977, 0.026646303176879882, 0.02620275115966797, 0.02610380744934082, 0.026455263137817382, 0.027392799377441407, 0.026728448867797853, 0.026576288223266603, 0.027406848907470704, 0.026787456512451173, 0.026614240646362305, 0.02631679916381836, 0.026705919265747072, 0.026529792785644532, 0.026421247482299806, 0.026161151885986327, 0.026508575439453126, 0.026304224014282226, 0.026403839111328126, 0.026425344467163086, 0.02690640068054199, 0.026300159454345703, 0.02640470314025879, 0.026459808349609374, 0.026327360153198243, 0.026282655715942384, 0.02632089614868164, 0.02633884811401367, 0.026120672225952147, 0.026300735473632812, 0.026688735961914064, 0.02625174331665039, 0.02619740867614746, 0.02619254493713379, 0.02613408088684082, 0.02627827262878418, 0.026543519973754884, 0.026554208755493164, 0.026323711395263672, 0.026484479904174806, 0.026335008621215822, 0.026310304641723632, 0.026311359405517577, 0.026580831527709962, 0.027404224395751953, 0.028201311111450196, 0.027394048690795897, 0.027249439239501953, 0.02650227165222168, 0.026217344284057618, 0.026578655242919923, 0.026470495223999024, 0.02613596725463867, 0.026179584503173828, 0.026032928466796876, 0.026176576614379884, 0.026001760482788086, 0.026048608779907226, 0.025915903091430666, 0.026013696670532226, 0.02585977554321289, 0.025852224349975587, 0.02593337631225586, 0.02642198371887207, 0.026099424362182617, 0.025972736358642577, 0.026908672332763672, 0.02646735954284668, 0.026254304885864256, 0.025968927383422852, 0.026096639633178712, 0.02630512046813965, 0.025960575103759764, 0.02592758369445801, 0.026042463302612305, 0.026103231430053712, 0.025967071533203125, 0.02645964813232422, 0.02696579170227051, 0.02608777618408203, 0.026270175933837892, 0.02615705680847168, 0.026426847457885743, 0.025864736557006836, 0.025999359130859375, 0.026011648178100585, 0.026089344024658203, 0.025954431533813476, 0.025939968109130858, 0.02671820831298828, 0.02613248062133789, 0.026086816787719725, 0.026157663345336913, 0.02624224090576172, 0.026073183059692383, 0.026133216857910157, 0.025890880584716797, 0.026298303604125977, 0.026275840759277344, 0.030109695434570313, 0.027120927810668945, 0.026553056716918946, 0.025851232528686523, 0.026221216201782225, 0.028276735305786133, 0.026429439544677736, 0.026210304260253905, 0.026195968627929687, 0.02632262420654297, 0.026527999877929687, 0.027556320190429688, 0.026664831161499022, 0.026531423568725586, 0.02668351936340332, 0.02637660789489746, 0.026441728591918946, 0.02623244857788086, 0.026370431900024412, 0.02623049545288086, 0.0262126407623291, 0.026701087951660155, 0.026555103302001955, 0.026351648330688475, 0.027228063583374023, 0.02629151916503906, 0.026620704650878905, 0.026594272613525392, 0.02656972885131836, 0.02689740753173828, 0.02631372833251953, 0.026892192840576173, 0.02761315155029297, 0.02687331199645996, 0.026489503860473634, 0.026351615905761717, 0.026619232177734375, 0.02621072006225586, 0.026206464767456056, 0.028094463348388672, 0.027322463989257813, 0.026348960876464843, 0.026054880142211915, 0.02613417625427246, 0.026095935821533203, 0.026252927780151366, 0.02653011131286621, 0.026530176162719726, 0.026371807098388673, 0.026876192092895507, 0.026290016174316408, 0.026286239624023437, 0.02613859176635742, 0.026115360260009764, 0.026091392517089845, 0.026364799499511718, 0.026332735061645508, 0.02612678337097168, 0.026211936950683593, 0.026276256561279295, 0.02607513618469238, 0.02587775993347168, 0.02684511947631836, 0.026249887466430664, 0.02595452880859375, 0.025846784591674804, 0.025922496795654296, 0.026392576217651367, 0.0261345272064209, 0.026026079177856445, 0.026277311325073244, 0.026095903396606446, 0.02589321517944336, 0.025917280197143556, 0.027365503311157228, 0.026208351135253907, 0.026167327880859376, 0.025980031967163086, 0.02598182487487793, 0.02587766456604004, 0.02598556709289551, 0.02596611213684082, 0.02650569534301758, 0.026212671279907226, 0.026247167587280275, 0.02614271926879883, 0.02592972755432129, 0.026402816772460938, 0.02596659278869629, 0.025896608352661134, 0.025854303359985353, 0.026066879272460937, 0.025869760513305664, 0.026528383255004884, 0.025900575637817384, 0.026014175415039063, 0.0258272647857666, 0.02604038429260254, 0.026012895584106445, 0.025848608016967773, 0.025869951248168946, 0.025997472763061524, 0.025894144058227538, 0.02685136032104492, 0.02605673599243164, 0.026184192657470705, 0.025946239471435546, 0.026018335342407228, 0.026248224258422853, 0.026272096633911134, 0.02607529640197754, 0.026073312759399413, 0.026267040252685548, 0.02607561683654785, 0.026073055267333986, 0.026052192687988283, 0.026245695114135742, 0.02612019157409668, 0.02611568069458008, 0.026110368728637694, 0.02693734359741211, 0.026187776565551758, 0.026583040237426758, 0.026048511505126954, 0.02607014465332031, 0.02659008026123047, 0.02623423957824707, 0.02620047950744629, 0.026172767639160155, 0.026047359466552733, 0.026086912155151368, 0.02604083251953125, 0.02660905647277832, 0.026251104354858397, 0.026088031768798828, 0.026148704528808592, 0.02618582344055176, 0.02747065544128418, 0.0267357120513916, 0.027333311080932617, 0.026831071853637697, 0.026584192276000975, 0.026600320816040038, 0.026355712890625, 0.026343423843383788, 0.026328800201416015, 0.026534175872802733, 0.026535200119018554, 0.026381023406982423, 0.02647039985656738, 0.026208192825317382, 0.0265947208404541, 0.02591606330871582, 0.02612224006652832, 0.026421375274658203, 0.026875776290893556, 0.026476415634155273, 0.026302175521850588, 0.026446239471435547, 0.026068992614746093, 0.026019615173339845, 0.026246591567993163, 0.026458879470825196, 0.026479967117309572, 0.026213247299194335, 0.02640412712097168, 0.02644432067871094, 0.026623104095458986, 0.026272640228271485, 0.02676531219482422, 0.02613862419128418, 0.026089471817016603, 0.026046464920043946, 0.025994688034057616, 0.026247743606567384, 0.02590105628967285, 0.025800703048706054, 0.02669955253601074, 0.026167520523071287, 0.026491039276123046, 0.02615894317626953, 0.026104000091552733, 0.02651087951660156, 0.02597324752807617, 0.02596019172668457, 0.02591334342956543, 0.026230815887451173, 0.026327039718627928, 0.02633260726928711, 0.026620479583740236, 0.02611404800415039, 0.025821184158325194, 0.026077184677124023, 0.02609766387939453, 0.025906303405761718, 0.02597999954223633, 0.025912832260131836, 0.02601603126525879, 0.025882623672485353, 0.025939968109130858, 0.027062719345092773, 0.026255264282226562, 0.026578592300415038, 0.026130847930908203, 0.02605801582336426, 0.026120927810668944, 0.026179584503173828, 0.026175071716308593, 0.026100128173828126, 0.026185855865478516, 0.026040128707885742, 0.026142911911010744, 0.025968416213989258, 0.026084703445434572, 0.026062816619873048, 0.026082015991210936, 0.02599065589904785, 0.026241600036621095, 0.02615705680847168, 0.02624620819091797, 0.026139200210571287, 0.02628646469116211, 0.02609516716003418, 0.026214847564697264, 0.026292224884033204, 0.026388479232788087, 0.026340543746948244, 0.026469087600708006, 0.026474592208862304, 0.026413055419921876, 0.02638643264770508, 0.02639023971557617, 0.026549568176269533, 0.027372447967529297, 0.027747840881347657, 0.027523647308349608, 0.026373567581176757, 0.02654879951477051, 0.0262524471282959, 0.026057567596435547, 0.0265229434967041, 0.02623967933654785, 0.02642051124572754, 0.026321632385253906, 0.0264105281829834, 0.026388479232788087, 0.026853599548339845, 0.026196224212646484, 0.0262938232421875, 0.02615750312805176, 0.026398719787597655, 0.02644742393493652, 0.026409343719482423, 0.026234943389892577, 0.02595193672180176, 0.026011327743530273, 0.02623142433166504, 0.02623641586303711, 0.02607155227661133, 0.02619980812072754, 0.026011680603027342, 0.02599776077270508, 0.02599888038635254, 0.02735113525390625, 0.026499296188354494, 0.026197664260864256, 0.026868064880371093, 0.026367488861083983, 0.02645625686645508, 0.026270015716552735, 0.026290176391601562, 0.02609766387939453, 0.02592972755432129, 0.025982208251953125, 0.02604310417175293, 0.026025056838989258, 0.0259880313873291, 0.02589014434814453, 0.02608780860900879, 0.026157344818115234, 0.02604243278503418, 0.02630771255493164, 0.02624764823913574, 0.02611814308166504, 0.02604217529296875, 0.026229280471801758, 0.02631679916381836, 0.026836992263793946, 0.027596799850463868, 0.026967391967773438, 0.02618844795227051, 0.026085376739501953, 0.02698240089416504, 0.02761315155029297, 0.02607516860961914, 0.026561887741088867, 0.02642799949645996, 0.026974271774291993, 0.02627174377441406, 0.026066944122314452, 0.02630451202392578, 0.026247167587280275, 0.02615091133117676, 0.026138143539428713, 0.02665920066833496, 0.026405055999755858, 0.026156896591186522, 0.02625551986694336, 0.026502880096435547, 0.02628348731994629, 0.02627043151855469, 0.026185056686401368, 0.02655302429199219, 0.026467679977416992, 0.026419551849365234, 0.02638876724243164, 0.02673459243774414, 0.027062271118164064, 0.02649616050720215, 0.026530656814575195, 0.026717792510986327, 0.026521600723266602, 0.026487199783325196, 0.02629804801940918, 0.02649648094177246, 0.02645903968811035, 0.027384191513061523, 0.026587295532226562, 0.02675302314758301, 0.030273536682128906, 0.026709600448608397, 0.026595775604248046, 0.026652639389038085, 0.026570816040039063, 0.026863231658935546, 0.02653398323059082, 0.02656073570251465, 0.02624928092956543, 0.02671392059326172, 0.026547903060913085, 0.02652614402770996, 0.026257408142089843, 0.026283103942871092, 0.026256288528442383, 0.026287679672241212, 0.026207904815673828, 0.026676000595092772, 0.026619903564453123, 0.026217472076416014, 0.02701148796081543, 0.02646086311340332, 0.026546079635620116, 0.026782751083374023, 0.026416095733642578, 0.02681875228881836, 0.026521343231201172, 0.026215904235839842, 0.026489280700683595, 0.026515392303466796, 0.026708192825317383, 0.026638111114501952, 0.027111648559570312, 0.026689855575561524, 0.026612831115722657, 0.026468544006347655, 0.02680678367614746, 0.026535488128662108, 0.026767711639404296, 0.02831974411010742, 0.02661974334716797, 0.026474655151367186, 0.02631385612487793, 0.0261760311126709, 0.026222816467285158, 0.02643891143798828, 0.026911327362060547, 0.026542367935180663, 0.02642723274230957, 0.026706079483032226, 0.02651955223083496, 0.026399904251098633, 0.026321344375610352, 0.026365919113159178, 0.02629033660888672, 0.026933536529541016, 0.026439680099487304, 0.026286176681518555, 0.0268002872467041, 0.026203903198242187]",tokens/s,37.93561845396473,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4899.647488,7907.24608,0.0,7511.998464,6895.682048,s,1,12.6356240234375,12.6356240234375,0.0,12.6356240234375,12.6356240234375,12.6356240234375,12.6356240234375,[12.6356240234375],,kWh,0.00016642297527916223,1.8348216214458956e-05,6.775477642601002e-05,0.0002525259679196312,,MB,2781.53216,7921.926144,0.0,7514.095616,6822.139904,s,10,29.33606640625,2.9336066406250003,0.0027620805749914213,2.9343319091796873,2.9365601806640624,2.9366376342773437,2.936699597167969,"[2.927570556640625, 2.931563232421875, 2.93044921875, 2.93380517578125, 2.9352333984375, 2.934239013671875, 2.9344248046875, 2.93552294921875, 2.936715087890625, 2.93654296875]",tokens/s,87.26459657367683,kWh,8.558926053708242e-05,9.439313111170493e-06,5.69508511161998e-05,0.00015197942476445273,tokens/kWh,1684438.5376294516,MB,2785.857536,7924.023296,0.0,7516.192768,6822.142464,s,10,18.646573852539063,1.8646573852539063,0.014621137052411227,1.8582719116210937,1.8802652954101562,1.888691973876953,1.8954333166503907,"[1.850527099609375, 1.855883056640625, 1.859441650390625, 1.8488958740234376, 1.8571021728515624, 1.8702486572265624, 1.8531971435546875, 1.8783927001953125, 1.875766845703125, 1.89711865234375]",tokens/s,33.78636767173259,kWh,5.55465222549977e-05,6.127621784968093e-06,3.683152946519774e-05,9.850567350516351e-05,tokens/kWh,639557.070757936,,s,630,18.644053899765005,0.029593736348833357,0.00044405198722840183,0.029464848518371584,0.03013679370880127,0.03030251989364624,0.030792031955718994,"[0.03071254348754883, 0.029541919708251953, 0.02955721664428711, 0.029360128402709962, 0.029306880950927733, 0.029634559631347656, 0.029908992767333983, 0.02945372772216797, 0.02946928024291992, 0.02952396774291992, 0.029399295806884766, 0.029515520095825195, 0.0293021125793457, 0.029321151733398436, 0.029186784744262697, 0.029198335647583007, 0.02927622413635254, 0.029253568649291992, 0.029136064529418946, 0.029186880111694336, 0.029591552734375, 0.02934169578552246, 0.029371648788452148, 0.02945075225830078, 0.02935219192504883, 0.0291778564453125, 0.029337087631225587, 0.02923369598388672, 0.029177824020385743, 0.02927324867248535, 0.029203296661376953, 0.029161088943481444, 0.029220735549926758, 0.029493471145629883, 0.029196575164794923, 0.029216768264770508, 0.029190143585205077, 0.029231103897094726, 0.029251583099365236, 0.029257728576660157, 0.02933964729309082, 0.029360128402709962, 0.029249120712280273, 0.029301151275634766, 0.0293253116607666, 0.029261823654174804, 0.029237247467041014, 0.02935398483276367, 0.02917180824279785, 0.029302688598632814, 0.029535903930664062, 0.02949567985534668, 0.02934988784790039, 0.029505504608154296, 0.02945996856689453, 0.029374975204467774, 0.02934067153930664, 0.029500160217285156, 0.029519296646118163, 0.029403968811035155, 0.02938431930541992, 0.029304800033569337, 0.02924176025390625, 0.030669727325439454, 0.02971174430847168, 0.02929142379760742, 0.030680320739746095, 0.03008310317993164, 0.02933340835571289, 0.029467168807983397, 0.029480960845947264, 0.029282304763793947, 0.029179744720458985, 0.029290464401245116, 0.029220991134643555, 0.029202495574951172, 0.02922700881958008, 0.029241344451904298, 0.02904697608947754, 0.029086847305297852, 0.029428031921386717, 0.029157632827758788, 0.029153408050537108, 0.029778976440429688, 0.029229856491088866, 0.02905516815185547, 0.0290283203125, 0.029209951400756835, 0.029684415817260744, 0.029206527709960937, 0.029100160598754882, 0.02932086372375488, 0.029094112396240233, 0.029203935623168944, 0.029286943435668945, 0.029403135299682616, 0.02925132751464844, 0.029257984161376954, 0.02914303970336914, 0.02911846351623535, 0.029132799148559572, 0.02918400001525879, 0.02925676727294922, 0.029100992202758788, 0.029458431243896483, 0.02911430358886719, 0.029224992752075196, 0.029204511642456056, 0.029195808410644532, 0.02911894416809082, 0.02998851203918457, 0.02949068832397461, 0.03374371337890625, 0.029265920639038087, 0.029418560028076173, 0.02928531265258789, 0.02918604850769043, 0.02922697639465332, 0.02996227264404297, 0.029130752563476563, 0.029243040084838866, 0.029380224227905274, 0.029438432693481446, 0.029620479583740235, 0.031041536331176758, 0.029624319076538085, 0.03053727912902832, 0.029822399139404297, 0.02981990432739258, 0.030123872756958006, 0.029986719131469726, 0.029599903106689453, 0.029515071868896483, 0.029471391677856444, 0.029599424362182616, 0.029673919677734376, 0.02934979248046875, 0.029194303512573242, 0.029220544815063476, 0.029201791763305663, 0.029344736099243166, 0.02939888000488281, 0.02955446434020996, 0.02947439956665039, 0.029309759140014647, 0.029791807174682616, 0.030344959259033202, 0.029921632766723633, 0.029704095840454102, 0.02973891258239746, 0.029693695068359376, 0.029698816299438477, 0.029925376892089843, 0.029902847290039062, 0.029738239288330078, 0.02970086479187012, 0.029650623321533204, 0.029507904052734374, 0.029325023651123047, 0.029376800537109376, 0.029462528228759766, 0.029619295120239256, 0.029608863830566406, 0.029738176345825196, 0.029552576065063476, 0.02936835289001465, 0.029289024353027344, 0.029423103332519532, 0.029306880950927733, 0.02927043151855469, 0.029180288314819336, 0.02950044822692871, 0.029602815628051758, 0.029435680389404296, 0.029538496017456055, 0.029378559112548826, 0.029210176467895508, 0.029299135208129882, 0.029206527709960937, 0.02919590377807617, 0.02920863914489746, 0.029323232650756835, 0.02931315231323242, 0.029165824890136718, 0.02919808006286621, 0.029110496520996093, 0.029138944625854493, 0.02918400001525879, 0.029151264190673827, 0.030474943161010744, 0.029593791961669922, 0.029539936065673827, 0.02919196891784668, 0.029149824142456055, 0.02916348838806152, 0.029130144119262694, 0.029136800765991212, 0.029338336944580077, 0.029095935821533202, 0.029458431243896483, 0.029310976028442383, 0.029206527709960937, 0.029280256271362305, 0.02935603141784668, 0.029239295959472656, 0.029437408447265626, 0.030534175872802733, 0.03058483123779297, 0.02935753631591797, 0.029340192794799803, 0.02924492835998535, 0.029546848297119142, 0.0292737922668457, 0.02919487953186035, 0.029175552368164062, 0.029231199264526365, 0.029112447738647462, 0.029310848236083983, 0.029138944625854493, 0.029198495864868165, 0.029159263610839845, 0.029148319244384765, 0.029137567520141603, 0.029339712142944337, 0.029177440643310546, 0.029293088912963866, 0.029276159286499022, 0.029212671279907225, 0.029241344451904298, 0.02920038414001465, 0.029211999893188477, 0.029330080032348632, 0.029175487518310547, 0.02939321517944336, 0.02932512092590332, 0.029165760040283203, 0.02915033531188965, 0.02918012809753418, 0.02913267135620117, 0.029246240615844726, 0.029294591903686523, 0.029390815734863282, 0.029324447631835938, 0.029275007247924804, 0.02929420852661133, 0.02930521583557129, 0.029413375854492187, 0.02935603141784668, 0.02930624008178711, 0.029335296630859375, 0.029690752029418944, 0.029834848403930664, 0.030721311569213868, 0.029937824249267577, 0.030032543182373046, 0.029799423217773437, 0.02957155227661133, 0.02943948745727539, 0.029430719375610353, 0.02955366325378418, 0.029494272232055665, 0.029425664901733397, 0.02948310470581055, 0.02938256072998047, 0.029409151077270507, 0.029450368881225587, 0.02932326316833496, 0.029327360153198243, 0.029351936340332032, 0.029671424865722655, 0.02998476791381836, 0.03020595169067383, 0.02958892822265625, 0.029339967727661134, 0.02936454391479492, 0.029387872695922853, 0.029423711776733398, 0.029518592834472657, 0.029437952041625977, 0.029287647247314454, 0.02928838348388672, 0.029204511642456056, 0.029217344284057617, 0.029395200729370116, 0.029393983840942384, 0.029565887451171877, 0.029929279327392578, 0.03021126365661621, 0.029519968032836914, 0.02934480094909668, 0.029282175064086913, 0.029241344451904298, 0.029734144210815428, 0.02920729637145996, 0.029421567916870117, 0.029243392944335936, 0.02922217559814453, 0.029207263946533203, 0.029199552536010743, 0.029208736419677736, 0.029305503845214843, 0.029378719329833984, 0.029288288116455077, 0.02954185676574707, 0.02959823989868164, 0.029483007431030273, 0.029372415542602538, 0.029304832458496095, 0.029220096588134764, 0.02926464080810547, 0.029279903411865236, 0.02930031967163086, 0.029352703094482423, 0.0293571834564209, 0.029410175323486328, 0.030738048553466797, 0.02966771125793457, 0.0294072322845459, 0.030195711135864257, 0.02919363212585449, 0.029252191543579102, 0.029183744430541992, 0.029293983459472657, 0.029631328582763673, 0.029304832458496095, 0.02941542434692383, 0.02948646354675293, 0.029382272720336913, 0.02944118309020996, 0.029408767700195314, 0.02941302490234375, 0.029481664657592773, 0.02971238327026367, 0.030040063858032227, 0.030084672927856444, 0.03004607963562012, 0.03000172805786133, 0.030136320114135744, 0.03004342460632324, 0.0302271671295166, 0.03028995132446289, 0.030136287689208983, 0.03024844741821289, 0.03030067253112793, 0.030271232604980467, 0.03010367965698242, 0.0305296630859375, 0.0316231689453125, 0.030043296813964844, 0.029959007263183592, 0.029837312698364257, 0.0299234561920166, 0.029363199234008788, 0.029377248764038084, 0.029335712432861327, 0.02946236801147461, 0.0293602237701416, 0.029421632766723632, 0.02938012886047363, 0.029370784759521484, 0.029296703338623047, 0.029373600006103517, 0.029233983993530274, 0.029304864883422852, 0.029310976028442383, 0.02941129684448242, 0.029381664276123046, 0.0293304328918457, 0.029343391418457033, 0.029614431381225586, 0.029376319885253906, 0.029363744735717772, 0.029460575103759764, 0.02976326370239258, 0.02947564888000488, 0.029556543350219726, 0.029427967071533202, 0.029468671798706055, 0.03061199951171875, 0.029850496292114257, 0.029348831176757812, 0.029140960693359374, 0.029380640029907226, 0.029149183273315428, 0.0293621768951416, 0.029222911834716796, 0.02948054313659668, 0.029390239715576173, 0.029723712921142578, 0.02931705665588379, 0.029405183792114258, 0.02932918357849121, 0.029499616622924805, 0.029345792770385744, 0.029586816787719728, 0.029299423217773436, 0.029278175354003906, 0.029398975372314454, 0.02935353660583496, 0.02998931121826172, 0.029310976028442383, 0.029245407104492187, 0.02931100845336914, 0.029302783966064453, 0.029165536880493163, 0.029380640029907226, 0.029193887710571288, 0.029171104431152343, 0.0291409912109375, 0.029336191177368163, 0.02938211250305176, 0.02939334487915039, 0.029276191711425783, 0.029254016876220704, 0.02926348876953125, 0.02926630401611328, 0.029249759674072267, 0.02918806457519531, 0.02921855926513672, 0.029394144058227538, 0.02952432060241699, 0.02958729553222656, 0.029608287811279298, 0.029345184326171874, 0.029358144760131835, 0.029518400192260742, 0.030132511138916015, 0.02962227249145508, 0.029550592422485353, 0.02929193687438965, 0.029309375762939453, 0.029405344009399415, 0.029894496917724608, 0.029599903106689453, 0.02944144058227539, 0.029447872161865233, 0.029397375106811525, 0.02924527931213379, 0.029253631591796874, 0.029292736053466797, 0.029230880737304687, 0.030592735290527345, 0.029710464477539063, 0.02971664047241211, 0.02952396774291992, 0.029761375427246092, 0.029657344818115234, 0.029403039932250977, 0.02940310478210449, 0.02949081611633301, 0.029411584854125976, 0.029478559494018553, 0.029385215759277345, 0.029294591903686523, 0.029368064880371095, 0.029644319534301758, 0.02951807975769043, 0.029606367111206056, 0.02983526420593262, 0.02953011131286621, 0.029421567916870117, 0.029601791381835937, 0.029792255401611328, 0.029877504348754882, 0.02946735954284668, 0.029454208374023436, 0.02959561538696289, 0.029309247970581053, 0.029450111389160157, 0.029493408203125, 0.0295664005279541, 0.029517248153686525, 0.029534303665161132, 0.029401216506958008, 0.029442815780639647, 0.029576448440551757, 0.02962713623046875, 0.030141504287719726, 0.030131135940551758, 0.03040460777282715, 0.030574880599975585, 0.030158559799194337, 0.03015884780883789, 0.030029823303222656, 0.030103551864624024, 0.02999091148376465, 0.03073548889160156, 0.030911584854125977, 0.030304031372070314, 0.030256895065307616, 0.030145792007446288, 0.030112735748291014, 0.030151775360107422, 0.0302109432220459, 0.029769792556762695, 0.029744800567626954, 0.02976291275024414, 0.029649152755737304, 0.02982374382019043, 0.030008384704589844, 0.030035968780517577, 0.03004921531677246, 0.03024070358276367, 0.030019647598266603, 0.030801952362060545, 0.029956192016601563, 0.029792255401611328, 0.029654464721679687, 0.02974777603149414, 0.02970403289794922, 0.02966534423828125, 0.029595296859741212, 0.029691423416137695, 0.02948793601989746, 0.02970159912109375, 0.029938304901123047, 0.02974300765991211, 0.029611488342285157, 0.0298789119720459, 0.02953398323059082, 0.029583583831787108, 0.02964816093444824, 0.029874912261962892, 0.02962985610961914, 0.02969046401977539, 0.02953625679016113, 0.02971238327026367, 0.029566015243530273, 0.029665695190429688, 0.029766176223754885, 0.029575168609619142, 0.029679616928100585, 0.029859296798706053, 0.02960643196105957, 0.02992915153503418, 0.030304576873779295, 0.02971238327026367, 0.029668384552001954, 0.029842144012451173, 0.02988688087463379, 0.02977916717529297, 0.02964339256286621, 0.02994175910949707, 0.029799936294555664, 0.029732767105102538, 0.0297762565612793, 0.029710559844970702, 0.029621856689453125, 0.029633983612060547, 0.030079008102416992, 0.029864864349365236, 0.02992848014831543, 0.029894784927368166, 0.029808576583862306, 0.02972153663635254, 0.029685760498046877, 0.029638656616210936, 0.030123104095458986, 0.029653919219970702, 0.029587135314941407, 0.02957321548461914, 0.029908512115478517, 0.029604543685913087, 0.0295927677154541, 0.029620159149169923, 0.03045369529724121, 0.02989356803894043, 0.030905855178833007, 0.030767744064331054, 0.030205503463745117, 0.0299935359954834, 0.0330747184753418, 0.030167327880859376, 0.030105632781982423, 0.03005414390563965, 0.030148992538452147, 0.030074911117553713, 0.030242816925048828, 0.02999504089355469, 0.029949920654296875, 0.029949056625366212, 0.029920127868652342, 0.029945472717285156, 0.030054784774780272, 0.030155839920043944, 0.029997024536132812, 0.029995616912841798, 0.029865888595581053, 0.029923807144165038, 0.029951263427734375, 0.02988924789428711, 0.030100608825683595, 0.029935743331909178, 0.030046720504760743, 0.030060447692871094, 0.030239072799682618, 0.030564096450805663, 0.030438943862915038, 0.03017919921875, 0.03025391960144043, 0.03075481605529785, 0.030466047286987305, 0.030244319915771485, 0.030120512008666993, 0.03024278450012207, 0.030282943725585938, 0.030171968460083007, 0.030023679733276368, 0.029816831588745117, 0.03001475143432617, 0.029989599227905273, 0.02993152046203613, 0.029853696823120116, 0.029988927841186522, 0.029703615188598632, 0.02971446418762207, 0.029639135360717772, 0.029888511657714844, 0.029670751571655274, 0.02958736038208008, 0.02952668762207031, 0.029651039123535155, 0.029499391555786132, 0.029507583618164062, 0.03058086395263672, 0.03027302360534668, 0.03011529541015625, 0.030141056060791014, 0.03005264091491699, 0.030207712173461913]",tokens/s,33.79093427786863,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,875.159552,690.946048,0.0,295.698432,277.263872,s,1,8.11909912109375,8.11909912109375,0.0,8.11909912109375,8.11909912109375,8.11909912109375,8.11909912109375,[8.11909912109375],,kWh,2.4179700970785235e-05,2.6600073390606967e-06,7.5708393899864834e-06,3.441054769983242e-05,,MB,1181.384704,743.374848,0.0,335.54432,312.39168,s,17,0.5514090576171875,0.03243582691865808,0.0016088973820605303,0.03220207977294922,0.03238558731079102,0.033655135345458984,0.03771621688842774,"[0.038731487274169925, 0.03220816040039062, 0.032158432006835935, 0.03160505676269531, 0.032301982879638674, 0.03220207977294922, 0.031267936706542966, 0.03217087936401367, 0.03163433647155762, 0.032350048065185544, 0.032299488067626954, 0.03149331283569336, 0.03238528060913086, 0.031835424423217774, 0.03221033477783203, 0.03216876983642578, 0.03238604736328125]",tokens/s,7892.507277276812,kWh,1.174110313102384e-06,1.2948132147932021e-07,7.818178945301146e-07,2.0854095291118186e-06,tokens/kWh,122757662.9080769,MB,1192.22272,770.637824,0.0,362.807296,312.39424,s,17,9.954290649414064,0.5855465087890624,0.00707897054558384,0.5835198364257812,0.5953194702148438,0.5984454467773438,0.6034490014648438,"[0.5942778930664062, 0.5881565551757812, 0.583928955078125, 0.587521484375, 0.5829520263671875, 0.5853764038085938, 0.5968818359375, 0.6046998901367188, 0.581895263671875, 0.579759765625, 0.5776902465820313, 0.5771841430664062, 0.5876484985351562, 0.5835198364257812, 0.5810463256835937, 0.58303857421875, 0.5787129516601562]",tokens/s,107.59179510827747,kWh,1.667274623395756e-05,1.8387166656205125e-06,6.767028013822432e-06,2.527849091340051e-05,tokens/kWh,2492237.3814096134,,s,1071,9.946050537109388,0.009286695179373833,0.00020343916655361172,0.009230015754699707,0.009521151542663574,0.009662544250488281,0.009936396598815917,"[0.009731743812561035, 0.009677151679992675, 0.00960921573638916, 0.009449472427368164, 0.009373536109924316, 0.009352992057800293, 0.009308095932006835, 0.00930844783782959, 0.009299231529235839, 0.009304224014282226, 0.009310367584228516, 0.009236960411071778, 0.00923862361907959, 0.00928876781463623, 0.009349663734436036, 0.009344544410705567, 0.009353119850158692, 0.00927023983001709, 0.00931158447265625, 0.009288191795349121, 0.009255071640014649, 0.009368991851806641, 0.009234335899353028, 0.009263808250427247, 0.009250368118286133, 0.009241024017333984, 0.00923680019378662, 0.009315936088562012, 0.009386303901672364, 0.009350943565368652, 0.009359040260314942, 0.009299360275268554, 0.009298015594482421, 0.009421504020690918, 0.009414400100708008, 0.009349504470825195, 0.009349120140075684, 0.009332159996032714, 0.009402272224426269, 0.009333024024963379, 0.009668416023254394, 0.009922975540161133, 0.009457023620605468, 0.009464608192443848, 0.009521151542663574, 0.009461888313293457, 0.00944320011138916, 0.009545727729797364, 0.009711615562438965, 0.009750144004821777, 0.0097609281539917, 0.009914848327636719, 0.009576191902160644, 0.009527296066284179, 0.009424896240234374, 0.00943120002746582, 0.009561951637268067, 0.00988486385345459, 0.009448224067687989, 0.009414688110351562, 0.009363455772399902, 0.009321791648864746, 0.009351584434509277, 0.009043487548828125, 0.009291808128356933, 0.009309120178222656, 0.00925267219543457, 0.009238719940185547, 0.00929910373687744, 0.009266016006469726, 0.009254688262939454, 0.009351391792297364, 0.009289728164672852, 0.009258655548095704, 0.009277888298034668, 0.009217951774597168, 0.009242624282836913, 0.009240032196044922, 0.00917967987060547, 0.00922812843322754, 0.009195584297180176, 0.009298015594482421, 0.00925596809387207, 0.009187999725341798, 0.009230655670166016, 0.009269248008728028, 0.009232352256774903, 0.009261311531066895, 0.009467552185058594, 0.009498751640319825, 0.009830400466918946, 0.009260640144348145, 0.009249024391174317, 0.00928758430480957, 0.00917734432220459, 0.009188480377197265, 0.00924454402923584, 0.009192447662353515, 0.01093222427368164, 0.011141119956970215, 0.009318400382995605, 0.009404224395751953, 0.009252927780151366, 0.009327903747558593, 0.009253888130187989, 0.009301888465881347, 0.009322688102722168, 0.009342752456665038, 0.00931811237335205, 0.009314656257629395, 0.009279711723327637, 0.009246432304382324, 0.009236191749572754, 0.009251296043395996, 0.009238335609436035, 0.009226240158081055, 0.009217696189880371, 0.00920406436920166, 0.009259072303771973, 0.009275327682495118, 0.009217568397521973, 0.009233887672424317, 0.009225215911865235, 0.009186592102050781, 0.009259743690490723, 0.009318400382995605, 0.008987839698791505, 0.009204480171203613, 0.009240223884582519, 0.009209535598754882, 0.009182175636291505, 0.009200639724731445, 0.009210111618041993, 0.009293984413146973, 0.009240511894226075, 0.009281951904296875, 0.009265151977539063, 0.009302016258239745, 0.009338879585266113, 0.009207807540893554, 0.009234399795532226, 0.009193152427673339, 0.009262911796569824, 0.00974287986755371, 0.009180319786071778, 0.009220959663391114, 0.00918876838684082, 0.00921455955505371, 0.009193471908569336, 0.00923363208770752, 0.009214079856872559, 0.009349791526794434, 0.009367551803588867, 0.009406463623046875, 0.009319519996643067, 0.009290816307067871, 0.009283552169799805, 0.009320351600646972, 0.009185279846191406, 0.00922976016998291, 0.009245216369628906, 0.009203071594238282, 0.009343615531921387, 0.009234399795532226, 0.009316384315490723, 0.009251008033752442, 0.00925654411315918, 0.00925062370300293, 0.009260576248168945, 0.009298815727233886, 0.009380064010620116, 0.009391519546508789, 0.00943171215057373, 0.009407456398010255, 0.009315072059631347, 0.009385984420776367, 0.009227840423583985, 0.009222047805786133, 0.00921180820465088, 0.009186976432800293, 0.00916374397277832, 0.009177087783813476, 0.009186528205871581, 0.009175775527954101, 0.00920787239074707, 0.009190815925598145, 0.00926371192932129, 0.009220095634460449, 0.00923401641845703, 0.009014559745788573, 0.009423871994018555, 0.009730048179626465, 0.009215616226196289, 0.009243007659912109, 0.009515263557434083, 0.009205056190490722, 0.009290240287780761, 0.009248384475708008, 0.009203264236450195, 0.00969600009918213, 0.009272800445556641, 0.009251359939575196, 0.009294015884399414, 0.009320256233215331, 0.009420000076293945, 0.009270048141479493, 0.009297247886657714, 0.009218943595886231, 0.009256735801696777, 0.009366815567016601, 0.009398943901062011, 0.00931334400177002, 0.009319583892822265, 0.009451583862304687, 0.00931817626953125, 0.00935321617126465, 0.00928707218170166, 0.009248640060424805, 0.00926313591003418, 0.009218400001525878, 0.00932694435119629, 0.00929321575164795, 0.009171551704406738, 0.009182751655578613, 0.009209568023681641, 0.009313055992126465, 0.009229887962341308, 0.00917955207824707, 0.009152640342712402, 0.009158528327941895, 0.00923635196685791, 0.009449600219726562, 0.009359519958496093, 0.009313247680664062, 0.009325440406799317, 0.009361408233642577, 0.009541791915893555, 0.00949232006072998, 0.00930611228942871, 0.009329792022705077, 0.009292767524719239, 0.009316255569458008, 0.009629088401794434, 0.00935587215423584, 0.00941977596282959, 0.00973840045928955, 0.009304832458496094, 0.009254112243652343, 0.009253824234008789, 0.009207743644714356, 0.009179136276245118, 0.00924073600769043, 0.009317983627319336, 0.009296287536621093, 0.00930611228942871, 0.009254655838012694, 0.009195679664611817, 0.009205120086669922, 0.009173888206481934, 0.009205344200134278, 0.009193663597106933, 0.009214015960693359, 0.009269248008728028, 0.009239744186401368, 0.009255104064941407, 0.009259519577026367, 0.009183327674865722, 0.009160736083984374, 0.009170944213867188, 0.009119775772094727, 0.009173215866088867, 0.009223039627075196, 0.009515680313110352, 0.009552127838134765, 0.009416000366210937, 0.009347904205322265, 0.009312095642089844, 0.009303263664245606, 0.009265952110290528, 0.00923852825164795, 0.00921395206451416, 0.009230015754699707, 0.009277759552001953, 0.009172863960266113, 0.009201791763305664, 0.009166848182678223, 0.009239551544189453, 0.009183679580688477, 0.009142848014831543, 0.009146368026733399, 0.009172736167907715, 0.009178720474243163, 0.00920847988128662, 0.009270336151123047, 0.009294783592224121, 0.009768959999084472, 0.009197312355041504, 0.009219903945922852, 0.009169343948364259, 0.009219264030456543, 0.0091942720413208, 0.009182720184326172, 0.00929856014251709, 0.009728032112121582, 0.009174880027770995, 0.009185312271118164, 0.009299679756164551, 0.009158944129943848, 0.009230015754699707, 0.009134336471557617, 0.009203776359558106, 0.009141856193542481, 0.009140159606933594, 0.009148896217346192, 0.00918563175201416, 0.009077119827270507, 0.009209088325500488, 0.009234623908996582, 0.009224127769470215, 0.009187968254089355, 0.009209856033325196, 0.009260448455810547, 0.009208415985107422, 0.009190815925598145, 0.009314687728881835, 0.009300191879272461, 0.009327775955200195, 0.00931715202331543, 0.009405823707580566, 0.00928179168701172, 0.009291872024536133, 0.009333087921142577, 0.009225855827331543, 0.009308544158935547, 0.009441280364990234, 0.009404416084289552, 0.00936143970489502, 0.009383904457092284, 0.009496576309204101, 0.00932863998413086, 0.009326623916625977, 0.009356575965881348, 0.009364192008972167, 0.009404383659362792, 0.00935321617126465, 0.009359040260314942, 0.009449248313903808, 0.009282079696655274, 0.009332736015319825, 0.009279808044433594, 0.009279168128967286, 0.009332736015319825, 0.009297120094299316, 0.009255711555480956, 0.009210911750793457, 0.009200927734375, 0.00928435230255127, 0.009196479797363282, 0.009171072006225587, 0.009148287773132324, 0.009137439727783202, 0.009206496238708495, 0.009230367660522461, 0.009295392036437988, 0.00948419189453125, 0.009329055786132813, 0.009330816268920898, 0.00929974365234375, 0.009289952278137207, 0.009328960418701172, 0.009340352058410644, 0.009244768142700196, 0.009261216163635254, 0.009223296165466308, 0.009198464393615723, 0.009148415565490722, 0.009164159774780273, 0.009168928146362304, 0.008978176116943359, 0.009177248001098632, 0.009235679626464844, 0.009191583633422852, 0.00923852825164795, 0.009224543571472168, 0.00926144027709961, 0.009307423591613769, 0.00932470417022705, 0.009371968269348144, 0.00944054412841797, 0.009470975875854493, 0.00949836826324463, 0.009414624214172364, 0.009404416084289552, 0.009512031555175781, 0.009319328308105468, 0.00943734359741211, 0.009784992218017579, 0.00940006446838379, 0.009429759979248047, 0.009385663986206055, 0.009420831680297852, 0.00944649600982666, 0.009388863563537598, 0.009461600303649903, 0.00950432014465332, 0.009396896362304688, 0.009358783721923829, 0.009509632110595703, 0.009485183715820312, 0.009565119743347167, 0.009440896034240722, 0.009501055717468261, 0.009449472427368164, 0.009455615997314454, 0.009541728019714356, 0.00956118392944336, 0.00956704044342041, 0.009621503829956055, 0.009667648315429687, 0.009649087905883789, 0.009653440475463867, 0.009630528450012207, 0.00954684829711914, 0.009528287887573242, 0.009607104301452637, 0.00964633560180664, 0.00960867214202881, 0.00953987216949463, 0.00951296043395996, 0.00957750415802002, 0.009636832237243651, 0.009539711952209472, 0.009707391738891602, 0.009590527534484864, 0.009701631546020507, 0.009517056465148926, 0.00933843231201172, 0.009386367797851562, 0.009386048316955567, 0.00942899227142334, 0.009451519966125489, 0.009762687683105469, 0.009825856208801269, 0.009831199645996094, 0.010010623931884765, 0.009989631652832032, 0.009808032035827637, 0.009736543655395507, 0.009832287788391113, 0.00966089630126953, 0.00966419219970703, 0.009672160148620606, 0.00967734432220459, 0.009697279930114745, 0.009645088195800781, 0.009729120254516601, 0.00966438388824463, 0.009651488304138183, 0.009741120338439942, 0.009580448150634765, 0.00962713623046875, 0.009621664047241211, 0.009630047798156738, 0.009644351959228515, 0.009603008270263672, 0.00969600009918213, 0.009614336013793945, 0.00963379192352295, 0.009574399948120118, 0.009659520149230957, 0.009605664253234864, 0.009557696342468261, 0.009563039779663086, 0.009530176162719726, 0.009477184295654298, 0.009541567802429199, 0.009611328125, 0.009811840057373046, 0.009597151756286622, 0.00961235237121582, 0.009691871643066407, 0.00961740779876709, 0.009537376403808594, 0.01026863956451416, 0.009468095779418945, 0.009485919952392579, 0.009451744079589843, 0.009508064270019531, 0.009446368217468262, 0.009453568458557129, 0.009437184333801269, 0.009490464210510254, 0.009437151908874512, 0.009441280364990234, 0.009365119934082032, 0.00928985595703125, 0.009369440078735351, 0.00936956787109375, 0.00932271957397461, 0.009318623542785644, 0.009276736259460449, 0.00928179168701172, 0.00924623966217041, 0.009255840301513671, 0.008937664031982422, 0.009189920425415039, 0.009183199882507324, 0.009181216239929199, 0.009158207893371581, 0.009146816253662109, 0.009114784240722656, 0.009098079681396485, 0.009133952140808105, 0.009097344398498535, 0.009136128425598144, 0.009514143943786621, 0.010959775924682617, 0.010619935989379882, 0.009345952033996583, 0.00922214412689209, 0.009215104103088378, 0.009132160186767577, 0.009325311660766602, 0.009209856033325196, 0.00925836753845215, 0.009120384216308593, 0.009431039810180664, 0.00919257640838623, 0.009148544311523437, 0.00935807991027832, 0.009140224456787109, 0.009138175964355469, 0.009151552200317383, 0.009164735794067383, 0.00926643180847168, 0.009193535804748534, 0.009223423957824706, 0.009178815841674804, 0.009118271827697753, 0.009164992332458497, 0.009144319534301757, 0.009164799690246582, 0.009109503746032714, 0.009140031814575195, 0.009151743888854981, 0.009104000091552735, 0.009156479835510253, 0.009187775611877442, 0.009119199752807618, 0.009103903770446778, 0.009136351585388183, 0.009238304138183594, 0.009223487854003906, 0.009155263900756836, 0.009121055603027345, 0.009149087905883788, 0.009164192199707032, 0.009195775985717773, 0.009300479888916016, 0.009119647979736328, 0.009152544021606446, 0.009131999969482421, 0.009174688339233399, 0.0091527681350708, 0.009105567932128905, 0.00916380786895752, 0.009155488014221192, 0.008931455612182617, 0.009167072296142578, 0.009155200004577637, 0.009161919593811034, 0.009159487724304199, 0.009138143539428711, 0.009121824264526367, 0.009138175964355469, 0.009111680030822755, 0.009137568473815917, 0.00918934440612793, 0.009284095764160156, 0.009119744300842286, 0.009228287696838379, 0.009516608238220215, 0.009161151885986329, 0.009191424369812011, 0.00922812843322754, 0.009181568145751953, 0.0092260160446167, 0.009199359893798827, 0.010575903892517089, 0.00928159999847412, 0.009180671691894531, 0.009163423538208008, 0.009210911750793457, 0.009171839714050292, 0.009148672103881835, 0.009175040245056153, 0.00957795238494873, 0.009149056434631347, 0.009210911750793457, 0.009133888244628905, 0.009147232055664062, 0.009126976013183593, 0.009194111824035645, 0.009130335807800294, 0.00909727954864502, 0.00916476821899414, 0.009108960151672363, 0.00924880027770996, 0.00909727954864502, 0.009240127563476563, 0.009094207763671876, 0.009115455627441406, 0.009126175880432129, 0.009091967582702637, 0.009237279891967773, 0.009153984069824218, 0.009165472030639648, 0.009119711875915527, 0.009147583961486816, 0.009141119956970215, 0.00912377643585205, 0.009103391647338867, 0.009366656303405762, 0.00915135955810547, 0.009183072090148925, 0.009146783828735352, 0.009182432174682616, 0.009101856231689453, 0.009140416145324707, 0.009089088439941407, 0.008951807975769043, 0.009136128425598144, 0.009203264236450195, 0.009165311813354492, 0.009153504371643067, 0.009099967956542969, 0.00914038372039795, 0.009157024383544921, 0.009111231803894042, 0.009099295616149902, 0.009142271995544434, 0.009237824440002441, 0.00911574363708496, 0.009192031860351562, 0.009180447578430177, 0.009135871887207032, 0.00918627166748047, 0.009244671821594238, 0.009224160194396973, 0.009387264251708985, 0.009204544067382813, 0.009205727577209473, 0.009121696472167968, 0.009244192123413085, 0.009113247871398927, 0.009105695724487304, 0.009184032440185548, 0.009099072456359863, 0.009142175674438476, 0.009134207725524902, 0.009368960380554198, 0.009149056434631347, 0.009109503746032714, 0.009179007530212403, 0.009164832115173339, 0.009148672103881835, 0.00912889575958252, 0.009165056228637695, 0.00920028781890869, 0.009172991752624511, 0.009224160194396973, 0.009209888458251953, 0.009102720260620117, 0.009120575904846192, 0.00915231990814209, 0.009148415565490722, 0.009135104179382325, 0.009145343780517578, 0.009101183891296386, 0.00911580753326416, 0.009088640213012695, 0.009158143997192383, 0.009223008155822754, 0.009103360176086426, 0.009326784133911133, 0.009200799942016601, 0.00914243221282959, 0.009173503875732422, 0.009150464057922364, 0.009154560089111329, 0.009117728233337402, 0.009184831619262696, 0.009133824348449707, 0.009065343856811523, 0.009183232307434081, 0.00910752010345459, 0.009154208183288574, 0.00922652816772461, 0.009280703544616698, 0.009169440269470214, 0.009213600158691406, 0.009253408432006836, 0.009184960365295411, 0.009170975685119629, 0.00918278408050537, 0.009163583755493164, 0.009129983901977539, 0.009142271995544434, 0.009116703987121583, 0.009164863586425782, 0.009319328308105468, 0.009152511596679687, 0.0091495361328125, 0.009184160232543946, 0.00912611198425293, 0.00917024040222168, 0.009154144287109376, 0.009175935745239257, 0.009188672065734864, 0.009221952438354492, 0.009149056434631347, 0.009183775901794434, 0.009166560173034667, 0.009153599739074708, 0.009116640090942383, 0.009129952430725098, 0.009123231887817383, 0.009124095916748048, 0.009144800186157227, 0.009178432464599609, 0.009120320320129395, 0.009143808364868163, 0.009180800437927246, 0.009158880233764648, 0.009105759620666504, 0.00909727954864502, 0.009113856315612793, 0.009125887870788574, 0.009111328125, 0.00912816047668457, 0.009135680198669434, 0.009130368232727051, 0.009127776145935058, 0.009156831741333008, 0.009219840049743652, 0.009105664253234863, 0.009129280090332032, 0.009108160018920898, 0.009127840042114258, 0.009095264434814452, 0.009136128425598144, 0.00913542366027832, 0.009140928268432618, 0.009181183815002441, 0.009131232261657714, 0.009141023635864259, 0.00924505615234375, 0.009340800285339356, 0.009297183990478516, 0.009321632385253907, 0.009225919723510742, 0.009322112083435058, 0.009471391677856445, 0.009503840446472168, 0.009500543594360351, 0.00949392032623291, 0.009496352195739747, 0.009342880249023437, 0.009273471832275391, 0.009305120468139649, 0.009289440155029296, 0.009334815979003906, 0.009473247528076172, 0.009223135948181152, 0.009225728034973145, 0.009240896224975585, 0.009232383728027344, 0.009271167755126953, 0.00928371238708496, 0.009256319999694824, 0.009211615562438965, 0.0091943998336792, 0.009150464057922364, 0.00919961643218994, 0.009150464057922364, 0.0092609281539917, 0.009905471801757812, 0.010527039527893066, 0.009738752365112305, 0.009246272087097168, 0.010176959991455079, 0.009252863883972168, 0.009269472122192382, 0.009390975952148438, 0.009339360237121582, 0.009269696235656738, 0.009256768226623534, 0.009357503890991211, 0.009222368240356445, 0.009342752456665038, 0.00922985553741455, 0.009181280136108399, 0.00925324821472168, 0.00920364761352539, 0.009138239860534667, 0.009240575790405273, 0.009166655540466308, 0.00917728042602539, 0.009146368026733399, 0.00911359977722168, 0.009225503921508789, 0.009280223846435547, 0.009166848182678223, 0.009168895721435547, 0.009342975616455078, 0.009181023597717285, 0.009262944221496582, 0.009121952056884765, 0.009146335601806641, 0.008925951957702637, 0.00912604808807373, 0.009217439651489258, 0.009218688011169434, 0.009207615852355957, 0.009249792098999024, 0.00919654369354248, 0.009201663970947266, 0.0091626558303833, 0.00913315200805664, 0.009123071670532227, 0.009481568336486817, 0.009398591995239259, 0.00996771240234375, 0.009885696411132813, 0.009217344284057617, 0.009226688385009766, 0.009304160118103028, 0.009156160354614258, 0.009183839797973633, 0.00922214412689209, 0.009154560089111329, 0.009265151977539063, 0.009157631874084473, 0.009341376304626465, 0.009555808067321777, 0.00927030372619629, 0.009237919807434082, 0.009228575706481933, 0.009273344039916993, 0.009148415565490722, 0.009122879981994628, 0.00915552043914795, 0.009129983901977539, 0.009211647987365722, 0.009545984268188477, 0.009383935928344727, 0.009209407806396484, 0.009247167587280274, 0.009224191665649414, 0.009240575790405273, 0.009265151977539063, 0.009203712463378906, 0.009187328338623046, 0.009125663757324218, 0.009158592224121093, 0.00923801612854004, 0.009235487937927247, 0.009148447990417481, 0.009144191741943359, 0.009176095962524415, 0.009222208023071289, 0.009261823654174805, 0.00922544002532959, 0.009201791763305664, 0.00923305606842041, 0.009265151977539063, 0.009445376396179199, 0.009269503593444825, 0.009272800445556641, 0.009285856246948242, 0.009344896316528321, 0.009212096214294434, 0.00928809642791748, 0.00932636833190918, 0.00935097599029541, 0.00931715202331543, 0.009386240005493164, 0.00927455997467041, 0.00927609634399414, 0.009412544250488281, 0.009304448127746582, 0.009311967849731446, 0.009277471542358398, 0.009259008407592773, 0.00920911979675293, 0.009237215995788575, 0.009269248008728028, 0.009244671821594238, 0.009191360473632813, 0.00914851188659668, 0.009119744300842286, 0.009146207809448243, 0.00909939193725586, 0.009143648147583007, 0.009107456207275391, 0.009171615600585938, 0.009170944213867188, 0.009129664421081544, 0.00916652774810791, 0.009194111824035645, 0.009184479713439941, 0.009255711555480956, 0.009234335899353028, 0.009171327590942384, 0.009793248176574707, 0.009351424217224121, 0.009511967658996583, 0.009210880279541016, 0.009209600448608399, 0.009166239738464355, 0.009149344444274902, 0.009198464393615723, 0.009142848014831543, 0.009097439765930176, 0.009168895721435547, 0.009140416145324707, 0.009142208099365234, 0.009113471984863281, 0.009154560089111329, 0.009138143539428711, 0.009145855903625488, 0.009169407844543457, 0.009099424362182617, 0.009154111862182616, 0.009095487594604492, 0.009158656120300293, 0.009150464057922364, 0.009158143997192383, 0.009187840461730956, 0.009178655624389648, 0.009157088279724121, 0.00912179183959961, 0.009179136276245118, 0.009312255859375, 0.009215999603271484, 0.009751328468322754, 0.009713760375976562, 0.009693087577819825, 0.009551520347595216, 0.009419103622436524, 0.009381855964660644, 0.009269280433654785, 0.009273152351379394, 0.009279232025146484, 0.009269087791442872, 0.009224639892578126, 0.009304224014282226, 0.009250016212463378, 0.009191583633422852, 0.009243264198303222, 0.009254303932189942, 0.009267680168151856, 0.009222271919250489, 0.00920588779449463, 0.009324416160583496, 0.00924403190612793, 0.00926540756225586, 0.009294207572937013, 0.00923801612854004, 0.009228832244873047, 0.00925487995147705, 0.009242719650268554, 0.009234335899353028, 0.00925209617614746, 0.009210623741149903, 0.00919046401977539, 0.009161664009094237, 0.00916431999206543, 0.009185759544372559, 0.009242464065551757, 0.009189536094665527, 0.009176959991455078, 0.009193599700927734, 0.009207807540893554, 0.009163840293884278, 0.009191360473632813, 0.009205984115600587, 0.009159456253051758, 0.009223615646362305, 0.009206080436706543, 0.009257216453552247, 0.009215680122375489, 0.00917478370666504, 0.009219776153564454, 0.009177984237670899, 0.009205023765563964, 0.009197407722473145, 0.009150527954101562, 0.009134528160095214, 0.009118047714233399, 0.00912816047668457, 0.009236384391784667, 0.009197471618652343, 0.009156864166259765, 0.009123295783996581, 0.009170656204223632, 0.009113183975219727, 0.009179743766784668, 0.00901696014404297, 0.009184800148010253, 0.009143136024475097, 0.009207039833068848, 0.00918115234375, 0.009212767601013184, 0.00915555191040039, 0.00948147201538086, 0.009291328430175782, 0.00942300796508789, 0.009319647789001465, 0.009251615524291993, 0.009183232307434081, 0.009215999603271484, 0.009227840423583985, 0.00917734432220459, 0.00919161605834961, 0.009148415565490722, 0.009116703987121583, 0.009092063903808593, 0.009166848182678223, 0.009156864166259765, 0.009148192405700684, 0.009189408302307129, 0.00909500789642334, 0.009139328002929688, 0.009126367568969727, 0.009171263694763183, 0.009146559715270995, 0.009132384300231933, 0.00914406394958496, 0.009091168403625489, 0.009164608001708985, 0.009154560089111329, 0.009168895721435547, 0.009141568183898927, 0.009135968208312988, 0.009179807662963868, 0.0091342716217041, 0.009101311683654785, 0.009111552238464356, 0.009136128425598144, 0.009098655700683593, 0.009124383926391602, 0.009152128219604492, 0.009093791961669921, 0.009149248123168946, 0.009151455879211425, 0.009194751739501954, 0.009193535804748534, 0.009228992462158202, 0.009351231575012208, 0.009172927856445313, 0.009326560020446777, 0.009420063972473145, 0.00916966438293457, 0.009134079933166504, 0.009134079933166504, 0.009166848182678223, 0.009180992126464844, 0.009150112152099609, 0.009140768051147462, 0.009155712127685546]",tokens/s,107.68093284907691,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1545.949184,1823.408128,0.0,1428.160512,1322.516992,s,1,8.6151826171875,8.6151826171875,0.0,8.6151826171875,8.6151826171875,8.6151826171875,8.6151826171875,[8.6151826171875],,kWh,4.841402075417136e-05,5.333193352905008e-06,1.8674737161994415e-05,7.242195126907079e-05,,MB,1518.239744,1842.282496,0.0,1434.451968,1322.072064,s,10,5.334609802246093,0.5334609802246093,0.001341341049820948,0.533219970703125,0.5344022033691406,0.5357007049560547,0.536739506225586,"[0.5369992065429687, 0.5320140380859375, 0.533542236328125, 0.5341136474609375, 0.5322193603515625, 0.532897705078125, 0.5327489013671876, 0.5337323608398438, 0.5335829467773437, 0.5327593994140625]",tokens/s,479.88514528693986,kWh,1.5942756853948133e-05,1.7576546860210916e-06,1.0655330161684307e-05,2.8355741701653534e-05,tokens/kWh,9028153.89890054,MB,1523.48672,1842.282496,0.0,1434.451968,1374.923264,s,10,10.940359008789061,1.094035900878906,0.01615004752358528,1.0859066162109374,1.1225823608398438,1.124010968017578,1.1251538537597656,"[1.08465234375, 1.0914818115234375, 1.08587841796875, 1.085934814453125, 1.08191357421875, 1.0795816650390626, 1.1023470458984375, 1.0808648681640625, 1.1254395751953126, 1.122264892578125]",tokens/s,57.58494757748648,kWh,3.2475488793968615e-05,3.5824973199971697e-06,1.7150303193914406e-05,5.3208289307880186e-05,tokens/kWh,1184026.038414087,,s,630,10.93764793968201,0.01736134593600318,0.0004135646128487301,0.017225695610046388,0.017894233131408692,0.01801259527206421,0.018358398532867432,"[0.01812284851074219, 0.017483776092529296, 0.017249311447143555, 0.017118175506591796, 0.017028160095214843, 0.01730636787414551, 0.017556991577148438, 0.01759097671508789, 0.01823311996459961, 0.017380800247192383, 0.017294111251831053, 0.017173856735229493, 0.01725916862487793, 0.017096704483032226, 0.017145856857299805, 0.017160192489624023, 0.01706598472595215, 0.0176507511138916, 0.01714787292480469, 0.017175519943237304, 0.017143808364868163, 0.017043455123901367, 0.017073568344116212, 0.017077856063842774, 0.016966655731201173, 0.016965631484985352, 0.01702707290649414, 0.016947200775146484, 0.017047872543334962, 0.016972576141357422, 0.017046432495117187, 0.01697996711730957, 0.017006591796875, 0.017032255172729494, 0.016996992111206054, 0.017006912231445313, 0.016983200073242187, 0.01702537536621094, 0.017084928512573243, 0.017035263061523438, 0.017145856857299805, 0.017051647186279297, 0.01701683235168457, 0.01700454330444336, 0.017063232421875, 0.017040063858032226, 0.017152000427246093, 0.01700864028930664, 0.0170098876953125, 0.017062463760375977, 0.01715407943725586, 0.017057151794433595, 0.017050432205200194, 0.019209535598754882, 0.01836716842651367, 0.017218656539916992, 0.017159072875976563, 0.017135616302490234, 0.017172479629516603, 0.017006591796875, 0.017224992752075195, 0.01708720016479492, 0.017307647705078123, 0.01833692741394043, 0.017773408889770508, 0.01769264030456543, 0.01767628860473633, 0.01792140769958496, 0.01785513687133789, 0.017717248916625978, 0.01778278350830078, 0.017794912338256835, 0.017631391525268554, 0.017579679489135743, 0.017453407287597655, 0.017270784378051757, 0.01702707290649414, 0.017131519317626954, 0.017132959365844726, 0.017037792205810545, 0.017334400177001955, 0.017100799560546876, 0.017199199676513673, 0.01723161506652832, 0.017213056564331055, 0.017185184478759767, 0.01722995185852051, 0.01724006462097168, 0.017094655990600584, 0.017358495712280275, 0.017461599349975584, 0.017292352676391603, 0.01710995292663574, 0.017766368865966796, 0.017072160720825194, 0.01704140853881836, 0.017104000091552735, 0.01703001594543457, 0.016955392837524414, 0.01698406410217285, 0.017072128295898437, 0.01721343994140625, 0.01724790382385254, 0.01786899185180664, 0.01728060722351074, 0.017236576080322266, 0.017178592681884767, 0.017159616470336914, 0.017332799911499025, 0.01722572708129883, 0.017262624740600585, 0.01726051139831543, 0.01724006462097168, 0.01737068748474121, 0.017318592071533204, 0.017329919815063478, 0.017242111206054688, 0.01727004814147949, 0.01717692756652832, 0.017249792098999024, 0.017230655670166014, 0.01715558433532715, 0.01707241630554199, 0.01708880043029785, 0.01711427116394043, 0.01720355224609375, 0.018409183502197266, 0.017785120010375976, 0.017479679107666016, 0.01756979179382324, 0.0176680965423584, 0.017510719299316406, 0.01745814323425293, 0.017219423294067383, 0.01713417625427246, 0.017258783340454102, 0.017102848052978514, 0.017423519134521483, 0.017380191802978517, 0.017258495330810548, 0.01781760025024414, 0.017262592315673828, 0.017029184341430664, 0.017031103134155273, 0.01731123161315918, 0.017326143264770506, 0.017048255920410156, 0.017011871337890627, 0.01704377555847168, 0.01710108757019043, 0.017130975723266603, 0.01707676887512207, 0.017283071517944337, 0.017094655990600584, 0.017014720916748046, 0.01696076774597168, 0.017211711883544922, 0.0173941764831543, 0.01763475227355957, 0.017416767120361328, 0.01722163200378418, 0.017063936233520507, 0.01703321647644043, 0.017026208877563478, 0.017216352462768553, 0.01712335968017578, 0.017219551086425783, 0.017161600112915038, 0.017171072006225585, 0.01717452812194824, 0.01716864013671875, 0.01706572723388672, 0.01717862319946289, 0.017014240264892577, 0.016929056167602537, 0.016888063430786134, 0.017035263061523438, 0.016936960220336913, 0.017091615676879883, 0.017036256790161134, 0.017051647186279297, 0.01705369567871094, 0.017074176788330078, 0.01727039909362793, 0.017213823318481446, 0.01731283187866211, 0.01735321617126465, 0.017324480056762695, 0.017338367462158204, 0.021570175170898438, 0.017733631134033204, 0.01795686340332031, 0.017459199905395507, 0.017342464447021484, 0.017377279281616212, 0.017207456588745118, 0.017417119979858398, 0.01730860710144043, 0.01722947120666504, 0.01746774482727051, 0.017252351760864256, 0.01722777557373047, 0.017201152801513672, 0.01724323272705078, 0.017308416366577147, 0.01728937530517578, 0.01817804718017578, 0.017291040420532228, 0.01721571159362793, 0.01710908889770508, 0.01716534423828125, 0.016999103546142577, 0.017014303207397462, 0.017086944580078124, 0.017071935653686525, 0.017179008483886718, 0.017055744171142577, 0.017104415893554686, 0.017050079345703125, 0.01724569511413574, 0.01704966354370117, 0.01698252868652344, 0.01695737648010254, 0.01700454330444336, 0.016953344345092772, 0.016975616455078123, 0.01699456024169922, 0.017022975921630858, 0.017093727111816406, 0.01710812759399414, 0.017041088104248047, 0.017192703247070312, 0.017135711669921876, 0.01727724838256836, 0.01722972869873047, 0.017113311767578125, 0.01699203109741211, 0.017092191696166992, 0.017023391723632812, 0.01698726463317871, 0.016946048736572267, 0.017045568466186524, 0.016957536697387695, 0.017009599685668945, 0.01705049514770508, 0.01711516761779785, 0.017121280670166016, 0.016995552062988282, 0.01695414352416992, 0.016977920532226562, 0.01700044822692871, 0.01694713592529297, 0.018298112869262695, 0.01759644889831543, 0.01751318359375, 0.017565696716308594, 0.017336320877075196, 0.01738751983642578, 0.017330175399780275, 0.01722777557373047, 0.017392831802368162, 0.01742313575744629, 0.017325151443481446, 0.01734752082824707, 0.01743257522583008, 0.017238016128540038, 0.017370975494384766, 0.017217567443847656, 0.017180671691894533, 0.01700876808166504, 0.01700044822692871, 0.017001855850219728, 0.016933504104614257, 0.01699430465698242, 0.016982015609741212, 0.017063360214233398, 0.017119808197021483, 0.017112064361572265, 0.01705881690979004, 0.016999584197998047, 0.016948095321655275, 0.01690943908691406, 0.016984928131103517, 0.017058944702148436, 0.017131744384765626, 0.01711516761779785, 0.01711372756958008, 0.01707766342163086, 0.01697865676879883, 0.01717571258544922, 0.017005279541015626, 0.01691164779663086, 0.016885856628417968, 0.017007232666015625, 0.017253759384155274, 0.017316255569458008, 0.01741360092163086, 0.017117952346801756, 0.016973823547363282, 0.016910335540771485, 0.016945152282714843, 0.01696371269226074, 0.017000608444213867, 0.016969791412353517, 0.01688972854614258, 0.01686854362487793, 0.017015392303466798, 0.016914432525634765, 0.016924671173095703, 0.017074176788330078, 0.018288415908813478, 0.017115232467651367, 0.017847583770751952, 0.01704431915283203, 0.017067136764526366, 0.018201471328735352, 0.017452224731445313, 0.017306432723999024, 0.017717248916625978, 0.017092607498168946, 0.01701068878173828, 0.017047552108764647, 0.016920576095581053, 0.017178304672241212, 0.016949567794799805, 0.01697996711730957, 0.01710393524169922, 0.017081279754638672, 0.01703321647644043, 0.017121280670166016, 0.017111040115356444, 0.017111263275146484, 0.017043231964111328, 0.017084415435791016, 0.017036767959594728, 0.01701481628417969, 0.017041919708251953, 0.016946239471435545, 0.0169498233795166, 0.017327680587768554, 0.017023807525634767, 0.017006591796875, 0.01715814399719238, 0.017192384719848634, 0.01718329620361328, 0.017311744689941407, 0.01723311996459961, 0.017242528915405272, 0.017257055282592772, 0.017304672241210937, 0.01727110481262207, 0.017250688552856445, 0.01722163200378418, 0.017172479629516603, 0.017233728408813476, 0.0171167049407959, 0.01711497688293457, 0.01698054313659668, 0.016908544540405274, 0.016951295852661134, 0.01711836814880371, 0.016974687576293945, 0.016932863235473633, 0.01708028793334961, 0.017059423446655272, 0.017103296279907226, 0.01705286407470703, 0.016988128662109376, 0.016960351943969727, 0.016959423065185546, 0.01692073631286621, 0.017088415145874024, 0.017235008239746094, 0.017169599533081056, 0.017178367614746094, 0.017303552627563477, 0.017143552780151367, 0.017084672927856447, 0.01813007926940918, 0.01750307273864746, 0.017317888259887695, 0.01739571189880371, 0.01737113571166992, 0.01758185577392578, 0.017690752029418947, 0.017667871475219726, 0.017805055618286134, 0.01776896095275879, 0.017817567825317383, 0.017747167587280274, 0.01780415916442871, 0.017803232192993165, 0.017678367614746095, 0.017741823196411134, 0.01781475257873535, 0.01777743911743164, 0.01776870346069336, 0.017792768478393554, 0.017741823196411134, 0.01763532829284668, 0.017695903778076172, 0.017642335891723634, 0.017657888412475585, 0.018129024505615234, 0.01802009582519531, 0.017889503479003907, 0.017905120849609376, 0.017615264892578125, 0.017682432174682617, 0.01737113571166992, 0.017383232116699218, 0.017840320587158204, 0.017311744689941407, 0.017192928314208985, 0.017225664138793947, 0.0170897274017334, 0.01700752067565918, 0.016956672668457032, 0.016986047744750977, 0.016982175827026366, 0.017199775695800782, 0.017051807403564455, 0.017047199249267578, 0.017031360626220703, 0.017034591674804686, 0.017154912948608398, 0.017232927322387695, 0.017121248245239258, 0.017238847732543944, 0.017229824066162108, 0.017218656539916992, 0.01734079933166504, 0.017328672409057617, 0.017372671127319335, 0.017408512115478517, 0.017536863327026368, 0.017283103942871095, 0.01755763244628906, 0.01763942337036133, 0.017640607833862305, 0.01747849655151367, 0.018118656158447266, 0.017305503845214842, 0.017150047302246094, 0.017086528778076173, 0.017245664596557617, 0.017340896606445312, 0.017317888259887695, 0.017104576110839844, 0.017135583877563476, 0.01703887939453125, 0.017034048080444335, 0.017121023178100585, 0.017080575942993163, 0.016987520217895506, 0.017076576232910156, 0.016980255126953125, 0.016932863235473633, 0.016973823547363282, 0.017031167984008787, 0.017082368850708008, 0.01703321647644043, 0.016916479110717773, 0.017111040115356444, 0.01720524787902832, 0.01716758346557617, 0.01716281509399414, 0.017239519119262695, 0.017084192276000977, 0.017017631530761718, 0.017008544921875, 0.017053983688354493, 0.01699635124206543, 0.0170250244140625, 0.017203296661376953, 0.017232864379882813, 0.017206207275390625, 0.01707827186584473, 0.01742198371887207, 0.01733580780029297, 0.017378368377685548, 0.017264415740966797, 0.01721343994140625, 0.0172126407623291, 0.017117216110229493, 0.0170927677154541, 0.017192832946777342, 0.0170296630859375, 0.017179935455322266, 0.017070720672607422, 0.017043424606323243, 0.017121599197387694, 0.017342464447021484, 0.017284320831298827, 0.01718966484069824, 0.01720319938659668, 0.017099775314331055, 0.017107967376708985, 0.017250303268432618, 0.017082368850708008, 0.01705187225341797, 0.017102624893188475, 0.017169952392578125, 0.01717910385131836, 0.018556512832641602, 0.01809654426574707, 0.01800601577758789, 0.017975296020507812, 0.01801420783996582, 0.017893375396728514, 0.017985536575317384, 0.017909151077270508, 0.017930816650390625, 0.01794175910949707, 0.01825811195373535, 0.017934207916259767, 0.01795964813232422, 0.01785241508483887, 0.017903615951538086, 0.017845279693603517, 0.017879295349121093, 0.017804128646850586, 0.017950592041015626, 0.01787494468688965, 0.018028543472290038, 0.01796623992919922, 0.01795907211303711, 0.01800262451171875, 0.017886592864990235, 0.017793664932250975, 0.01783193588256836, 0.017868703842163085, 0.017997888565063475, 0.01788870429992676, 0.018012767791748048, 0.017853887557983398, 0.017840703964233397, 0.017977312088012697, 0.01805036735534668, 0.018016992568969728, 0.018045984268188476, 0.017994144439697265, 0.018055423736572266, 0.017959232330322265, 0.017934335708618163, 0.017737728118896484, 0.017829248428344727, 0.01786534309387207, 0.017835391998291015, 0.017848960876464842, 0.017741823196411134, 0.017679807662963867, 0.01758633613586426, 0.017600128173828125, 0.017599264144897462, 0.017582080841064454, 0.017675680160522463, 0.017559648513793946, 0.017639936447143553, 0.017515615463256837, 0.01755228805541992, 0.01767033576965332, 0.01757574462890625, 0.01759846305847168, 0.017682687759399414, 0.01764246368408203, 0.017578784942626952, 0.01811020851135254, 0.017701568603515624, 0.017364992141723632, 0.017335359573364257, 0.01726969528198242, 0.01728102493286133, 0.017320032119750976, 0.017170112609863283, 0.01725667190551758, 0.01718230438232422, 0.017237535476684572, 0.01722457695007324, 0.017238016128540038, 0.01778396797180176, 0.019261503219604494, 0.01764579200744629, 0.01768227195739746, 0.01780614471435547, 0.017790271759033204, 0.017779232025146485, 0.017788991928100586, 0.01785856056213379, 0.017796192169189453, 0.017875680923461913, 0.017831647872924804, 0.01779964828491211, 0.01777663993835449, 0.017844224929809572, 0.017854463577270507, 0.01778483200073242, 0.017810592651367186, 0.019446239471435547, 0.017901952743530274, 0.01785651206970215, 0.017811456680297853, 0.017954816818237306, 0.017703136444091796, 0.01785385513305664, 0.017858943939208984, 0.017739776611328126, 0.017725439071655275, 0.01772742462158203, 0.017874399185180665, 0.01791756820678711, 0.017943647384643553, 0.017907583236694335, 0.017876991271972655, 0.01786476707458496, 0.017964736938476562, 0.017813760757446288, 0.017817312240600586, 0.017809535980224608, 0.01780137634277344, 0.01787494468688965, 0.017987583160400392, 0.017864704132080078, 0.018001920700073244, 0.018147327423095702, 0.018060543060302733, 0.018118528366088866, 0.01810918426513672, 0.01801228713989258, 0.01801238441467285]",tokens/s,57.599220917903885,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1844.895744,2861.498368,0.0,2466.250752,2401.696256,s,1,9.251365234375,9.251365234375,0.0,9.251365234375,9.251365234375,9.251365234375,9.251365234375,[9.251365234375],,kWh,6.886548576247454e-05,7.588989187789802e-06,2.5445020356035242e-05,0.00010189949530629958,,MB,1858.654208,3087.990784,0.0,2680.160256,2582.173696,s,10,5.429983032226562,0.5429983032226563,0.002382280051480081,0.5424951171875,0.5447377258300781,0.546928646850586,0.5486813836669922,"[0.5491195678710937, 0.5414999389648437, 0.53960498046875, 0.5442508544921875, 0.5422152099609375, 0.5430211791992188, 0.5424531860351562, 0.5425370483398437, 0.5415059814453125, 0.5437750854492187]",tokens/s,471.45635351097457,kWh,1.623639444745482e-05,1.7905768134000214e-06,1.0804900619222778e-05,2.883187188007762e-05,tokens/kWh,8879062.763069924,MB,1858.654208,3087.990784,0.0,2680.160256,2582.176256,s,10,13.049833618164062,1.3049833618164064,0.007199874367472601,1.3030529174804688,1.314584814453125,1.3170796997070313,1.3190756079101562,"[1.3195745849609375, 1.3140303955078125, 1.2986925048828124, 1.3074959716796875, 1.2964578857421876, 1.302795654296875, 1.3090628662109376, 1.3018612060546875, 1.2965523681640625, 1.3033101806640626]",tokens/s,48.27647757310124,kWh,3.8289459017962514e-05,4.223309476993651e-06,2.0908124751180103e-05,6.342089324613628e-05,tokens/kWh,993363.4923036674,,s,630,13.047791374206549,0.020710779959058005,0.0004612778528481184,0.020622047424316406,0.02097887725830078,0.021442092704772948,0.022525767879486087,"[0.02205936050415039, 0.02130191993713379, 0.0210729923248291, 0.021205503463745116, 0.02084262466430664, 0.020707391738891603, 0.020690143585205076, 0.02068355178833008, 0.02060492706298828, 0.02081996726989746, 0.020704383850097655, 0.020716415405273438, 0.02071347236633301, 0.020641311645507813, 0.02115836715698242, 0.02083635139465332, 0.020979711532592774, 0.021147872924804686, 0.02109347152709961, 0.021140159606933592, 0.020791296005249024, 0.02077270317077637, 0.020770303726196288, 0.020738624572753907, 0.0208787841796875, 0.020781728744506837, 0.020749631881713866, 0.020687551498413087, 0.020733951568603515, 0.020782527923583986, 0.020783327102661134, 0.020899551391601563, 0.020688735961914062, 0.02077788734436035, 0.02072153663635254, 0.020682207107543947, 0.021793312072753906, 0.022495231628417968, 0.02146713638305664, 0.02076176071166992, 0.020931423187255858, 0.020862079620361327, 0.020771583557128905, 0.02112227249145508, 0.020870048522949217, 0.020733951568603515, 0.020717567443847656, 0.020809471130371095, 0.02080384063720703, 0.02066201591491699, 0.020832063674926758, 0.02071187210083008, 0.020708383560180663, 0.020844512939453125, 0.020714496612548827, 0.020742143630981445, 0.02068070411682129, 0.02253824043273926, 0.022119775772094726, 0.020828128814697266, 0.0207891845703125, 0.020785823822021484, 0.02090617561340332, 0.02188924789428711, 0.020978784561157225, 0.02098883247375488, 0.020808895111083983, 0.02162361526489258, 0.022987936019897463, 0.02112188720703125, 0.021212671279907228, 0.02082252883911133, 0.020618976593017577, 0.02062099266052246, 0.02070742416381836, 0.020662784576416016, 0.020614656448364257, 0.020677120208740234, 0.02073948860168457, 0.02077142333984375, 0.02085478401184082, 0.020776960372924806, 0.020789247512817383, 0.020839487075805664, 0.020699520111083985, 0.02077110481262207, 0.020859167098999022, 0.020840320587158203, 0.020883583068847657, 0.02127667236328125, 0.02088140869140625, 0.020738048553466795, 0.020716863632202147, 0.020605056762695313, 0.020523584365844727, 0.020535295486450195, 0.02047567939758301, 0.021408000946044923, 0.022445535659790038, 0.020861440658569336, 0.021606399536132814, 0.020637855529785157, 0.020483104705810547, 0.02046441650390625, 0.020743743896484375, 0.02053779220581055, 0.020682271957397462, 0.0206115837097168, 0.020577375411987304, 0.020626335144042968, 0.02070230484008789, 0.02052182388305664, 0.020737375259399414, 0.021017311096191406, 0.020914176940917968, 0.02062131118774414, 0.02064588737487793, 0.0206777286529541, 0.020788127899169923, 0.020744192123413087, 0.0209039363861084, 0.020939008712768555, 0.020622783660888672, 0.020433216094970702, 0.020446271896362306, 0.020526016235351562, 0.02182143974304199, 0.02087222480773926, 0.020816864013671874, 0.020650272369384767, 0.020520671844482422, 0.020496383666992187, 0.020336639404296874, 0.02041004753112793, 0.0204039363861084, 0.020666976928710938, 0.020607040405273436, 0.020488128662109376, 0.020667455673217774, 0.020908992767333986, 0.020723712921142577, 0.020678016662597658, 0.020973695755004882, 0.020863168716430663, 0.020729663848876954, 0.02070992088317871, 0.02063680076599121, 0.020724576950073244, 0.020692991256713866, 0.020723520278930666, 0.02060857582092285, 0.02079529571533203, 0.020669151306152343, 0.02065203285217285, 0.020588544845581053, 0.020977664947509765, 0.02045747184753418, 0.02049843215942383, 0.02046073532104492, 0.020501184463500976, 0.02098361587524414, 0.020594783782958984, 0.02054956817626953, 0.02057040023803711, 0.020471807479858398, 0.020506528854370116, 0.020545215606689454, 0.020469568252563478, 0.02040687942504883, 0.02047590446472168, 0.02047920036315918, 0.020470560073852537, 0.02044623947143555, 0.0204932804107666, 0.02054083251953125, 0.020707935333251954, 0.020821855545043944, 0.02100239944458008, 0.02059878349304199, 0.020367328643798827, 0.020395168304443358, 0.020439935684204102, 0.02038374328613281, 0.0204716796875, 0.020453119277954103, 0.020382080078125, 0.020348928451538087, 0.020383487701416014, 0.02041200065612793, 0.02153654479980469, 0.020964351654052735, 0.020740095138549804, 0.02082815933227539, 0.02088153648376465, 0.02069286346435547, 0.02072310447692871, 0.020707168579101563, 0.020943616867065428, 0.020776960372924806, 0.020879167556762696, 0.02073619270324707, 0.020754432678222655, 0.02059996795654297, 0.02053558349609375, 0.020539487838745117, 0.020511199951171875, 0.0206429443359375, 0.020590560913085938, 0.020583328247070314, 0.020588544845581053, 0.02059836769104004, 0.02058255958557129, 0.02052534484863281, 0.020594655990600588, 0.020455423355102538, 0.020510719299316405, 0.020588544845581053, 0.020538496017456054, 0.020554527282714844, 0.020566112518310548, 0.020495967864990236, 0.020527551651000977, 0.020520320892333986, 0.021537376403808595, 0.02082611274719238, 0.020749952316284178, 0.020796031951904298, 0.020673343658447266, 0.02073846435546875, 0.020716127395629884, 0.020893632888793947, 0.020791296005249024, 0.020717567443847656, 0.0206561279296875, 0.020679872512817384, 0.020903743743896485, 0.020738367080688477, 0.02059129524230957, 0.02062131118774414, 0.02058563232421875, 0.020605791091918946, 0.020579967498779297, 0.020566400527954103, 0.020576000213623047, 0.020581727981567384, 0.020478752136230467, 0.020533376693725586, 0.020496383666992187, 0.02036083221435547, 0.02043497657775879, 0.02038979148864746, 0.025665151596069337, 0.022013952255249023, 0.021399551391601563, 0.021126815795898438, 0.020969472885131835, 0.020926464080810548, 0.02084899139404297, 0.020651071548461915, 0.02055059242248535, 0.020462848663330077, 0.020687200546264647, 0.02063385581970215, 0.02059596824645996, 0.02060176086425781, 0.020608991622924806, 0.020529184341430664, 0.02042470359802246, 0.020372831344604492, 0.02042947196960449, 0.020368736267089845, 0.020406944274902344, 0.02042464065551758, 0.020361215591430663, 0.02039200019836426, 0.020396032333374024, 0.02042265510559082, 0.020459007263183594, 0.020507135391235352, 0.020563968658447264, 0.020688608169555665, 0.02056787109375, 0.020424768447875975, 0.02051728057861328, 0.020451391220092773, 0.02036479949951172, 0.020477760314941407, 0.02258188819885254, 0.02088256072998047, 0.020593536376953124, 0.020393983840942383, 0.020307296752929686, 0.02029542350769043, 0.02048847961425781, 0.020316799163818358, 0.02049542427062988, 0.02035807991027832, 0.02046156883239746, 0.020295839309692382, 0.020372320175170898, 0.02031718444824219, 0.020588544845581053, 0.02025062370300293, 0.020309343338012695, 0.020373247146606446, 0.02027369689941406, 0.02038617515563965, 0.02041164779663086, 0.020351743698120116, 0.020387712478637694, 0.02047942352294922, 0.02056867218017578, 0.02051081657409668, 0.020750335693359375, 0.020776832580566406, 0.02162892723083496, 0.021072927474975585, 0.021029855728149412, 0.020744192123413087, 0.020564992904663085, 0.020435136795043947, 0.02033132743835449, 0.02039743995666504, 0.02082809638977051, 0.020497087478637696, 0.020475967407226563, 0.020491424560546874, 0.020607776641845703, 0.020387104034423828, 0.020316064834594725, 0.02202908706665039, 0.020552928924560548, 0.020548416137695313, 0.0206376953125, 0.020625408172607423, 0.02065407943725586, 0.020774911880493165, 0.020785120010375975, 0.020705312728881837, 0.0208035831451416, 0.020750335693359375, 0.02069868850708008, 0.020764671325683593, 0.020810176849365234, 0.020714784622192384, 0.020520671844482422, 0.020513792037963868, 0.020533248901367186, 0.02060655975341797, 0.020658655166625975, 0.020651584625244142, 0.02086515235900879, 0.02065177536010742, 0.02061568069458008, 0.02065203285217285, 0.020645471572875978, 0.02050079917907715, 0.02045961570739746, 0.020616384506225587, 0.020578496932983397, 0.020523231506347658, 0.020607391357421876, 0.020690208435058595, 0.020619680404663086, 0.020506784439086913, 0.0207238712310791, 0.02129644775390625, 0.020973663330078125, 0.020656736373901367, 0.02048975944519043, 0.020559520721435548, 0.020486976623535155, 0.020783103942871094, 0.020602880477905275, 0.02073097610473633, 0.02056048011779785, 0.02050899124145508, 0.020547584533691408, 0.02179088020324707, 0.02084864044189453, 0.020719615936279297, 0.020679935455322266, 0.020574975967407226, 0.022984703063964843, 0.021057024002075195, 0.020564640045166015, 0.020538400650024415, 0.020937536239624025, 0.02062745666503906, 0.020725727081298828, 0.02073193550109863, 0.020703231811523438, 0.020665695190429687, 0.0207663688659668, 0.021684959411621095, 0.02074448013305664, 0.02064793586730957, 0.02082611274719238, 0.023066272735595705, 0.02162723159790039, 0.020570112228393556, 0.020393407821655274, 0.020533023834228517, 0.020554527282714844, 0.020555231094360353, 0.020584800720214843, 0.020512960433959962, 0.020444543838500976, 0.02049078369140625, 0.020389984130859375, 0.02042780876159668, 0.020466272354125976, 0.02045580863952637, 0.020411840438842775, 0.02044780731201172, 0.02049225616455078, 0.020460704803466796, 0.020469919204711914, 0.020566368103027345, 0.020619199752807616, 0.02081964874267578, 0.020767648696899413, 0.020801376342773438, 0.020922367095947265, 0.020720672607421876, 0.020690944671630858, 0.020804576873779297, 0.02067865562438965, 0.02072166442871094, 0.020731903076171874, 0.020731903076171874, 0.020888799667358397, 0.020808479309082032, 0.02074131202697754, 0.02067683219909668, 0.02060102462768555, 0.020653600692749022, 0.020577119827270507, 0.020770847320556642, 0.020709280014038087, 0.020671712875366212, 0.02193484878540039, 0.02124736022949219, 0.021462656021118163, 0.02186332893371582, 0.02097350311279297, 0.020758687973022463, 0.020542816162109376, 0.02044380760192871, 0.02047385597229004, 0.020797216415405273, 0.02066044807434082, 0.02068070411682129, 0.020508159637451173, 0.020434688568115235, 0.020429567337036134, 0.020409887313842773, 0.020506591796875, 0.02037727928161621, 0.020411199569702148, 0.020477951049804686, 0.02050668716430664, 0.02054547119140625, 0.020530752182006836, 0.02215318489074707, 0.02078767967224121, 0.020840448379516603, 0.020781055450439453, 0.020612863540649413, 0.020627712249755858, 0.02044483184814453, 0.020744064331054687, 0.020627935409545897, 0.020520191192626953, 0.020638463973999023, 0.020631168365478517, 0.020659936904907226, 0.020652608871459963, 0.02086921691894531, 0.020746240615844725, 0.020665599822998048, 0.020728288650512697, 0.020703104019165038, 0.02072822380065918, 0.020854623794555664, 0.020590784072875977, 0.02049363136291504, 0.02045404815673828, 0.020459232330322267, 0.020475488662719726, 0.020550336837768555, 0.02052297592163086, 0.02047996711730957, 0.02039574432373047, 0.02042300796508789, 0.02037555122375488, 0.02046335983276367, 0.02041468811035156, 0.02041859245300293, 0.02046156883239746, 0.020389888763427736, 0.020402175903320312, 0.02045747184753418, 0.02044313621520996, 0.021608192443847655, 0.020867231369018555, 0.020637407302856445, 0.02058323287963867, 0.02046156883239746, 0.020501535415649415, 0.02044822311401367, 0.02044108772277832, 0.02044108772277832, 0.02051420783996582, 0.020429407119750977, 0.021032960891723632, 0.020477216720581056, 0.020419071197509766, 0.020318431854248045, 0.02047385597229004, 0.020522176742553713, 0.02035139274597168, 0.020558048248291015, 0.020474048614501954, 0.02039129638671875, 0.0205216007232666, 0.020442367553710938, 0.02039596748352051, 0.020507455825805664, 0.02042470359802246, 0.020461055755615236, 0.020395647048950194, 0.02050486373901367, 0.020554336547851562, 0.021661600112915038, 0.020693088531494142, 0.020325536727905272, 0.020403039932250976, 0.020445184707641603, 0.020363103866577147, 0.02036751937866211, 0.020391424179077147, 0.020361984252929687, 0.020495679855346678, 0.02052467155456543, 0.02059753608703613, 0.020733983993530273, 0.020651391983032227, 0.020653823852539062, 0.02067955207824707, 0.02068889617919922, 0.020659744262695314, 0.020693471908569336, 0.020684799194335936, 0.020715167999267578, 0.02061961555480957, 0.020526847839355468, 0.02056163215637207, 0.02050102424621582, 0.020586048126220703, 0.02050092887878418, 0.020594623565673827, 0.0207524471282959, 0.020733951568603515, 0.02067046356201172, 0.02067068862915039, 0.020669216156005858, 0.022004735946655272, 0.021416959762573243, 0.021354496002197267, 0.021351776123046874, 0.02108671951293945, 0.020891807556152345, 0.02092051124572754, 0.02079315185546875, 0.020695039749145508, 0.020822015762329102, 0.02569215965270996, 0.02064384078979492, 0.020544864654541015, 0.020630048751831054, 0.020367071151733397, 0.020423072814941406, 0.02042464065551758, 0.02045529556274414, 0.02042902374267578, 0.020412384033203126, 0.020289535522460937, 0.020338464736938476, 0.02032252883911133, 0.020309823989868164, 0.020387840270996094, 0.02034694480895996, 0.02041254425048828, 0.020510719299316405, 0.02040227127075195, 0.020299007415771484, 0.020472480773925782, 0.02038140869140625, 0.020414751052856447, 0.020518655776977538, 0.020541696548461913, 0.020532800674438478, 0.02067296028137207, 0.020694528579711914, 0.021086336135864258, 0.02120537567138672, 0.020854496002197267, 0.020738208770751953, 0.020715295791625978, 0.020664224624633788, 0.02053494453430176, 0.020542240142822264, 0.020502304077148436, 0.020477855682373047, 0.02045574378967285, 0.02065772819519043, 0.020531328201293945, 0.020451648712158203, 0.020353023529052734, 0.020377344131469726, 0.02043110466003418, 0.020483936309814453, 0.020336448669433595, 0.02039638328552246, 0.020406272888183592, 0.020330495834350586, 0.020387840270996094, 0.020352352142333986, 0.0206342716217041]",tokens/s,48.28403382088191,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4244.074496,4725.866496,0.0,4347.396096,4328.833024,s,1,10.40515625,10.40515625,0.0,10.40515625,10.40515625,10.40515625,10.40515625,[10.40515625],,kWh,9.143228809164157e-05,1.0074670626374124e-05,2.9218912263984098e-05,0.00013072587098199979,,MB,1385.684992,5220.794368,0.0,4812.96384,4756.928512,s,10,3.9759581909179684,0.39759581909179686,0.004339503763737649,0.39798226928710934,0.40114043884277345,0.4012296890258789,0.4013010891723633,"[0.3852103271484375, 0.39959088134765625, 0.3977213439941406, 0.3977471923828125, 0.39811508178710936, 0.3976213684082031, 0.3996629943847656, 0.39784945678710937, 0.40112060546875, 0.40131893920898437]",tokens/s,643.869949600488,kWh,1.1604514901762577e-05,1.2797745318105492e-06,7.739621576307716e-06,2.0623911009880844e-05,tokens/kWh,12412776.60077913,MB,1390.194688,5325.651968,0.0,4917.82144,4876.091904,s,10,24.099277099609377,2.4099277099609377,0.016611488617423925,2.406385498046875,2.435732788085937,2.436072814941406,2.436344836425781,"[2.4356572265625, 2.4194453125, 2.400362548828125, 2.38584423828125, 2.391546875, 2.4111103515625, 2.419173828125, 2.40166064453125, 2.398063232421875, 2.436412841796875]",tokens/s,26.141862986015113,kWh,7.10939256199046e-05,7.841736990095445e-06,4.7047033364089756e-05,0.0001259826959740898,tokens/kWh,500068.67620103067,,s,630,24.096367031097394,0.038248201636662565,0.0005840847493881605,0.03814447975158691,0.03880594062805176,0.03910825805664063,0.04053592147827149,"[0.03988630294799805, 0.03840668869018555, 0.03834470367431641, 0.038246654510498045, 0.03829939270019531, 0.038160385131835936, 0.03832012939453125, 0.038454814910888674, 0.038746593475341796, 0.03876444625854492, 0.038707168579101565, 0.03896742248535156, 0.03877478408813476, 0.0388087043762207, 0.03926105499267578, 0.03917532730102539, 0.0384048957824707, 0.0386519660949707, 0.038948863983154294, 0.0385546875, 0.03867292785644531, 0.03873628616333008, 0.03865804672241211, 0.038645759582519534, 0.038594558715820314, 0.03981052780151367, 0.03885939025878906, 0.038725440979003906, 0.038705249786376954, 0.03861503982543945, 0.0384634895324707, 0.03856700897216797, 0.038832511901855465, 0.03856643295288086, 0.038632705688476564, 0.03870934295654297, 0.03874374389648438, 0.0384950065612793, 0.03887635040283203, 0.03868707275390625, 0.03877750396728516, 0.03865190505981445, 0.03861913681030273, 0.03878297424316406, 0.038648033142089845, 0.03893430328369141, 0.038563838958740236, 0.03856588745117188, 0.03850649642944336, 0.0385904655456543, 0.03837673568725586, 0.03853385543823242, 0.03856304168701172, 0.0385208969116211, 0.038505184173583985, 0.038292545318603516, 0.0384439697265625, 0.03852627182006836, 0.03840480041503906, 0.03856588745117188, 0.038313312530517576, 0.03844339370727539, 0.03874425506591797, 0.04010860824584961, 0.0384571533203125, 0.038365345001220706, 0.038363166809082035, 0.038400001525878906, 0.038043071746826175, 0.03810569763183594, 0.03804313659667969, 0.03813833618164063, 0.03805539321899414, 0.03809302520751953, 0.038185279846191404, 0.03826224136352539, 0.0385497932434082, 0.03872998428344727, 0.0388403205871582, 0.03875632095336914, 0.03999081420898438, 0.03880563354492188, 0.038694782257080076, 0.03879100799560547, 0.03895158386230469, 0.038798721313476565, 0.038906494140625, 0.0387665901184082, 0.03841740798950195, 0.03860329437255859, 0.0385192642211914, 0.03833184051513672, 0.03807289505004883, 0.038085697174072265, 0.03816255950927734, 0.037921600341796875, 0.03835903930664063, 0.038258689880371094, 0.0384901123046875, 0.03983564758300781, 0.03822182464599609, 0.03834265518188477, 0.03842031860351563, 0.03839619064331055, 0.0386712646484375, 0.03830172729492187, 0.03850131225585938, 0.03855513763427734, 0.0381671028137207, 0.03808777618408203, 0.03811568069458008, 0.03807897567749023, 0.038282527923583984, 0.038144577026367185, 0.03824025726318359, 0.037840991973876956, 0.038113601684570314, 0.037668609619140626, 0.03790028762817383, 0.03850569534301758, 0.038466335296630856, 0.03792086410522461, 0.037881343841552735, 0.03870105743408203, 0.03762422561645508, 0.03777667236328125, 0.03981721496582031, 0.038437950134277345, 0.03827727890014648, 0.038071071624755856, 0.03789020919799805, 0.03792844772338867, 0.03787139129638672, 0.03806675338745117, 0.038136863708496095, 0.038061119079589846, 0.037924766540527344, 0.03784198379516602, 0.037829566955566406, 0.03788576126098633, 0.037676288604736326, 0.03774105453491211, 0.0376583023071289, 0.037730175018310545, 0.04069055938720703, 0.03803750228881836, 0.03792822265625, 0.03773430252075195, 0.03785766220092773, 0.03785551834106445, 0.0376668815612793, 0.03769747161865234, 0.03771171188354492, 0.03800428771972656, 0.03832406234741211, 0.03832310485839844, 0.03865766525268555, 0.03998963165283203, 0.0388485107421875, 0.03868822479248047, 0.03837996673583984, 0.038370880126953125, 0.03806057739257813, 0.03812694549560547, 0.0381300163269043, 0.03788627243041992, 0.03785113525390625, 0.037756671905517576, 0.039511489868164065, 0.03910031890869141, 0.03817155075073242, 0.03819456100463867, 0.03834044647216797, 0.037929759979248044, 0.0385986557006836, 0.03774700927734375, 0.03783545684814453, 0.03778867340087891, 0.03760323333740234, 0.03770582580566406, 0.037506462097167965, 0.037450336456298826, 0.03768230438232422, 0.037553024291992185, 0.03746723175048828, 0.03767520141601562, 0.03746480178833008, 0.03764835357666016, 0.037700927734375, 0.03943695831298828, 0.03844927978515625, 0.03794521713256836, 0.03787926483154297, 0.037873985290527344, 0.03775929641723633, 0.03782236862182617, 0.037574142456054685, 0.03777177429199219, 0.037570110321044924, 0.03757686233520508, 0.037717952728271484, 0.03769507217407227, 0.03777817535400391, 0.03784479904174805, 0.037689342498779296, 0.037816513061523435, 0.03791462326049805, 0.0380682258605957, 0.03775692749023438, 0.037760128021240236, 0.03775084686279297, 0.03784377670288086, 0.03768025588989258, 0.03773952102661133, 0.03770259094238281, 0.03772649765014648, 0.03822444915771484, 0.03783292770385742, 0.037830528259277345, 0.037705726623535156, 0.03797196960449219, 0.03779379272460937, 0.0378853759765625, 0.037494495391845704, 0.03773436737060547, 0.03770457458496094, 0.03772825622558594, 0.037654655456542965, 0.03780799865722656, 0.03799039840698242, 0.037820415496826174, 0.03802259063720703, 0.03791452789306641, 0.03785897445678711, 0.03775129699707031, 0.03765913772583008, 0.03763529586791992, 0.038010913848876955, 0.03769184112548828, 0.03780588912963867, 0.03779372787475586, 0.0377960319519043, 0.038187713623046876, 0.03863062286376953, 0.03785776138305664, 0.037992095947265624, 0.03793340682983398, 0.03781964874267578, 0.037904735565185546, 0.03770937728881836, 0.03805987167358398, 0.03819996643066406, 0.04021148681640625, 0.03832924652099609, 0.038699073791503905, 0.03825836944580078, 0.03815667343139648, 0.03794326400756836, 0.03829555130004883, 0.038362430572509765, 0.03782457733154297, 0.03789507293701172, 0.03777097702026367, 0.03764223861694336, 0.03788822555541992, 0.037700927734375, 0.037638622283935545, 0.03766796875, 0.03770185470581055, 0.037595294952392576, 0.03768534469604492, 0.03786383819580078, 0.03818905639648437, 0.03787081527709961, 0.03784991836547852, 0.03778713607788086, 0.03807484817504883, 0.03786956787109375, 0.03766681671142578, 0.03788800048828125, 0.037743614196777346, 0.03772927856445313, 0.03771596908569336, 0.03785881423950195, 0.0384640007019043, 0.038537216186523435, 0.038020896911621097, 0.03811555099487305, 0.03781836700439453, 0.03777347183227539, 0.038047584533691406, 0.03799792098999023, 0.037843616485595706, 0.03809894561767578, 0.03787161636352539, 0.03801200103759766, 0.03791584014892578, 0.03806588745117188, 0.03821529769897461, 0.03797772979736328, 0.037817089080810544, 0.037797889709472655, 0.03780198287963867, 0.03774259185791016, 0.037617664337158206, 0.037566463470458986, 0.03775660705566406, 0.037746593475341796, 0.037681537628173827, 0.03757206344604492, 0.03765919876098633, 0.038626495361328124, 0.03801948928833008, 0.037843456268310545, 0.03787926483154297, 0.039798465728759766, 0.038054206848144534, 0.037996543884277346, 0.03787571334838867, 0.03911475372314453, 0.04053753662109375, 0.03810396957397461, 0.03764188766479492, 0.037623329162597655, 0.03759939193725586, 0.03758931350708008, 0.03749884796142578, 0.03788803100585938, 0.03773974227905273, 0.03757136154174805, 0.03773807907104492, 0.03775539016723633, 0.03795452880859375, 0.038351806640625, 0.03828940963745117, 0.03853107070922852, 0.03836108779907227, 0.0383631362915039, 0.03847782516479492, 0.038090431213378906, 0.03827948760986328, 0.03820451354980469, 0.03813849639892578, 0.03820982360839844, 0.038966815948486326, 0.03832835388183594, 0.03845084762573242, 0.03829993438720703, 0.03825644683837891, 0.03824828720092773, 0.03837628936767578, 0.03817798233032227, 0.03817158508300781, 0.03898518371582031, 0.03894723129272461, 0.038508544921875, 0.03822963333129883, 0.03832617568969727, 0.038614879608154296, 0.038594497680664065, 0.038267936706542965, 0.038340255737304686, 0.03799385452270508, 0.03803798294067383, 0.03807043075561523, 0.03779971313476563, 0.03841215896606445, 0.03844745635986328, 0.038219745635986326, 0.038139232635498045, 0.0380382080078125, 0.037929183959960935, 0.03762969589233398, 0.037812255859375, 0.03777526473999023, 0.0375048942565918, 0.041302433013916014, 0.03823193740844726, 0.04130393600463867, 0.03825017547607422, 0.038338623046875, 0.03788838577270508, 0.03812934494018555, 0.037859649658203126, 0.037795841217041014, 0.03792835235595703, 0.037845600128173826, 0.03811875152587891, 0.03835318374633789, 0.04014899063110351, 0.038982017517089844, 0.03846553421020508, 0.03836435317993164, 0.03831276702880859, 0.03852447891235351, 0.039739585876464846, 0.03888102340698242, 0.042092479705810544, 0.03893731307983399, 0.038783840179443356, 0.0387217903137207, 0.038727455139160157, 0.03894704055786133, 0.03883491134643555, 0.03866812896728516, 0.03850259017944336, 0.038526657104492185, 0.038314048767089846, 0.03835315322875977, 0.03828867340087891, 0.037996543884277346, 0.03847446441650391, 0.03823622512817383, 0.03853113555908203, 0.03848537445068359, 0.038362945556640625, 0.03821779251098633, 0.03850454330444336, 0.03844723129272461, 0.03806864166259766, 0.038118465423583985, 0.03853952026367188, 0.03877503967285156, 0.03843100738525391, 0.037851200103759766, 0.03785481643676758, 0.03780454254150391, 0.037840511322021486, 0.03774006271362305, 0.03775174331665039, 0.03763596725463867, 0.037621055603027344, 0.0377371826171875, 0.03761356735229492, 0.03811894226074219, 0.03829782485961914, 0.03799020767211914, 0.03773894500732422, 0.03769452667236328, 0.03765055847167969, 0.0378171501159668, 0.04027865600585938, 0.03822796630859375, 0.038106590270996096, 0.03833091354370117, 0.03809689712524414, 0.03810508728027344, 0.03808857727050781, 0.03826451110839844, 0.03812524795532227, 0.03823795318603516, 0.03814438247680664, 0.03816716766357422, 0.038061790466308594, 0.03793948745727539, 0.03801497650146484, 0.03798387145996094, 0.03818534469604492, 0.03800038528442383, 0.038000415802001954, 0.037927391052246094, 0.03798339080810547, 0.03793596649169922, 0.03781612777709961, 0.03841046524047852, 0.03817878341674805, 0.03800678253173828, 0.03772787094116211, 0.03782905578613281, 0.03805996704101562, 0.03775078582763672, 0.03801497650146484, 0.037869823455810546, 0.037827903747558594, 0.0376693115234375, 0.03751935958862305, 0.03756851196289063, 0.03771603012084961, 0.03783878326416015, 0.03796131134033203, 0.038262462615966795, 0.0380681266784668, 0.037955902099609376, 0.03813836669921875, 0.03823740768432617, 0.03843939208984375, 0.03843932723999023, 0.038466590881347656, 0.03827382278442383, 0.038983776092529294, 0.03838521575927734, 0.038814144134521486, 0.038449153900146485, 0.038178398132324216, 0.03815190505981445, 0.03814595031738281, 0.038195999145507815, 0.038144001007080076, 0.038202880859375, 0.03798681640625, 0.0378616943359375, 0.037957054138183594, 0.037918785095214846, 0.03773574447631836, 0.04011990356445312, 0.0381192626953125, 0.03823878479003906, 0.03851468658447266, 0.038152065277099606, 0.0378966064453125, 0.038001888275146486, 0.03797804641723633, 0.037839424133300784, 0.03799859237670898, 0.037607425689697264, 0.03766204833984375, 0.03761996841430664, 0.037714305877685546, 0.03780521774291992, 0.03830435180664062, 0.0377305908203125, 0.037806079864501956, 0.03760947036743164, 0.037713920593261716, 0.037736480712890624, 0.037966976165771486, 0.037808544158935545, 0.0377592658996582, 0.03776528167724609, 0.037774463653564454, 0.0377127685546875, 0.03815411376953125, 0.03797846221923828, 0.03833343887329101, 0.038305728912353516, 0.03793094253540039, 0.03780857467651367, 0.037781505584716796, 0.037738239288330075, 0.03768726348876953, 0.037722625732421876, 0.03758720016479492, 0.03773235321044922, 0.03796899032592774, 0.03801603317260742, 0.03845312118530273, 0.03800892639160156, 0.037981182098388674, 0.03815526580810547, 0.03800259017944336, 0.03988838577270508, 0.03797452926635742, 0.038051136016845705, 0.03795833587646484, 0.038109184265136715, 0.03817609786987305, 0.03806684875488281, 0.03793100738525391, 0.03782355117797852, 0.03766982269287109, 0.03876979064941406, 0.03907843017578125, 0.03795929718017578, 0.038951488494873045, 0.03811529541015625, 0.03792300796508789, 0.03901203155517578, 0.03984979248046875, 0.038410430908203126, 0.03827916717529297, 0.03800678253173828, 0.03833651351928711, 0.038473567962646484, 0.03889705657958984, 0.03856870269775391, 0.038585697174072266, 0.03876931381225586, 0.03855081558227539, 0.038214111328125, 0.038356639862060546, 0.038476192474365234, 0.03833875274658203, 0.03835087966918945, 0.03806560134887695, 0.038263294219970705, 0.03831609725952148, 0.03829961776733398, 0.039018497467041016, 0.038407585144042966, 0.03843913650512695, 0.03872943878173828, 0.03829520034790039, 0.03838431930541992, 0.038502464294433596, 0.03838518524169922, 0.03849084854125977, 0.04077363204956055, 0.039215103149414066, 0.03843100738525391, 0.03852873611450195, 0.03837952041625976, 0.03852848052978516, 0.03890230560302734, 0.0385846061706543, 0.03908512115478516, 0.03842838287353516, 0.0384071044921875, 0.03875430297851563, 0.040531967163085936, 0.038850208282470704, 0.038349151611328125, 0.03836723327636719, 0.038381568908691405, 0.038242305755615234, 0.039354366302490236, 0.03835324859619141, 0.03831772613525391, 0.038274368286132815, 0.03828547286987305, 0.04205145645141602, 0.03860137557983399, 0.03838304138183594, 0.03832883071899414, 0.038240478515625, 0.03950732803344727, 0.038975967407226565, 0.03836044692993164, 0.038675071716308594, 0.0391162223815918, 0.03843948745727539]",tokens/s,26.145020084851687,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1308.635136,1083.113472,0.0,704.643072,675.01056,s,1,8.2532353515625,8.2532353515625,0.0,8.2532353515625,8.2532353515625,8.2532353515625,8.2532353515625,[8.2532353515625],,kWh,3.6389911249989386e-05,4.006824042660556e-06,1.103473104999142e-05,5.143146634264136e-05,,MB,1354.149888,1401.880576,0.0,994.050048,942.608384,s,10,0.3598317146301269,0.0359831714630127,0.00035803077395493477,0.03605726432800293,0.036346069717407226,0.036394411277770995,0.03643308452606201,"[0.036157215118408206, 0.03584652709960937, 0.0359681282043457, 0.0363353271484375, 0.036146400451660156, 0.035140865325927736, 0.035953792572021484, 0.036442752838134765, 0.035644161224365235, 0.0361965446472168]",tokens/s,7114.436821199705,kWh,1.1674981185092925e-06,1.2870139494380178e-07,7.71797872860529e-07,2.0679973863136232e-06,tokens/kWh,123791258.9707578,MB,1382.817792,1412.366336,0.0,1004.535808,942.610944,s,10,20.393948486328124,2.0393948486328126,0.01703011907324134,2.038789245605469,2.056566943359375,2.067591943359375,2.076411943359375,"[2.078616943359375, 2.022477294921875, 2.0407437744140626, 2.0478001708984377, 2.01685546875, 2.02565283203125, 2.0399276123046874, 2.03765087890625, 2.0301065673828127, 2.054116943359375]",tokens/s,30.891516688018747,kWh,5.969731842607485e-05,6.584410114205411e-06,2.342149925953884e-05,8.970322779981909e-05,tokens/kWh,702315.8647154841,,s,630,20.391595325469936,0.032367611627730115,0.0011623937798228494,0.03235038375854492,0.03276208305358887,0.03306687450408936,0.03457327724456788,"[0.032945758819580076, 0.03270697784423828, 0.03256467056274414, 0.032557151794433595, 0.03245043182373047, 0.032653919219970705, 0.03277004623413086, 0.03283148956298828, 0.032628734588623046, 0.03257548904418945, 0.03254185485839844, 0.03249641418457031, 0.03252844619750977, 0.03244611358642578, 0.03246236801147461, 0.03246368026733398, 0.03249356842041016, 0.03268198394775391, 0.032573440551757815, 0.032427967071533205, 0.03231955337524414, 0.03232998275756836, 0.032701759338378905, 0.03254316711425781, 0.03232153701782227, 0.03231875228881836, 0.03239737701416016, 0.032473758697509764, 0.03245209503173828, 0.03293833541870117, 0.033382720947265625, 0.03253228759765625, 0.03245264053344726, 0.032514049530029294, 0.03258691024780273, 0.033245601654052735, 0.03255340957641602, 0.032478816986083986, 0.03237081527709961, 0.0324997444152832, 0.032422046661376956, 0.03253667068481445, 0.03251740646362305, 0.032688865661621096, 0.032537662506103515, 0.03286671829223633, 0.033009185791015624, 0.05799628829956055, 0.032523902893066406, 0.03263116836547852, 0.03252975845336914, 0.032328353881835935, 0.032519966125488284, 0.03238140869140625, 0.03269132614135742, 0.03281164932250977, 0.03255295944213867, 0.03281875228881836, 0.03240915298461914, 0.03272512054443359, 0.03272985458374023, 0.03243008041381836, 0.03250719833374023, 0.03238451385498047, 0.03219494247436523, 0.03195916748046875, 0.031889408111572266, 0.03160211181640625, 0.03178553581237793, 0.03163113594055176, 0.0315446720123291, 0.0314069766998291, 0.03137740707397461, 0.03139174461364746, 0.03187628746032715, 0.031488096237182614, 0.03141647911071777, 0.03145376014709473, 0.03141340827941894, 0.03149519920349121, 0.031450944900512694, 0.03138710403442383, 0.03148633575439453, 0.031493759155273436, 0.031769119262695315, 0.03166543960571289, 0.03191500854492187, 0.031811296463012693, 0.031680511474609374, 0.03176563262939453, 0.03262758255004883, 0.03213721466064453, 0.03201811218261719, 0.03224947357177734, 0.03204166412353516, 0.032077823638916016, 0.03212595367431641, 0.03212102508544922, 0.03229087829589844, 0.03224038314819336, 0.03244972610473633, 0.032482112884521484, 0.032394496917724606, 0.03239926528930664, 0.03256816101074219, 0.032833534240722655, 0.03276556777954102, 0.03247123336791992, 0.03241388702392578, 0.032524383544921875, 0.03256924819946289, 0.03275689697265625, 0.0327784309387207, 0.03262736129760742, 0.03248332977294922, 0.03246284866333008, 0.03251814270019531, 0.03245167922973633, 0.03292422485351563, 0.03252054214477539, 0.03236380767822265, 0.032341854095458984, 0.032600128173828125, 0.03242886352539062, 0.032308639526367186, 0.0321495361328125, 0.03229542541503906, 0.032237632751464844, 0.03215359878540039, 0.03252838516235351, 0.03196454429626465, 0.03180790328979492, 0.0316942081451416, 0.03197014427185058, 0.03168870353698731, 0.03179427146911621, 0.03270134353637695, 0.031899648666381834, 0.03229695892333984, 0.032898399353027345, 0.032417888641357424, 0.03241548919677734, 0.03241247940063476, 0.0324771842956543, 0.03251814270019531, 0.032487422943115234, 0.03262054443359375, 0.03268403244018555, 0.032646560668945314, 0.032532096862792965, 0.03300067138671875, 0.03275526428222656, 0.03234016036987305, 0.0322902717590332, 0.032207393646240236, 0.03193446350097656, 0.03170528030395508, 0.031592256546020506, 0.03159859275817871, 0.03169190406799317, 0.031449024200439456, 0.031463359832763674, 0.03177948760986328, 0.03204726409912109, 0.03167865562438965, 0.03156991958618164, 0.03181363105773926, 0.0323210563659668, 0.03204963302612305, 0.03239113616943359, 0.031831424713134764, 0.03178681564331055, 0.03217903900146484, 0.03362201690673828, 0.03216707229614258, 0.03198796844482422, 0.031871519088745116, 0.03590300750732422, 0.036490814208984375, 0.03466099166870117, 0.03233388900756836, 0.032553249359130856, 0.032595966339111326, 0.03259392166137695, 0.03248931121826172, 0.032390430450439454, 0.03309571075439453, 0.03256550216674805, 0.03254716873168945, 0.03239161682128906, 0.032583393096923825, 0.033255424499511715, 0.03249596786499023, 0.03322608184814453, 0.03275225448608399, 0.03313663864135742, 0.033576480865478514, 0.03252828979492187, 0.032797248840332034, 0.03262828826904297, 0.032557502746582034, 0.03256729507446289, 0.03251814270019531, 0.03244646453857422, 0.03242803192138672, 0.032521888732910155, 0.03272848129272461, 0.03248812866210937, 0.032504062652587894, 0.03296246337890625, 0.0340805778503418, 0.03313897705078125, 0.032857662200927736, 0.03274924850463867, 0.03258771133422852, 0.03278726577758789, 0.032489471435546875, 0.03242598342895508, 0.03245043182373047, 0.03275584030151367, 0.03256630325317383, 0.03263382339477539, 0.032575233459472656, 0.03218198394775391, 0.03246688079833984, 0.03239731216430664, 0.032664161682128906, 0.0326814079284668, 0.03214393615722656, 0.0321943359375, 0.03207551956176758, 0.03205363082885742, 0.03191203117370606, 0.03184623908996582, 0.0318939208984375, 0.032148223876953125, 0.03230774307250977, 0.0323070068359375, 0.03241027069091797, 0.03229872131347656, 0.032198944091796876, 0.03225347137451172, 0.03225843048095703, 0.03237897491455078, 0.03229500961303711, 0.03203881454467773, 0.032094207763671875, 0.032298561096191405, 0.03232368087768555, 0.03223382568359375, 0.0320463981628418, 0.031979455947875976, 0.03290192031860351, 0.03212265777587891, 0.03229695892333984, 0.032343681335449216, 0.032906944274902344, 0.033233600616455077, 0.03262259292602539, 0.032350208282470705, 0.03301513671875, 0.032457374572753904, 0.03234406280517578, 0.032292736053466796, 0.03303641510009766, 0.03435852813720703, 0.032541343688964844, 0.03250124740600586, 0.032451488494873046, 0.0322413444519043, 0.03218841552734375, 0.032282752990722655, 0.03205043029785156, 0.03179583930969238, 0.03180083274841308, 0.03169100761413574, 0.03177497673034668, 0.03186073684692383, 0.031680511474609374, 0.03153267288208008, 0.03165632057189941, 0.03160678482055664, 0.03189760017395019, 0.031752191543579104, 0.03153510475158691, 0.03165184020996094, 0.0315567684173584, 0.031863199234008786, 0.03174608039855957, 0.0316208324432373, 0.03213996887207031, 0.03197686386108398, 0.031797599792480466, 0.03161500740051269, 0.031518943786621095, 0.03153097534179688, 0.031419456481933596, 0.03148080062866211, 0.03159568023681641, 0.03147878456115723, 0.03145907211303711, 0.03146761512756348, 0.031641855239868164, 0.03202624130249023, 0.031909887313842776, 0.032018207550048826, 0.03222768020629883, 0.0318764476776123, 0.031912607192993166, 0.031851871490478516, 0.03170707130432129, 0.03166896057128906, 0.031604736328125, 0.031643648147583005, 0.031528959274291994, 0.031535871505737306, 0.03139379119873047, 0.031494144439697266, 0.03159449577331543, 0.03158835220336914, 0.031692800521850584, 0.03161507225036621, 0.031493120193481446, 0.031649791717529296, 0.03149440002441406, 0.03162998390197754, 0.031389696121215824, 0.03179724884033203, 0.031680511474609374, 0.03157401657104492, 0.03162889671325684, 0.03187139129638672, 0.031892959594726565, 0.032037216186523436, 0.03231497573852539, 0.03191804885864258, 0.03211228942871094, 0.03190473556518555, 0.032481281280517575, 0.032086017608642575, 0.032233470916748046, 0.03218227386474609, 0.032476673126220705, 0.032495391845703124, 0.03254550552368164, 0.03249097442626953, 0.0324736328125, 0.032555007934570314, 0.03239052963256836, 0.03258841705322266, 0.0329727668762207, 0.033654014587402345, 0.03287871932983399, 0.0327616958618164, 0.032144191741943356, 0.032882625579833985, 0.032295230865478516, 0.03280051040649414, 0.03225600051879883, 0.03239868927001953, 0.032131614685058596, 0.03207158279418945, 0.03205961608886719, 0.03202006530761719, 0.03225027084350586, 0.03205961608886719, 0.031880319595336916, 0.032127201080322264, 0.03204463958740234, 0.0322476806640625, 0.03229372787475586, 0.03237030410766602, 0.032274944305419925, 0.032389118194580076, 0.032394496917724606, 0.03252096176147461, 0.0324156494140625, 0.03246697616577148, 0.03265475082397461, 0.03263343811035156, 0.032342239379882814, 0.03241097640991211, 0.032379329681396486, 0.03238931274414063, 0.03215955352783203, 0.03224371337890625, 0.032317440032958986, 0.032343647003173825, 0.03244809722900391, 0.03241862487792969, 0.032589534759521484, 0.032493854522705076, 0.03245817565917969, 0.032520225524902344, 0.033309215545654296, 0.03259945678710938, 0.032514656066894534, 0.032686080932617184, 0.03273523330688476, 0.03346953582763672, 0.03259033584594727, 0.03243459320068359, 0.03248028945922852, 0.03245935821533203, 0.032159648895263675, 0.032149982452392575, 0.031936511993408204, 0.031899648666381834, 0.031643648147583005, 0.03198921585083008, 0.03212460708618164, 0.03164828872680664, 0.03155590438842774, 0.031959039688110355, 0.03186278343200684, 0.03192563247680664, 0.031879583358764646, 0.03197929573059082, 0.031995647430419924, 0.032065216064453124, 0.03216486358642578, 0.03225356674194336, 0.03229734420776367, 0.03218399810791016, 0.03221737670898438, 0.032349857330322265, 0.03230992126464844, 0.03634966278076172, 0.03226419067382812, 0.032246784210205076, 0.03287068939208984, 0.03252883148193359, 0.0324150390625, 0.03212796783447266, 0.03210444641113281, 0.03215510559082031, 0.03226883316040039, 0.03244815826416016, 0.03250435256958008, 0.032324542999267576, 0.0324389762878418, 0.0323768310546875, 0.03217027282714844, 0.032148670196533204, 0.03224355316162109, 0.03225084686279297, 0.03249935913085938, 0.032228736877441405, 0.032091102600097655, 0.03196025657653809, 0.0320827522277832, 0.032307201385498044, 0.032020481109619144, 0.031942655563354495, 0.03176652717590332, 0.033576961517333984, 0.031748096466064454, 0.03194655990600586, 0.0317926082611084, 0.031750879287719726, 0.0318156795501709, 0.031704095840454104, 0.03182691192626953, 0.031768224716186524, 0.0318057918548584, 0.031983200073242186, 0.03218207931518555, 0.03248393630981445, 0.03242758560180664, 0.03232403182983398, 0.032468097686767575, 0.0332276496887207, 0.03243318557739258, 0.03259695816040039, 0.03246284866333008, 0.03241164779663086, 0.03244169616699219, 0.034328895568847655, 0.03273353576660156, 0.032640640258789065, 0.03241516876220703, 0.03224252700805664, 0.03545651245117187, 0.033497150421142576, 0.03217798233032226, 0.03228745651245117, 0.03224915313720703, 0.03237343978881836, 0.032152641296386716, 0.0323546257019043, 0.03243276977539063, 0.03243174362182617, 0.032373119354248046, 0.03234201431274414, 0.03243027114868164, 0.03244182586669922, 0.032516448974609376, 0.03221209716796875, 0.032186592102050784, 0.0320211181640625, 0.03200601577758789, 0.03206307220458984, 0.03222742462158203, 0.03158883285522461, 0.03158527946472168, 0.03160223960876465, 0.031360416412353515, 0.031498720169067386, 0.03150691223144531, 0.03156732749938965, 0.031504255294799804, 0.03152070426940918, 0.0317541446685791, 0.03200211334228516, 0.032064254760742185, 0.03314201736450195, 0.03246889495849609, 0.03203363037109375, 0.03201638412475586, 0.032042816162109376, 0.03200019073486328, 0.031883264541625975, 0.03204857635498047, 0.03228236770629883, 0.03222345733642578, 0.03223612976074219, 0.03202867126464844, 0.0322696647644043, 0.032234142303466794, 0.032036865234375, 0.03195084762573242, 0.032062591552734374, 0.032089054107666014, 0.032126399993896486, 0.032255455017089846, 0.03221820831298828, 0.03207363128662109, 0.03226121520996094, 0.03217011260986328, 0.03235055923461914, 0.0322432975769043, 0.03235929489135742, 0.03235177612304688, 0.03220912170410156, 0.0322644157409668, 0.03208806228637695, 0.03257753753662109, 0.03224700927734375, 0.03215030288696289, 0.03236211013793945, 0.0322861442565918, 0.03226310348510742, 0.03237587356567383, 0.03231817626953125, 0.032333824157714845, 0.032639198303222657, 0.03237478256225586, 0.0335596809387207, 0.033777599334716794, 0.03280160140991211, 0.032552734375, 0.032544097900390624, 0.03248844909667969, 0.03287449645996094, 0.03255091094970703, 0.03236249542236328, 0.03243737411499024, 0.032521759033203125, 0.0322564811706543, 0.03340854263305664, 0.03487136077880859, 0.03273104095458984, 0.03243673706054687, 0.0324956169128418, 0.0324587516784668, 0.03254671859741211, 0.03305196762084961, 0.032540767669677735, 0.03225465774536133, 0.03235747146606445, 0.0323919677734375, 0.032274559020996095, 0.032396736145019533, 0.03245449447631836, 0.03248406219482422, 0.03240755081176758, 0.03242393493652344, 0.032767681121826174, 0.03246931076049805, 0.03259555053710937, 0.032725406646728517, 0.032727039337158204, 0.03271212768554688, 0.03307907104492187, 0.0339463996887207, 0.03253855895996094, 0.03263641738891602, 0.032718753814697264, 0.03233039855957031, 0.03260006332397461, 0.03265651321411133, 0.03257433700561523, 0.03258367919921875, 0.03247462463378906, 0.032346622467041015, 0.032710655212402344, 0.03256320190429687, 0.03259801483154297, 0.03244646453857422, 0.03248134231567383, 0.03237267303466797, 0.03259392166137695, 0.03235414505004883, 0.03260636901855469, 0.03252579116821289, 0.03252278518676758, 0.032486785888671876, 0.03240153503417969, 0.032661151885986325, 0.03235504150390625, 0.03231167984008789, 0.03244831848144531, 0.032638111114501954, 0.032514400482177734, 0.032549312591552734, 0.032353313446044925, 0.03254719924926758, 0.032473697662353515, 0.03271440124511719, 0.03239475250244141]",tokens/s,30.89508152474481,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3812.012032,4371.447808,0.0,3992.977408,3875.045888,s,1,9.7830205078125,9.7830205078125,0.0,9.7830205078125,9.7830205078125,9.7830205078125,9.7830205078125,[9.7830205078125],,kWh,7.739799573333338e-05,8.521289569338107e-06,2.5187797927997413e-05,0.0001111070832306689,,MB,2050.576384,4595.843072,0.0,4188.012544,4099.587072,s,10,3.3127397155761718,0.3312739715576172,0.0017208316625747086,0.3311283721923828,0.3333113311767578,0.3334079696655273,0.33348528045654297,"[0.3335046081542969, 0.3300729675292969, 0.329398193359375, 0.3296027526855469, 0.32876229858398437, 0.3304183654785156, 0.33183837890625, 0.3329078369140625, 0.3329444580078125, 0.33328985595703126]",tokens/s,772.7742653499565,kWh,1.0025084371944438e-05,1.1055919539487317e-06,6.651486802666575e-06,1.7782163128559746e-05,tokens/kWh,14396448.741876688,MB,2059.534336,4700.700672,0.0,4292.870144,4197.814272,s,10,27.489537109375,2.7489537109375,0.017626090166925304,2.743663330078125,2.7678792480468752,2.78010087890625,2.78987818359375,"[2.755270263671875, 2.74456005859375, 2.739586669921875, 2.731625244140625, 2.7427666015625, 2.7393564453125, 2.765163330078125, 2.74955908203125, 2.729326904296875, 2.792322509765625]",tokens/s,22.917810419774057,kWh,8.129318500138914e-05,8.966749398900194e-06,5.0396086613134033e-05,0.00014065602101342336,tokens/kWh,447901.19574040605,,s,630,27.48669079208376,0.04362966792394245,0.0006390068181520789,0.04353998374938965,0.04422013168334961,0.04451613445281982,0.04659644077301026,"[0.043960319519042966, 0.04343603134155274, 0.04382009506225586, 0.04642649459838867, 0.043063072204589846, 0.04311721420288086, 0.04314739227294922, 0.04322937774658203, 0.042990753173828125, 0.04321062469482422, 0.04330310440063476, 0.04361801528930664, 0.043721790313720706, 0.04364582443237305, 0.04357830429077148, 0.043684993743896484, 0.043581920623779295, 0.04368172836303711, 0.04380720138549805, 0.043620128631591794, 0.043534561157226564, 0.04345158386230469, 0.04359609603881836, 0.04347055816650391, 0.043428638458251956, 0.043756576538085935, 0.04377289581298828, 0.04364191818237305, 0.04446201705932617, 0.04382646560668945, 0.043779041290283205, 0.04377657699584961, 0.04396428680419922, 0.04352851104736328, 0.044257278442382815, 0.04351619338989258, 0.043714176177978514, 0.0436014404296875, 0.04376633453369141, 0.043802623748779294, 0.043665374755859375, 0.043687488555908205, 0.0439136962890625, 0.04392460632324219, 0.04402995300292969, 0.04397116851806641, 0.04393308639526367, 0.04408204650878906, 0.044014816284179685, 0.04406480026245117, 0.04415155029296875, 0.04397830581665039, 0.04410806274414063, 0.04427132797241211, 0.04377347183227539, 0.043616798400878905, 0.04358614349365234, 0.04340671920776367, 0.043520416259765625, 0.043747329711914064, 0.04352614212036133, 0.043345569610595706, 0.04338723373413086, 0.043905025482177736, 0.04396569442749024, 0.04389555358886719, 0.044041664123535156, 0.043870174407958984, 0.04398271942138672, 0.043746017456054685, 0.043796478271484376, 0.04339712142944336, 0.0433963851928711, 0.043450942993164064, 0.04341708755493164, 0.04315203094482422, 0.04307558441162109, 0.04315135955810547, 0.043278335571289066, 0.043153408050537106, 0.04376166534423828, 0.04363468933105469, 0.04377190399169922, 0.04357324981689453, 0.04360988616943359, 0.04352550506591797, 0.04377401733398437, 0.044357856750488284, 0.04352057647705078, 0.04375961685180664, 0.04628889465332031, 0.043824352264404294, 0.04346345520019531, 0.0431383056640625, 0.04344294357299805, 0.04330214309692383, 0.04356739044189453, 0.04343036651611328, 0.043786239624023435, 0.04363600158691406, 0.04362745666503906, 0.043757183074951175, 0.04372662353515625, 0.04365350341796875, 0.04384902572631836, 0.043383487701416014, 0.04331257629394531, 0.043493087768554685, 0.04338979339599609, 0.04314931106567383, 0.043617408752441404, 0.04298640060424805, 0.04428086471557617, 0.043129791259765626, 0.04372889709472656, 0.043524097442626954, 0.04437926483154297, 0.042997760772705076, 0.04302499389648438, 0.043039169311523434, 0.04321644973754883, 0.043103809356689456, 0.043356094360351566, 0.04269750213623047, 0.04318371200561524, 0.04286233520507812, 0.04375551986694336, 0.04331315231323242, 0.04337670516967773, 0.04347385787963867, 0.04353126525878906, 0.04367055892944336, 0.043581439971923826, 0.043592575073242185, 0.04350268936157226, 0.043463680267333986, 0.043093791961669924, 0.04362982559204102, 0.04345289611816406, 0.04329318237304688, 0.043044193267822266, 0.042869407653808596, 0.04293840026855469, 0.042933280944824216, 0.04319942474365234, 0.043730945587158204, 0.043905120849609375, 0.04302998352050781, 0.043382911682128905, 0.04292256164550781, 0.042897151947021483, 0.04376166534423828, 0.04305500793457031, 0.043634529113769534, 0.04368368148803711, 0.04383580780029297, 0.04393292617797852, 0.04371513748168945, 0.04372636795043945, 0.04355718231201172, 0.043649375915527346, 0.043676895141601564, 0.04375632095336914, 0.043649024963378906, 0.04383119964599609, 0.04352355194091797, 0.04321958541870117, 0.04323328018188476, 0.04321279907226563, 0.04320569610595703, 0.043576416015625, 0.04412188720703125, 0.04381827163696289, 0.04354300689697266, 0.04367366409301758, 0.04356451034545898, 0.04385871887207031, 0.04357120132446289, 0.04337417602539063, 0.0432193603515625, 0.04349737548828125, 0.04418569564819336, 0.04431824111938477, 0.0433732795715332, 0.04309695816040039, 0.04302115249633789, 0.04303292846679688, 0.043294654846191404, 0.04369382476806641, 0.04511510467529297, 0.04321548843383789, 0.04317219161987305, 0.042929824829101564, 0.042950721740722654, 0.042983360290527343, 0.043012096405029294, 0.04297552108764648, 0.04274288177490235, 0.042770496368408205, 0.042743423461914065, 0.042756927490234374, 0.04298144149780273, 0.04356313705444336, 0.04340694427490235, 0.04312303924560547, 0.042999519348144534, 0.0429071044921875, 0.04290233612060547, 0.04293632125854492, 0.043084030151367185, 0.04690713500976563, 0.04466281509399414, 0.04363827133178711, 0.04368841552734375, 0.043558399200439454, 0.04380108642578125, 0.0433328971862793, 0.0432663688659668, 0.04350611114501953, 0.04318819046020508, 0.043065055847167966, 0.04294655990600586, 0.04293247985839844, 0.042799457550048825, 0.04296044921875, 0.04326822280883789, 0.043558048248291015, 0.043514976501464846, 0.04353200149536133, 0.043369632720947265, 0.04335036849975586, 0.043393566131591794, 0.04361830520629883, 0.0435847053527832, 0.04315404891967774, 0.043794559478759765, 0.04336569595336914, 0.0431778564453125, 0.04320105743408203, 0.04328278350830078, 0.04301824188232422, 0.04297318267822266, 0.042962944030761716, 0.04329644775390625, 0.043374752044677736, 0.043964160919189456, 0.04342211151123047, 0.043741024017333985, 0.04365558242797852, 0.043361183166503905, 0.04355158233642578, 0.04331836700439453, 0.04393574523925781, 0.04398262405395508, 0.04390729522705078, 0.04304883193969727, 0.04317971038818359, 0.04315180969238281, 0.04320870590209961, 0.04324761581420898, 0.04292550277709961, 0.04287577438354492, 0.04276192092895508, 0.042649856567382814, 0.042925697326660156, 0.04310822296142578, 0.0433175048828125, 0.043524097442626954, 0.04353567886352539, 0.043456897735595704, 0.044017982482910153, 0.04381647872924805, 0.04358396911621094, 0.04343132781982422, 0.04456745529174805, 0.043455360412597656, 0.04327302551269531, 0.043522209167480466, 0.042942527770996095, 0.043269920349121097, 0.04310198211669922, 0.043176158905029294, 0.04347488021850586, 0.04341542434692383, 0.04363455963134766, 0.043619998931884764, 0.043597694396972655, 0.043682590484619144, 0.0436096305847168, 0.04352774429321289, 0.04358224105834961, 0.043409534454345707, 0.043490943908691404, 0.04353267288208008, 0.04335184097290039, 0.04352854537963867, 0.043449760437011715, 0.04330169677734375, 0.04328569412231445, 0.04341715240478516, 0.04401379013061523, 0.043880382537841794, 0.043980735778808594, 0.04396966552734375, 0.04452729415893555, 0.043911167144775394, 0.04455014419555664, 0.044025856018066405, 0.04396156692504883, 0.04404643249511719, 0.043767711639404294, 0.04385385513305664, 0.04374556732177735, 0.043420063018798825, 0.043039905548095704, 0.0436940803527832, 0.04530172729492187, 0.04351513671875, 0.043313953399658205, 0.043218944549560545, 0.04302643203735351, 0.04287078475952148, 0.04318627166748047, 0.04309187316894531, 0.043253761291503906, 0.043286529541015625, 0.0430398063659668, 0.044209087371826175, 0.04473251342773438, 0.04459881591796875, 0.04361648178100586, 0.043819393157958984, 0.04365599822998047, 0.04372579193115234, 0.04353868865966797, 0.043481952667236326, 0.04335094451904297, 0.04340038299560547, 0.04332204818725586, 0.04338035202026367, 0.04348396682739258, 0.043636417388916014, 0.04310835266113281, 0.042971134185791016, 0.04331135940551758, 0.043181217193603516, 0.043151905059814456, 0.04345657730102539, 0.043466751098632815, 0.043046432495117186, 0.043032638549804686, 0.04299817657470703, 0.043194366455078126, 0.043259902954101564, 0.04289292907714844, 0.04280105590820313, 0.04287126541137695, 0.042831871032714845, 0.043442176818847655, 0.0437022705078125, 0.0429936637878418, 0.04322099304199219, 0.04436195373535156, 0.043818782806396485, 0.04395212936401367, 0.044183551788330076, 0.04368105697631836, 0.04350844955444336, 0.043574432373046874, 0.043661376953125, 0.04366175842285156, 0.04393340682983399, 0.04352473449707031, 0.044678302764892576, 0.04341609573364258, 0.04316543960571289, 0.04325555038452149, 0.04302102279663086, 0.04402252960205078, 0.0438455696105957, 0.043910335540771485, 0.04900908660888672, 0.04666585540771485, 0.04391142272949219, 0.04383327865600586, 0.04349753570556641, 0.04351523208618164, 0.04366118240356445, 0.04331372833251953, 0.043122913360595705, 0.043142848968505856, 0.0459062385559082, 0.04357551956176758, 0.043589408874511716, 0.04370822525024414, 0.04341350555419922, 0.04353414535522461, 0.04328691101074219, 0.04753952026367188, 0.04396819305419922, 0.04406310272216797, 0.04390531158447265, 0.04386608123779297, 0.043792766571044923, 0.04346879959106445, 0.043380897521972654, 0.04351270294189453, 0.043888641357421876, 0.04342473602294922, 0.043370113372802735, 0.04319798278808594, 0.04325462341308594, 0.04317948913574219, 0.04323916625976563, 0.043934497833251956, 0.04370841598510742, 0.0436861457824707, 0.04391692733764648, 0.046976318359375, 0.04406175994873047, 0.04359382247924805, 0.04351264190673828, 0.04347574234008789, 0.04363270568847656, 0.04364006423950195, 0.04353305435180664, 0.04365046310424805, 0.04334441757202148, 0.043280448913574215, 0.04341356658935547, 0.04359980773925781, 0.04380387115478516, 0.043813663482666014, 0.04370431900024414, 0.04356300735473633, 0.04375551986694336, 0.04361625671386719, 0.043687934875488284, 0.04353843307495117, 0.04331257629394531, 0.04359142303466797, 0.04429865646362305, 0.043526336669921874, 0.04316156768798828, 0.043228191375732423, 0.042988704681396483, 0.043490463256835935, 0.042997856140136716, 0.04326031875610352, 0.04296518325805664, 0.04317177581787109, 0.04294278335571289, 0.043635902404785154, 0.043432449340820314, 0.04309811019897461, 0.043116127014160156, 0.04338294219970703, 0.0434158706665039, 0.04400300979614258, 0.043888702392578124, 0.04373302459716797, 0.04367177581787109, 0.04426777648925781, 0.04381875228881836, 0.044056159973144535, 0.04364944076538086, 0.043463775634765625, 0.043404193878173826, 0.04366854476928711, 0.043533470153808596, 0.04350278472900391, 0.04341411209106445, 0.04351100921630859, 0.04359862518310547, 0.043826366424560545, 0.044165950775146484, 0.044154815673828125, 0.04395219039916992, 0.04435385513305664, 0.044220096588134764, 0.04474998474121094, 0.044190559387207035, 0.04446966552734375, 0.04418518447875976, 0.04420083236694336, 0.044324993133544925, 0.04388454437255859, 0.04349542236328125, 0.04314726257324219, 0.04336975860595703, 0.04337276840209961, 0.04329318237304688, 0.04315264129638672, 0.04337740707397461, 0.043663360595703124, 0.04383667373657227, 0.04388227081298828, 0.043768798828125, 0.04361593627929688, 0.04352143859863281, 0.043458782196044925, 0.04354323196411133, 0.04348748779296875, 0.04333132934570313, 0.04378617477416992, 0.043361087799072266, 0.043046913146972655, 0.04306534576416016, 0.04280115127563477, 0.04283391952514649, 0.042880382537841794, 0.04301427078247071, 0.042969470977783204, 0.04307366561889649, 0.04291491317749024, 0.04291990280151367, 0.042828510284423825, 0.04400559997558594, 0.04293427276611328, 0.04327529525756836, 0.04288972854614258, 0.04291433715820313, 0.043794368743896486, 0.04292812728881836, 0.04310966491699219, 0.0433076171875, 0.04283772659301758, 0.043063358306884764, 0.04307388687133789, 0.04302643203735351, 0.04285590362548828, 0.04293891143798828, 0.042848255157470705, 0.04289251327514648, 0.04315177536010742, 0.04710435104370117, 0.04351798248291015, 0.04332038497924805, 0.04309907150268555, 0.04323123168945313, 0.04350566482543945, 0.04315071868896484, 0.04340364837646484, 0.04325174331665039, 0.04298921585083008, 0.04298553466796875, 0.04289353561401367, 0.04303862380981445, 0.04321260833740234, 0.04417593765258789, 0.04333100891113281, 0.043507457733154294, 0.04323001480102539, 0.043665409088134766, 0.043655166625976564, 0.04370140838623047, 0.04367257690429688, 0.04384553527832031, 0.04385580825805664, 0.04365107345581055, 0.043319297790527345, 0.0433704948425293, 0.04346579360961914, 0.043541278839111325, 0.04361846542358398, 0.043883838653564454, 0.04350588989257813, 0.043829280853271486, 0.04364582443237305, 0.043601024627685545, 0.04357004928588867, 0.04378534317016602, 0.04382400131225586, 0.04366070556640625, 0.04395257568359375, 0.04406697463989258, 0.04422969436645508, 0.044157920837402345, 0.04424448013305664, 0.04415686416625977, 0.044184288024902346, 0.04402473449707031, 0.044219295501708986, 0.04415081787109375, 0.04445532989501953, 0.043890975952148435, 0.043912513732910156, 0.044552352905273436, 0.044813182830810545, 0.04421638488769531, 0.044211105346679686, 0.046255264282226566, 0.04423011016845703, 0.04457302474975586, 0.04452761459350586, 0.044406398773193356, 0.044405120849609375, 0.045647422790527345, 0.04502732849121094, 0.04425881576538086, 0.04445087814331055, 0.04473990249633789, 0.045109825134277345, 0.04464652633666992, 0.04450249481201172, 0.044529182434082035, 0.04434009552001953, 0.04438355255126953, 0.0442204475402832, 0.04476176071166992, 0.044480159759521486, 0.044376033782958985, 0.04449955368041992, 0.0443021125793457, 0.044050079345703125, 0.044138622283935544, 0.044226783752441406, 0.04387145614624023, 0.0440239372253418, 0.0438895378112793, 0.044262176513671876, 0.043864959716796874, 0.043935073852539065, 0.04415180969238281, 0.0473210563659668, 0.044031806945800785, 0.044066814422607424, 0.043998336791992186, 0.04411481475830078, 0.043991073608398434]",tokens/s,22.920183617790837,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3798, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ self.model = DeciCoderModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ self.self_attn = DeciCoderAttention(config=config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ self._init_rope() File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1709, in __getattr__ raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2193.03936,2479.816704,0.0,2101.346304,1978.345472,s,1,9.1729736328125,9.1729736328125,0.0,9.1729736328125,9.1729736328125,9.1729736328125,9.1729736328125,[9.1729736328125],,kWh,5.379501379585842e-05,5.926686920874674e-06,1.6791680099992323e-05,7.651338081672542e-05,,MB,2226.229248,2817.458176,0.0,2409.627648,2250.334208,s,10,1.6971580505371096,0.16971580505371092,0.0006403378374378439,0.16959552001953127,0.17055413208007814,0.1706893051147461,0.17079744354248047,"[0.16978909301757814, 0.16916297912597655, 0.1688795166015625, 0.169161376953125, 0.16909584045410156, 0.17082447814941407, 0.16940194702148437, 0.17005043029785155, 0.1705240936279297, 0.17026829528808593]",tokens/s,1508.4040046770083,kWh,5.167874112865783e-06,5.696213223745739e-07,3.4221128741403476e-06,9.159608309380703e-06,tokens/kWh,27948793.371199146,MB,2240.802816,2859.401216,0.0,2451.570688,2333.38624,s,10,23.642529052734375,2.3642529052734376,0.01051823980940855,2.363437744140625,2.3755578857421873,2.379169201660156,2.3820582543945314,"[2.36325146484375, 2.34633642578125, 2.373258544921875, 2.354318115234375, 2.357580078125, 2.370532470703125, 2.37475537109375, 2.356092041015625, 2.3636240234375, 2.382780517578125]",tokens/s,26.64689545669131,kWh,6.805154646213757e-05,7.506256485265811e-06,3.445113964666052e-05,0.00011000894259406391,tokens/kWh,572680.7158984499,,s,630,23.63888668060304,0.037522042350163536,0.0005600282889955902,0.03745980834960938,0.037916993331909184,0.038079124450683594,0.03949451812744141,"[0.03775897598266602, 0.03779379272460937, 0.038078369140625, 0.03741500854492188, 0.03747840118408203, 0.037695297241210936, 0.03734102249145508, 0.037335391998291015, 0.037410175323486325, 0.037184127807617186, 0.03738447952270508, 0.037238529205322266, 0.03717724609375, 0.037398208618164064, 0.03736409759521484, 0.03734912109375, 0.03719193649291992, 0.03741247940063477, 0.03741491317749023, 0.03741734313964844, 0.041785343170166016, 0.03758694458007812, 0.037553184509277346, 0.03763644790649414, 0.03762649536132812, 0.037659774780273436, 0.03764108657836914, 0.037555648803710935, 0.03748921585083008, 0.038176769256591796, 0.03778556823730469, 0.037515296936035156, 0.0376545295715332, 0.03802316665649414, 0.03841228866577148, 0.03777536010742188, 0.0376360969543457, 0.03759513473510742, 0.03750707244873047, 0.03772415924072266, 0.03765043258666992, 0.03746406555175781, 0.03733708953857422, 0.03752755355834961, 0.037136192321777346, 0.03715091323852539, 0.03733913421630859, 0.03724662399291992, 0.037095775604248045, 0.037123233795166015, 0.03704713439941406, 0.03699043273925781, 0.03705904006958008, 0.03711801528930664, 0.03701760101318359, 0.03708108901977539, 0.0371671028137207, 0.03705855941772461, 0.03710976028442383, 0.0370683822631836, 0.03711427307128906, 0.03750092697143555, 0.037311904907226565, 0.037195297241210935, 0.03727420806884765, 0.03734732818603516, 0.036923393249511716, 0.037070846557617186, 0.036806560516357424, 0.03680207824707031, 0.0368741455078125, 0.03698041534423828, 0.037053409576416015, 0.03721011352539062, 0.037269504547119144, 0.03717468643188476, 0.03718000030517578, 0.03706675338745117, 0.03712128067016601, 0.037206783294677734, 0.037592960357666017, 0.03731212615966797, 0.037515777587890625, 0.037564414978027344, 0.03721033477783203, 0.03724620819091797, 0.03702790451049805, 0.03704636764526367, 0.03699955368041992, 0.037070846557617186, 0.03708671951293945, 0.037118465423583984, 0.03706265640258789, 0.03720393753051758, 0.03718143844604492, 0.037170497894287106, 0.037053150177001955, 0.03717900848388672, 0.03728627014160156, 0.0372674560546875, 0.03721555328369141, 0.03769955062866211, 0.03855571365356445, 0.03753209686279297, 0.03760547256469727, 0.03785059356689453, 0.03735004806518555, 0.037504638671875, 0.03748281478881836, 0.037455936431884766, 0.037420639038085936, 0.03731036758422852, 0.0371932144165039, 0.03717552185058594, 0.03716790390014649, 0.036980735778808595, 0.036988929748535154, 0.03703603363037109, 0.03684281539916992, 0.036843936920166014, 0.03693612670898438, 0.03801891326904297, 0.03712979125976563, 0.03765497589111328, 0.03709247970581055, 0.037181983947753905, 0.03742697525024414, 0.03733071899414062, 0.03742969512939453, 0.03702783966064453, 0.03722854232788086, 0.03709302520751953, 0.037953056335449216, 0.03773049545288086, 0.037472862243652344, 0.037324127197265626, 0.038072608947753904, 0.03735798263549805, 0.03742924880981445, 0.03725628662109375, 0.03719465637207031, 0.037351425170898435, 0.03766678237915039, 0.03764543914794922, 0.037643169403076174, 0.0380673599243164, 0.03806636810302735, 0.03745846557617188, 0.03755952072143555, 0.03758172988891602, 0.037617374420166015, 0.03760924911499024, 0.03788851165771484, 0.037740287780761717, 0.03776691055297852, 0.03771814346313476, 0.03770518493652344, 0.037353439331054686, 0.037491649627685544, 0.037711742401123044, 0.037533824920654296, 0.037289535522460934, 0.03724435043334961, 0.03721887969970703, 0.037378494262695315, 0.03758489608764649, 0.038430721282958984, 0.039335006713867186, 0.03879209518432617, 0.03773440170288086, 0.03777283096313477, 0.03781475067138672, 0.03753574371337891, 0.03771731185913086, 0.03781644821166992, 0.03751129531860352, 0.03741881561279297, 0.03736025619506836, 0.03739033508300781, 0.038929920196533206, 0.03746441650390625, 0.03734339141845703, 0.0374574089050293, 0.03740016174316406, 0.037380031585693356, 0.03789718246459961, 0.03990486526489258, 0.037756767272949215, 0.03751785659790039, 0.037240768432617186, 0.037275390625, 0.037042495727539065, 0.03672265625, 0.037036064147949216, 0.03704627227783203, 0.03698483276367188, 0.03707040023803711, 0.036991199493408206, 0.036569313049316404, 0.03697868728637695, 0.036875328063964846, 0.03729094314575195, 0.037318656921386716, 0.03746406555175781, 0.037324798583984374, 0.03741408157348633, 0.03748332977294922, 0.03749478530883789, 0.03752278518676758, 0.03727951812744141, 0.03738508987426758, 0.03735491180419922, 0.03725075149536133, 0.03741788864135742, 0.03729612731933594, 0.03734649658203125, 0.03723468780517578, 0.037352256774902344, 0.037532958984375, 0.037474174499511716, 0.037690208435058596, 0.03752671813964844, 0.03752979278564453, 0.03753228759765625, 0.03762956619262695, 0.03766425704956055, 0.03747964859008789, 0.03762041473388672, 0.03766748809814453, 0.03750534439086914, 0.03714361572265625, 0.037216991424560544, 0.03716041564941406, 0.03721907043457031, 0.037736446380615234, 0.037705726623535156, 0.03735321426391602, 0.03747865676879883, 0.03781167984008789, 0.03741750335693359, 0.03735251235961914, 0.03740563201904297, 0.03752675247192383, 0.03755254364013672, 0.03767539215087891, 0.03807027053833008, 0.037240543365478516, 0.037500320434570314, 0.037144447326660154, 0.03721932983398438, 0.037039424896240236, 0.038120128631591796, 0.03791011047363281, 0.03751567840576172, 0.03754703903198242, 0.037262271881103516, 0.037161121368408205, 0.0374015998840332, 0.037282688140869144, 0.03699647903442383, 0.03713702392578125, 0.03702947235107422, 0.03731292724609375, 0.037162784576416016, 0.03714275360107422, 0.036896766662597655, 0.03702067184448242, 0.03687936019897461, 0.037077056884765626, 0.037232574462890626, 0.04012441635131836, 0.03716505432128906, 0.03718143844604492, 0.037105056762695314, 0.036894527435302735, 0.03709545516967774, 0.03761356735229492, 0.03732345581054688, 0.03726454544067383, 0.03724176025390625, 0.03714665603637695, 0.03705033493041992, 0.03695135879516601, 0.0372558708190918, 0.03705241775512695, 0.03708265686035156, 0.03716553497314453, 0.0373430404663086, 0.037947582244873046, 0.03759308624267578, 0.037477535247802736, 0.0374444465637207, 0.03737571334838867, 0.03731689453125, 0.03743923187255859, 0.03772848129272461, 0.037654560089111326, 0.037563392639160156, 0.03746223831176758, 0.03731740951538086, 0.037319934844970704, 0.037235073089599606, 0.03752793502807617, 0.037438816070556644, 0.037488670349121095, 0.03751356887817383, 0.03760976028442383, 0.037582847595214845, 0.03764223861694336, 0.038079742431640626, 0.03800051116943359, 0.037965888977050784, 0.037705982208251956, 0.03771244812011719, 0.03801702499389648, 0.03767468643188476, 0.0375951042175293, 0.03792521667480469, 0.03787545776367188, 0.037763198852539065, 0.037460063934326174, 0.037695518493652345, 0.037615230560302734, 0.037541889190673826, 0.0372259521484375, 0.03718656158447266, 0.03723811340332031, 0.037369953155517575, 0.03760121536254883, 0.03746870422363281, 0.037375904083251955, 0.03855708694458008, 0.03821020889282226, 0.037978145599365236, 0.03766067123413086, 0.03824435043334961, 0.037620864868164065, 0.03738438415527344, 0.037278400421142575, 0.03757056045532227, 0.0373125114440918, 0.03728815841674805, 0.03745769500732422, 0.037607295989990235, 0.03721993637084961, 0.037542015075683596, 0.03731907272338867, 0.03750707244873047, 0.037353473663330077, 0.03755382537841797, 0.03911033630371094, 0.03843052673339844, 0.037326976776123046, 0.03725795364379883, 0.037160961151123044, 0.03701964950561523, 0.03705759811401367, 0.0370656623840332, 0.037236736297607424, 0.037572608947753904, 0.03762790298461914, 0.03747401428222656, 0.03750358581542969, 0.03751494216918945, 0.037760704040527344, 0.03767062377929688, 0.03762361526489258, 0.038617889404296876, 0.037879104614257815, 0.03803596878051758, 0.03900998306274414, 0.03786953735351563, 0.037626399993896484, 0.03748863983154297, 0.03746524810791016, 0.03762876892089844, 0.03745177459716797, 0.037408767700195314, 0.03765248107910156, 0.03742095947265625, 0.03797615814208984, 0.03743743896484375, 0.03765584182739258, 0.037318431854248046, 0.03743593597412109, 0.03728425598144531, 0.03759132766723633, 0.03755548858642578, 0.03751776123046875, 0.03746332931518555, 0.037704414367675784, 0.037557662963867186, 0.037573215484619144, 0.03774399948120117, 0.03785587310791016, 0.037659774780273436, 0.03779593658447265, 0.03774294281005859, 0.03753209686279297, 0.037453792572021485, 0.037604927062988285, 0.03750931167602539, 0.037289886474609374, 0.037080703735351564, 0.037157630920410155, 0.037101600646972654, 0.03736716842651367, 0.03711036682128906, 0.03741686248779297, 0.037201217651367184, 0.037167743682861326, 0.03711616134643555, 0.03733084869384766, 0.03727974319458008, 0.03737190246582031, 0.037294078826904296, 0.04583161544799805, 0.03788857650756836, 0.037367809295654295, 0.03730022430419922, 0.039833377838134766, 0.03738236618041992, 0.03743097686767578, 0.03802758407592773, 0.03748230361938477, 0.03732265472412109, 0.037488319396972655, 0.037963871002197266, 0.03834726333618164, 0.0378185920715332, 0.03787139129638672, 0.037580608367919925, 0.03734956741333008, 0.038158336639404294, 0.03750841522216797, 0.03760726547241211, 0.037380958557128904, 0.03736166381835938, 0.037461505889892575, 0.03765209579467774, 0.03763494491577148, 0.037566463470458986, 0.037722110748291016, 0.03759718322753906, 0.037533695220947266, 0.037502944946289064, 0.03746614456176758, 0.037920768737792966, 0.037765121459960936, 0.03767091369628906, 0.03767232131958008, 0.03799302291870117, 0.037889408111572265, 0.03748115158081055, 0.037483551025390624, 0.037497825622558593, 0.03741667175292969, 0.03767715072631836, 0.03725331115722656, 0.037253120422363284, 0.0372305908203125, 0.03743552017211914, 0.037316478729248044, 0.03728998565673828, 0.03703564834594727, 0.037017982482910154, 0.036939777374267575, 0.03721420669555664, 0.03711385726928711, 0.03721011352539062, 0.037253120422363284, 0.03718454360961914, 0.037026782989501954, 0.03698271942138672, 0.03712416076660156, 0.03710566329956055, 0.037144577026367184, 0.03727155303955078, 0.03725641632080078, 0.03713449478149414, 0.03720153427124023, 0.03718896102905273, 0.037589950561523436, 0.03752387237548828, 0.037484416961669924, 0.03752975845336914, 0.03790195083618164, 0.03746486282348633, 0.03738201522827148, 0.03745955276489258, 0.03717161560058594, 0.03737104034423828, 0.037283870697021486, 0.03721433639526367, 0.037493217468261716, 0.0374950065612793, 0.03723263931274414, 0.038856639862060546, 0.037209888458251954, 0.037208351135253906, 0.03705241775512695, 0.03745177459716797, 0.03710079956054688, 0.037190399169921874, 0.037434688568115236, 0.03727171325683594, 0.037304862976074216, 0.03733299255371094, 0.03742281723022461, 0.03701148986816406, 0.03746022415161133, 0.03747635269165039, 0.03747779083251953, 0.037478111267089845, 0.03740147018432617, 0.03736166381835938, 0.037449726104736326, 0.037519073486328124, 0.037480735778808595, 0.03953664016723633, 0.037482143402099606, 0.03733334350585937, 0.03733078384399414, 0.03736940765380859, 0.03719782257080078, 0.03729059219360351, 0.03724844741821289, 0.03728236770629883, 0.03718352127075195, 0.03737187194824219, 0.037318656921386716, 0.0375186882019043, 0.0374191665649414, 0.037423614501953126, 0.03713148880004883, 0.03875619125366211, 0.03731296157836914, 0.037497344970703124, 0.037945343017578126, 0.0377446403503418, 0.037599231719970705, 0.03726457595825195, 0.037409599304199216, 0.0373996810913086, 0.03803839874267578, 0.037350849151611326, 0.037450302124023435, 0.03759308624267578, 0.03736700820922852, 0.03721091079711914, 0.03732070541381836, 0.03716505432128906, 0.03726131057739258, 0.03702169418334961, 0.03715878295898437, 0.03724476623535156, 0.03731875228881836, 0.037388481140136716, 0.03769900894165039, 0.03761337661743164, 0.0402250862121582, 0.03759711837768555, 0.037513408660888675, 0.037479774475097656, 0.037425537109375, 0.03757497787475586, 0.037986560821533205, 0.03748614501953125, 0.03744198226928711, 0.03753334426879883, 0.037437793731689456, 0.037574657440185545, 0.03746831893920898, 0.03762076950073242, 0.037661376953125, 0.03761695861816406, 0.038282047271728514, 0.03799763107299805, 0.03772099304199219, 0.037873409271240235, 0.03772982406616211, 0.03762400054931641, 0.037997119903564455, 0.037658145904541016, 0.03756284713745117, 0.037582847595214845, 0.037556224822998044, 0.037895198822021484, 0.03768131256103516, 0.037753662109375, 0.037738494873046875, 0.03783065414428711, 0.037730175018310545, 0.03791680145263672, 0.03768016052246094, 0.03784188842773437, 0.03772956848144531, 0.03761190414428711, 0.037937503814697265, 0.039344127655029294, 0.037795841217041014, 0.03765657424926758, 0.037831806182861326, 0.03783155059814453, 0.037550079345703126, 0.03769865417480469, 0.03780495834350586, 0.03763811111450195, 0.03777123260498047, 0.037801185607910154, 0.03765129470825195, 0.03775446319580078, 0.03780239868164063, 0.037765121459960936, 0.037760799407958984, 0.03795785522460938, 0.03786678314208984, 0.03768787384033203, 0.03939139175415039, 0.03813478469848633, 0.03792998504638672, 0.038580223083496096, 0.037754878997802735, 0.03793100738525391, 0.03791872024536133, 0.03775888061523437, 0.03782457733154297, 0.03768112182617187, 0.03767055892944336, 0.038071903228759765]",tokens/s,26.651001314581737,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,7396.02432,7808.679936,0.0,7430.209536,7414.23104,s,1,11.616208984375,11.616208984375,0.0,11.616208984375,11.616208984375,11.616208984375,11.616208984375,[11.616208984375],,kWh,0.00013592565134166532,1.4986332407818976e-05,4.339864583000064e-05,0.00019431062957948493,,MB,1713.31584,8584.626176,0.0,8176.795648,8052.041728,s,10,8.148107971191406,0.8148107971191407,0.0047949582500165676,0.8145799255371093,0.8220338745117187,0.8220945739746094,0.8221431335449219,"[0.8057255249023437, 0.8153632202148438, 0.8114178466796875, 0.8112008056640625, 0.8127526245117187, 0.8146031494140625, 0.8221552734375, 0.8145567016601563, 0.8220203857421875, 0.8183124389648437]",tokens/s,314.18336735978227,kWh,2.38148693461525e-05,2.6263624427553085e-06,1.577599552676885e-05,4.221722731567666e-05,tokens/kWh,6063875.253715175,MB,1718.489088,8731.426816,0.0,8323.596288,8263.496192,s,10,37.15159716796875,3.715159716796875,0.007612190100120031,3.7147086181640625,3.7254406738281247,3.7262239990234374,3.7268506591796875,"[3.707013671875, 3.70830029296875, 3.706685791015625, 3.72264990234375, 3.706224365234375, 3.719031982421875, 3.714274169921875, 3.7252666015625, 3.72700732421875, 3.71514306640625]",tokens/s,16.957548208537624,kWh,0.00010844988760218326,1.196271960611995e-05,7.21384914714313e-05,0.0001925510986797345,tokens/kWh,327185.8765386031,,s,630,37.148054981231695,0.058965166636875704,0.0006265357516154737,0.05891768074035644,0.05953139877319336,0.05977858848571777,0.0611574122619629,"[0.06155952072143555, 0.05867504119873047, 0.05842534255981445, 0.05816038513183594, 0.058156993865966795, 0.058202945709228515, 0.05829609680175781, 0.058293792724609376, 0.058321182250976565, 0.05824143981933594, 0.05830556869506836, 0.05821539306640625, 0.05851094436645508, 0.05906428909301758, 0.05841459274291992, 0.05844220733642578, 0.0591341438293457, 0.05887823867797851, 0.058810367584228515, 0.05874620819091797, 0.058552993774414065, 0.05821388626098633, 0.05816400146484375, 0.058182975769042966, 0.058550464630126954, 0.05833135986328125, 0.05854003143310547, 0.058649600982666014, 0.05864873504638672, 0.058506080627441406, 0.05862604904174805, 0.05876303863525391, 0.058894336700439455, 0.059088191986083984, 0.05876588821411133, 0.058955745697021486, 0.05917103958129883, 0.058931007385253906, 0.058799617767333986, 0.058415969848632815, 0.058423297882080075, 0.058531841278076174, 0.05868339157104492, 0.05872972869873047, 0.05864934539794922, 0.0585994873046875, 0.05894771194458008, 0.05948806381225586, 0.0595513916015625, 0.05966883087158203, 0.05993267059326172, 0.05944934463500977, 0.05970345687866211, 0.05951628875732422, 0.059555679321289065, 0.059445728302001954, 0.05924265670776367, 0.05924863815307617, 0.059393726348876956, 0.05915475082397461, 0.05891228866577149, 0.05901974487304688, 0.059241855621337894, 0.06065334320068359, 0.059480224609375, 0.057825344085693356, 0.05785465621948242, 0.058525184631347656, 0.05789952087402344, 0.05812223815917969, 0.0581646728515625, 0.05834195327758789, 0.058175487518310545, 0.05834332656860351, 0.05842339324951172, 0.05833456039428711, 0.05839734268188477, 0.058218494415283206, 0.058322719573974606, 0.059764896392822266, 0.0593317756652832, 0.05877004623413086, 0.05840544128417969, 0.05844755172729492, 0.05837107086181641, 0.05831987380981445, 0.05828188705444336, 0.05842281723022461, 0.05868339157104492, 0.05853241729736328, 0.05862732696533203, 0.0585893440246582, 0.05847251129150391, 0.058659358978271486, 0.05886566543579102, 0.059184768676757815, 0.05895151901245117, 0.05925328063964844, 0.059104961395263673, 0.05880863952636719, 0.05852959823608399, 0.058856800079345704, 0.058710880279541015, 0.05882463836669922, 0.05909100723266601, 0.0587894401550293, 0.058863903045654295, 0.05890860748291016, 0.059111873626708986, 0.05914396667480469, 0.059137855529785156, 0.05913824081420899, 0.060082176208496096, 0.059323486328125, 0.059208606719970705, 0.05908889770507812, 0.05901926422119141, 0.05927923202514648, 0.05905420684814453, 0.05915852737426758, 0.059486209869384764, 0.05936947250366211, 0.05912588882446289, 0.059232128143310546, 0.06117279815673828, 0.05930192184448242, 0.06105887985229492, 0.058622142791748044, 0.05799468612670899, 0.057872352600097654, 0.057696159362792966, 0.05780080032348633, 0.05805551910400391, 0.05788441467285156, 0.05787065505981445, 0.057974464416503904, 0.058251102447509764, 0.05823430252075195, 0.05854003143310547, 0.058554496765136715, 0.058361953735351565, 0.05870217514038086, 0.058974369049072266, 0.059225601196289064, 0.0589634895324707, 0.0585797119140625, 0.05841123199462891, 0.05850521469116211, 0.058482688903808595, 0.058552127838134765, 0.05846428680419922, 0.05842736053466797, 0.058590721130371094, 0.05858972930908203, 0.05860367965698242, 0.05893116760253906, 0.05871414566040039, 0.05902473449707031, 0.05915619277954102, 0.05919334411621094, 0.05907334518432617, 0.059072544097900394, 0.05904780960083008, 0.05874502563476563, 0.05917004776000977, 0.058428062438964846, 0.059036895751953124, 0.05901609420776367, 0.0588963851928711, 0.05871567916870117, 0.05911190414428711, 0.05909503936767578, 0.059049983978271485, 0.05898649597167969, 0.0591987190246582, 0.0594310417175293, 0.05919964981079102, 0.05942729568481445, 0.05942489624023437, 0.05915225601196289, 0.05937177658081055, 0.059328193664550784, 0.05929587173461914, 0.05936633682250977, 0.0592803840637207, 0.05909503936767578, 0.05956198501586914, 0.059551742553710936, 0.059364673614501956, 0.060958526611328126, 0.058916160583496094, 0.05817638397216797, 0.05812428665161133, 0.05804431915283203, 0.05820630264282227, 0.05828403091430664, 0.0583592643737793, 0.05829891204833984, 0.05842124938964844, 0.05846835327148438, 0.05854534530639648, 0.05888288116455078, 0.05847244644165039, 0.058578495025634766, 0.05874534225463867, 0.05933356857299805, 0.05934796905517578, 0.06441340637207031, 0.058366302490234376, 0.058775489807128906, 0.05864588928222656, 0.058783775329589845, 0.05835222244262695, 0.05841715240478516, 0.058382431030273435, 0.05854304122924805, 0.058673694610595704, 0.05901737594604492, 0.058852863311767575, 0.058970878601074216, 0.059189281463623046, 0.06028006362915039, 0.06014345550537109, 0.05935811233520508, 0.059015167236328124, 0.058916160583496094, 0.05884403228759766, 0.05886137771606445, 0.058902015686035154, 0.058923519134521485, 0.05922304153442383, 0.05905311965942383, 0.059061248779296874, 0.059258975982666016, 0.059345504760742185, 0.05982028961181641, 0.059601119995117184, 0.05956697463989258, 0.05947894287109375, 0.059625473022460934, 0.059381153106689455, 0.05888470458984375, 0.05915852737426758, 0.05959804916381836, 0.05927222442626953, 0.0593098258972168, 0.0591847038269043, 0.05935968017578125, 0.05936268615722656, 0.05915497589111328, 0.05937155151367188, 0.05940966415405274, 0.06278979110717774, 0.058775390625, 0.05820172882080078, 0.05806908798217773, 0.0582064323425293, 0.05831180953979492, 0.058376640319824216, 0.05856512069702149, 0.05990134429931641, 0.0583512954711914, 0.058853790283203124, 0.05920719909667969, 0.05811692810058594, 0.05973622512817383, 0.058156959533691405, 0.058912864685058595, 0.05918243026733398, 0.059128353118896484, 0.05884121704101562, 0.058449920654296876, 0.05839052963256836, 0.05809971237182617, 0.058242240905761716, 0.058135265350341796, 0.05816511917114258, 0.05832931137084961, 0.05834511947631836, 0.05838848114013672, 0.058218494415283206, 0.058300769805908204, 0.05847836685180664, 0.05853411102294922, 0.05921564865112305, 0.058953216552734375, 0.058827327728271483, 0.058544288635253905, 0.05852127838134766, 0.05891449737548828, 0.058582656860351565, 0.0585164794921875, 0.05841100692749023, 0.058482688903808595, 0.058791393280029296, 0.05850284957885742, 0.05887062454223633, 0.059036895751953124, 0.05890127944946289, 0.05867283248901367, 0.05895110321044922, 0.05938470458984375, 0.059237857818603516, 0.0591693115234375, 0.059090240478515625, 0.05936198425292969, 0.05942617416381836, 0.059094718933105465, 0.05923116683959961, 0.05922921752929688, 0.05924249649047852, 0.05928879928588867, 0.059254528045654294, 0.05917900848388672, 0.05922611236572266, 0.06081315231323242, 0.058845279693603515, 0.0583930549621582, 0.05824512100219727, 0.05812211227416992, 0.05831840133666992, 0.05873100662231445, 0.05830047988891601, 0.0584510383605957, 0.05843868637084961, 0.058433406829833986, 0.05864652633666992, 0.05838166427612305, 0.0585447998046875, 0.05848643112182617, 0.05866019058227539, 0.05891897583007812, 0.05880928039550781, 0.05886361694335938, 0.059141887664794925, 0.058908927917480466, 0.05853148651123047, 0.05876361465454102, 0.05870620727539062, 0.05859708786010742, 0.05850316619873047, 0.05856665420532227, 0.061023841857910155, 0.05878806304931641, 0.058601665496826175, 0.058670368194580075, 0.05914303970336914, 0.05934080123901367, 0.059442142486572265, 0.060010398864746094, 0.059560928344726566, 0.05962051010131836, 0.05910025787353516, 0.058909950256347654, 0.05908089447021484, 0.05898796844482422, 0.058581886291503904, 0.059154430389404294, 0.058785888671875, 0.05880531311035156, 0.059206497192382815, 0.059189247131347655, 0.059140094757080076, 0.059858943939208986, 0.060200767517089845, 0.059474113464355466, 0.059514591217041016, 0.05916672134399414, 0.05948963165283203, 0.05960287857055664, 0.05929632186889648, 0.05924070358276367, 0.05914656066894531, 0.05934822463989258, 0.059479808807373045, 0.059233150482177734, 0.059086654663085936, 0.0593031997680664, 0.0615338249206543, 0.059033153533935546, 0.05836611175537109, 0.05816384124755859, 0.05827151870727539, 0.058477920532226564, 0.058380416870117184, 0.058315521240234376, 0.05854313659667969, 0.058411998748779295, 0.0584252815246582, 0.058460224151611326, 0.05854003143310547, 0.058662750244140624, 0.05869561767578125, 0.05872867202758789, 0.06111974334716797, 0.05922275161743164, 0.05884112167358398, 0.05951897430419922, 0.058265598297119144, 0.05832400131225586, 0.05837641525268555, 0.06045929718017578, 0.058923553466796875, 0.05846214294433594, 0.05923964691162109, 0.05889459228515625, 0.05879657745361328, 0.058711360931396485, 0.058585727691650394, 0.059168830871582034, 0.059138046264648435, 0.05899059295654297, 0.059187137603759765, 0.05924665451049805, 0.05937958526611328, 0.05906854248046875, 0.05909881591796875, 0.05894294357299805, 0.05873715209960938, 0.05887961578369141, 0.05886816024780273, 0.05893545532226562, 0.058982433319091795, 0.05915558242797852, 0.05905097579956055, 0.059133918762207034, 0.05932147216796875, 0.05939007949829102, 0.059253631591796876, 0.05886111831665039, 0.05902380752563476, 0.05880617523193359, 0.058859424591064455, 0.05891638565063476, 0.05877203369140625, 0.05888819122314453, 0.05884108734130859, 0.058922721862792966, 0.0590269775390625, 0.05924236679077149, 0.059103744506835934, 0.06097020721435547, 0.05860214233398438, 0.058246337890625, 0.05826950454711914, 0.05827276611328125, 0.05852150344848633, 0.05854601669311523, 0.05842559814453125, 0.05843555068969727, 0.05841084671020508, 0.05868563079833984, 0.05878374481201172, 0.05855641555786133, 0.058687393188476565, 0.058687583923339844, 0.05895372772216797, 0.05952899169921875, 0.05929391860961914, 0.05909503936767578, 0.05888143920898437, 0.05864486312866211, 0.058719905853271484, 0.058654815673828124, 0.05867497634887695, 0.05879167938232422, 0.0588419189453125, 0.05872243118286133, 0.05879769515991211, 0.0589725456237793, 0.05883881759643555, 0.05916422271728516, 0.05909523010253906, 0.059349441528320314, 0.05940636825561523, 0.05969465637207031, 0.05945180892944336, 0.06079286575317383, 0.05905408096313477, 0.059038719177246096, 0.058993663787841794, 0.059224063873291016, 0.05899059295654297, 0.059082752227783204, 0.059117568969726565, 0.059284736633300784, 0.05917567825317383, 0.05924172973632812, 0.059286270141601566, 0.05964556884765625, 0.05948387145996094, 0.0597429428100586, 0.06101417541503906, 0.059641441345214846, 0.0594312973022461, 0.059393856048583986, 0.05938969421386719, 0.05925628662109375, 0.059466529846191406, 0.05953692626953125, 0.059473983764648436, 0.059307998657226565, 0.0595307846069336, 0.0595874252319336, 0.06163027191162109, 0.05904150390625, 0.05828492736816406, 0.05820182418823242, 0.05837564849853515, 0.05839689636230469, 0.058415233612060545, 0.05839632034301758, 0.058294815063476564, 0.058659103393554686, 0.05864448165893555, 0.05849257659912109, 0.05848099136352539, 0.0585780143737793, 0.05862617492675781, 0.059210529327392576, 0.0595283203125, 0.059581310272216796, 0.05913600158691406, 0.05858508682250976, 0.05878521728515625, 0.05877814483642578, 0.05881449508666992, 0.058714111328125, 0.058729759216308595, 0.058646366119384764, 0.05902012634277344, 0.058834976196289065, 0.05893939208984375, 0.05880831909179687, 0.058793983459472655, 0.059264671325683596, 0.05997321701049805, 0.059357505798339844, 0.059522560119628906, 0.059585472106933594, 0.059057472229003906, 0.05913792037963867, 0.05922694396972656, 0.06008553695678711, 0.05900931167602539, 0.058937793731689454, 0.05936291122436523, 0.05947843170166016, 0.05976063919067383, 0.05942195129394531, 0.059398910522460935, 0.059434078216552735, 0.05926937484741211, 0.05965420913696289, 0.0598185920715332, 0.05958860778808594, 0.06053443145751953, 0.059662303924560546, 0.05959100723266601, 0.05938998413085937, 0.05929388809204102, 0.05929555130004883, 0.05931622314453125, 0.05920953750610351, 0.059211551666259764, 0.059514816284179685, 0.059789791107177734, 0.06131824111938477, 0.059089759826660156, 0.05840825653076172, 0.05809836959838867, 0.057864192962646485, 0.05793571090698242, 0.05817993545532227, 0.05835260772705078, 0.05852451324462891, 0.058519615173339846, 0.058560192108154295, 0.05868803024291992, 0.05860086441040039, 0.058681312561035155, 0.05842319869995117, 0.05902975845336914, 0.059226303100585936, 0.05947910308837891, 0.05885971069335937, 0.05865500640869141, 0.05839510345458984, 0.058243038177490235, 0.058744609832763674, 0.058582527160644535, 0.05859814453125, 0.05870182418823242, 0.058904064178466796, 0.05861996841430664, 0.058671550750732424, 0.05858246231079101, 0.058624576568603516, 0.0590909423828125, 0.05885721588134766, 0.05910124969482422, 0.059375808715820315, 0.05897830581665039, 0.05899059295654297, 0.05883903884887695, 0.05895372772216797, 0.05882463836669922, 0.05910739135742187, 0.059108768463134766, 0.05910905456542969, 0.05932271957397461, 0.059011646270751957, 0.058931137084960936, 0.05908832168579101, 0.05925104141235352, 0.059308319091796874, 0.059809185028076174, 0.0594950065612793, 0.0597437744140625, 0.0594967041015625, 0.05938943862915039, 0.059377822875976566, 0.05933932876586914, 0.05953740692138672, 0.05930704116821289, 0.059146976470947264, 0.05948031997680664, 0.05926707077026367, 0.05953945541381836, 0.05947391891479492]",tokens/s,16.959165165398158,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,8202.801152,11078.139904,0.0,10699.669504,10468.411392,s,1,13.217875,13.217875,0.0,13.217875,13.217875,13.217875,13.217875,[13.217875],,kWh,0.00018690370244588242,2.0609268657214217e-05,5.958199211003601e-05,0.0002670949632131327,,MB,4041.961472,11570.970624,0.0,11163.140096,10922.852352,s,10,8.030579650878906,0.8030579650878906,0.00806742905327945,0.8072570495605469,0.8102368774414063,0.8103816223144531,0.8104974182128907,"[0.788420654296875, 0.7902467651367188, 0.7945225830078125, 0.8073043823242188, 0.8102047119140625, 0.8064562377929687, 0.807209716796875, 0.8076004028320313, 0.8105263671875, 0.8080878295898437]",tokens/s,318.7814717359563,kWh,2.336502454327157e-05,2.576762183389112e-06,1.549719188492355e-05,4.143897861158423e-05,tokens/kWh,6177758.4433134515,MB,4046.286848,11573.067776,0.0,11165.237248,10922.854912,s,10,37.21511938476563,3.7215119384765627,0.012686653035861332,3.7212767333984376,3.7377667236328125,3.7392240600585938,3.740389929199219,"[3.70255322265625, 3.727070556640625, 3.731174072265625, 3.729641845703125, 3.71548291015625, 3.71462255859375, 3.73744287109375, 3.70910986328125, 3.740681396484375, 3.707340087890625]",tokens/s,16.928603492748614,kWh,0.00010802571604214453,1.1915910298840572e-05,7.187379467847737e-05,0.00019181542101946248,tokens/kWh,328440.7461358789,,s,630,37.211775604248054,0.059066310482933405,0.0008960666131213604,0.05889635276794433,0.05975107345581055,0.06053022289276123,0.06229926246643067,"[0.06148745727539062, 0.05839769744873047, 0.05801062393188477, 0.05809151840209961, 0.057753280639648436, 0.057452064514160156, 0.06068479919433594, 0.05785164642333984, 0.05817164611816406, 0.057743263244628903, 0.058143104553222656, 0.062117889404296876, 0.05817702484130859, 0.05793622589111328, 0.05767375946044922, 0.058060928344726564, 0.05828403091430664, 0.057869503021240234, 0.05786297607421875, 0.05798428726196289, 0.058347423553466796, 0.05810054397583008, 0.058244831085205076, 0.05803023910522461, 0.058396446228027345, 0.05840332794189453, 0.05850812911987305, 0.05859254455566406, 0.0584901123046875, 0.05854051208496094, 0.058431488037109375, 0.05866291046142578, 0.05899257659912109, 0.05898783874511719, 0.05926934432983398, 0.05922243118286133, 0.059224193572998046, 0.05875302505493164, 0.058918880462646483, 0.05870595169067383, 0.05883084869384766, 0.05884486389160156, 0.058886463165283204, 0.0588963851928711, 0.05958412933349609, 0.058941825866699216, 0.060030975341796876, 0.05910704040527344, 0.059033119201660156, 0.05949747085571289, 0.05952486419677734, 0.05921791839599609, 0.05939199829101562, 0.059873279571533204, 0.059002239227294924, 0.058972801208496094, 0.059009025573730466, 0.058865215301513674, 0.05890816116333008, 0.05878879928588867, 0.05885542297363281, 0.05877334213256836, 0.05882015991210938, 0.0621143684387207, 0.06004108810424805, 0.059168224334716794, 0.05882067108154297, 0.05891078567504883, 0.05901980972290039, 0.05900697708129883, 0.05937356948852539, 0.05915852737426758, 0.0599183349609375, 0.05907660675048828, 0.05938751983642578, 0.058794368743896486, 0.058826366424560544, 0.05890403366088867, 0.05902428817749023, 0.05897987365722656, 0.05902998352050781, 0.05956198501586914, 0.05869158554077149, 0.05967248153686523, 0.05871830368041992, 0.05885542297363281, 0.059089950561523434, 0.059513278961181644, 0.05967484664916992, 0.058891616821289065, 0.05886198425292969, 0.05841936111450195, 0.05863222503662109, 0.05855065536499023, 0.05847379302978516, 0.05850182342529297, 0.05897321701049805, 0.05884822463989258, 0.05939523315429687, 0.05896892929077149, 0.05978678512573242, 0.05867340850830078, 0.059109375, 0.05916694259643555, 0.05859532928466797, 0.05972172927856445, 0.05910454559326172, 0.058858207702636715, 0.05877145767211914, 0.05887295913696289, 0.06152076721191406, 0.05909670257568359, 0.059158145904541014, 0.05913043212890625, 0.060088287353515624, 0.059019489288330076, 0.058439678192138675, 0.05845766448974609, 0.058453601837158205, 0.0587149772644043, 0.058611358642578125, 0.05971798324584961, 0.05981184005737305, 0.059211071014404294, 0.05922067260742187, 0.05956512069702148, 0.06141548919677734, 0.059000831604003906, 0.058933246612548826, 0.05879811096191406, 0.058853343963623045, 0.05904793548583984, 0.05893280029296875, 0.05906259155273438, 0.059022815704345706, 0.059224735260009764, 0.05933260726928711, 0.05895596694946289, 0.058969825744628904, 0.059188896179199216, 0.0590709114074707, 0.061290401458740235, 0.06307171249389648, 0.059205535888671876, 0.060450817108154295, 0.05887583923339844, 0.05871641540527344, 0.058433536529541016, 0.05965641784667969, 0.05984508895874024, 0.05876617431640625, 0.058659839630126956, 0.05981568145751953, 0.059281665802001955, 0.05909708786010742, 0.05883622360229492, 0.058813182830810544, 0.058920673370361325, 0.05911580657958984, 0.059250526428222657, 0.059749790191650394, 0.059178783416748044, 0.06074265670776367, 0.05875299072265625, 0.058916862487792966, 0.05852182388305664, 0.0589453125, 0.059115169525146484, 0.058759105682373046, 0.05914835357666016, 0.05936777496337891, 0.05912371063232422, 0.05884928131103516, 0.05907059097290039, 0.059008895874023436, 0.059009025573730466, 0.059066368103027345, 0.05971318435668945, 0.05937392044067383, 0.059386943817138674, 0.058952640533447266, 0.05868134307861328, 0.058627967834472654, 0.05875711822509765, 0.058676799774169924, 0.05889900970458985, 0.05944729614257813, 0.05862944030761719, 0.05846435165405273, 0.06137913513183594, 0.058831134796142576, 0.058518657684326174, 0.05842380905151367, 0.05861824035644531, 0.05839379119873047, 0.058579551696777345, 0.05876863861083984, 0.05915337753295898, 0.05929759979248047, 0.05909523010253906, 0.05988108825683594, 0.05895750427246094, 0.05969113540649414, 0.059652671813964844, 0.059045345306396484, 0.05915875244140625, 0.059713855743408206, 0.05913727951049805, 0.05888691329956055, 0.05917900848388672, 0.05899468612670898, 0.059385185241699216, 0.05974697494506836, 0.05826544189453125, 0.058966175079345706, 0.05904115295410156, 0.05866969680786133, 0.0590643196105957, 0.05826342391967773, 0.0584664306640625, 0.05865267181396484, 0.05910732650756836, 0.05890867233276367, 0.059920543670654296, 0.05942256164550781, 0.06433702087402343, 0.05947824096679687, 0.05933737564086914, 0.05907251358032226, 0.059504638671875, 0.059979774475097655, 0.05989779281616211, 0.05967059326171875, 0.059504638671875, 0.059186752319335935, 0.05889446258544922, 0.05865875244140625, 0.05887622451782227, 0.05861548614501953, 0.05877577590942383, 0.059228321075439454, 0.05879107284545899, 0.058850143432617186, 0.058850753784179685, 0.05903984069824219, 0.05906832122802735, 0.05890105438232422, 0.05916876983642578, 0.059582462310791014, 0.059463390350341795, 0.05868163299560547, 0.058644222259521483, 0.06125766372680664, 0.059450111389160155, 0.05873459243774414, 0.05852691268920898, 0.05923667144775391, 0.058511039733886716, 0.05846928024291992, 0.05891635131835937, 0.05924204635620117, 0.05925523376464844, 0.05919171142578125, 0.05929779052734375, 0.05943270492553711, 0.05909529495239258, 0.058687488555908204, 0.05852364730834961, 0.05944934463500977, 0.05906227111816406, 0.058961536407470705, 0.05867507171630859, 0.05802854537963867, 0.057915393829345706, 0.05797846221923828, 0.05810790252685547, 0.058535839080810545, 0.05831219100952149, 0.05841574478149414, 0.05820230484008789, 0.05950073623657227, 0.059246593475341794, 0.0585274543762207, 0.06479293060302735, 0.061724609375, 0.058499073028564455, 0.0591313591003418, 0.0588191032409668, 0.05864243316650391, 0.0583383674621582, 0.060135902404785155, 0.058597694396972655, 0.05850537490844727, 0.05822649765014649, 0.060080257415771485, 0.058770942687988284, 0.058556991577148436, 0.058456062316894535, 0.058494655609130856, 0.058386016845703125, 0.058689376831054685, 0.05837257766723633, 0.058896320343017575, 0.05924256134033203, 0.059244384765625, 0.058618144989013674, 0.058640670776367185, 0.05859692764282227, 0.058898239135742186, 0.05860416030883789, 0.05896806335449219, 0.05874470520019531, 0.058901695251464846, 0.0589813117980957, 0.05886569595336914, 0.061982719421386716, 0.059232257843017576, 0.05973606491088867, 0.05940633773803711, 0.05938796615600586, 0.058988479614257815, 0.060635009765625, 0.05931020736694336, 0.05919948959350586, 0.06010879898071289, 0.0595333137512207, 0.05893734359741211, 0.05946540832519531, 0.059334976196289066, 0.05990195083618164, 0.059099136352539064, 0.05928345489501953, 0.05884438323974609, 0.05868531036376953, 0.0585, 0.05860086441040039, 0.058456672668457034, 0.05858256149291992, 0.058445377349853514, 0.05856694412231445, 0.0584620475769043, 0.05965030288696289, 0.0588719367980957, 0.05871651077270508, 0.05887964630126953, 0.059578334808349606, 0.059254783630371094, 0.05876780700683594, 0.059535358428955076, 0.0590274543762207, 0.05906774520874023, 0.05877622222900391, 0.058910400390625, 0.05836572647094727, 0.05840131378173828, 0.05811977767944336, 0.058458526611328124, 0.05872953414916992, 0.058846336364746094, 0.058482398986816404, 0.05817967987060547, 0.05842124938964844, 0.05864876937866211, 0.0589944953918457, 0.05861785507202148, 0.05851359939575195, 0.05871948623657226, 0.0588304328918457, 0.058948577880859374, 0.05882022476196289, 0.05909952163696289, 0.058520896911621094, 0.05827980804443359, 0.05854864120483398, 0.05842975997924805, 0.0588043212890625, 0.058601280212402344, 0.05820435333251953, 0.06160412979125977, 0.05942937469482422, 0.06014976119995117, 0.0587325439453125, 0.058226688385009766, 0.05843344116210938, 0.05847065734863281, 0.0582562255859375, 0.05818982315063476, 0.05829119873046875, 0.05836387252807617, 0.058342750549316404, 0.058309310913085936, 0.05845811080932617, 0.05847017669677734, 0.05849724960327148, 0.059115520477294924, 0.05895372772216797, 0.059114688873291014, 0.058640480041503906, 0.05905686569213867, 0.0585011215209961, 0.058621952056884766, 0.05861785507202148, 0.059415809631347655, 0.05901798248291015, 0.05874016189575195, 0.059027809143066406, 0.05885507202148438, 0.05876588821411133, 0.058750175476074216, 0.058938144683837894, 0.06077030563354492, 0.0592342414855957, 0.059651966094970706, 0.05981817626953125, 0.059170814514160154, 0.058861568450927736, 0.05856051254272461, 0.05865884780883789, 0.05893423843383789, 0.058668033599853515, 0.059151966094970705, 0.059351455688476565, 0.0596049919128418, 0.059762622833251955, 0.059522209167480467, 0.06239292907714844, 0.06011897659301758, 0.06503462219238282, 0.05987641525268555, 0.05971039962768555, 0.05979702377319336, 0.05911520004272461, 0.06146476745605469, 0.05952163314819336, 0.05940838241577148, 0.059270721435546875, 0.06237334442138672, 0.06083478546142578, 0.05881241607666016, 0.05853936004638672, 0.05875510406494141, 0.06188246536254883, 0.05904377746582031, 0.05880428695678711, 0.05875302505493164, 0.058883262634277345, 0.05865555191040039, 0.05869881439208984, 0.05910214233398438, 0.059187198638916014, 0.05916876983642578, 0.05933407974243164, 0.05905215835571289, 0.058456321716308594, 0.0587204475402832, 0.058984447479248046, 0.05925033569335938, 0.058769760131835935, 0.05856256103515625, 0.05931008148193359, 0.058744831085205076, 0.05883903884887695, 0.05842057418823242, 0.058263359069824217, 0.05827423858642578, 0.05841961669921875, 0.0583939208984375, 0.0585645751953125, 0.05918406295776367, 0.05863529586791992, 0.05868560028076172, 0.05875772857666016, 0.05863977432250977, 0.05889904022216797, 0.05888175964355469, 0.05942915344238281, 0.05887340927124023, 0.059035743713378906, 0.05888444900512695, 0.059009025573730466, 0.05874687957763672, 0.05894144058227539, 0.05905606460571289, 0.05871760177612305, 0.05906703948974609, 0.05923404693603516, 0.058595680236816404, 0.05845596694946289, 0.05836111831665039, 0.0585489616394043, 0.059541439056396486, 0.05853590393066406, 0.05860905456542969, 0.05859187316894531, 0.05868495941162109, 0.058649120330810545, 0.05881388854980469, 0.05876588821411133, 0.05861904144287109, 0.058815135955810546, 0.05882489776611328, 0.05858508682250976, 0.05955788803100586, 0.05897625732421875, 0.06199705505371094, 0.058982398986816405, 0.05885478210449219, 0.0586901741027832, 0.058672416687011716, 0.058995136260986326, 0.060391105651855466, 0.058505825042724606, 0.058257408142089843, 0.05820006561279297, 0.05845196914672852, 0.058363903045654295, 0.058517505645751956, 0.05855641555786133, 0.05856256103515625, 0.059637599945068356, 0.059130016326904296, 0.05876646423339844, 0.061150081634521486, 0.05935923385620117, 0.05931808090209961, 0.059316417694091794, 0.05909836959838867, 0.05848163223266602, 0.0583818244934082, 0.05815059280395508, 0.058356319427490234, 0.058517505645751956, 0.058498817443847655, 0.0590456657409668, 0.05875763320922851, 0.05862396621704102, 0.05898649597167969, 0.05886361694335938, 0.05864652633666992, 0.05861171340942383, 0.06049577713012695, 0.05860976028442383, 0.058671104431152345, 0.0591646728515625, 0.059351295471191404, 0.05904358291625977, 0.05899468612670898, 0.05973116683959961, 0.05936822509765625, 0.059512832641601565, 0.06052601623535156, 0.06142214584350586, 0.061736831665039064, 0.06036249542236328, 0.060380958557128904, 0.059832542419433594, 0.06110179138183594, 0.06821932983398438, 0.05889766311645508, 0.0588175048828125, 0.05872140884399414, 0.05887788772583008, 0.05913081741333008, 0.0587243537902832, 0.059791358947753906, 0.058616928100585934, 0.06053366470336914, 0.061727550506591795, 0.05840281677246094, 0.058482654571533205, 0.05823855972290039, 0.05853638458251953, 0.058603134155273434, 0.058799583435058596, 0.05871283340454102, 0.058775711059570315, 0.05856665420532227, 0.059114593505859375, 0.05909337615966797, 0.05887030410766601, 0.05895167922973633, 0.05930294418334961, 0.05924272155761719, 0.059127967834472654, 0.05894384002685547, 0.05953971099853515, 0.05897216033935547, 0.05927459335327148, 0.05896259307861328, 0.058907745361328125, 0.05912236785888672, 0.05907849502563477, 0.05875545501708984, 0.058810367584228515, 0.05933260726928711, 0.059660289764404295, 0.05936947250366211, 0.05878521728515625, 0.05932304000854492, 0.05824812698364258, 0.05862089538574219, 0.05846323013305664, 0.05850009536743164, 0.058267742156982424, 0.058588481903076174, 0.0585643196105957, 0.05836489486694336, 0.058611488342285155, 0.058406944274902346, 0.05826473617553711, 0.05829462432861328, 0.05865900802612305, 0.05884355163574219, 0.058461406707763675, 0.058616798400878904, 0.05870937728881836, 0.05849747085571289, 0.05893734359741211, 0.058686622619628905, 0.05922083282470703, 0.05925299072265625, 0.05899411010742187, 0.058732864379882815, 0.058705982208251954, 0.058576831817626955, 0.05893734359741211, 0.05851107025146484, 0.058321182250976565, 0.05839052963256836, 0.059356544494628904]",tokens/s,16.930124665378237,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,814.710784,537.853952,0.0,159.383552,141.514752,s,1,7.5653583984375,7.5653583984375,0.0,7.5653583984375,7.5653583984375,7.5653583984375,7.5653583984375,[7.5653583984375],,kWh,1.7524344425002408e-05,1.92576686632031e-06,5.607504486027093e-06,2.505761577734981e-05,,MB,1168.605184,630.12864,0.0,222.298112,184.525824,s,21,0.18403315448760987,0.00876348354702904,0.00010122327704860388,0.008751775741577148,0.00884108829498291,0.008887968063354491,0.009068320083618164,"[0.008619168281555175, 0.008813952445983887, 0.008810527801513672, 0.008887968063354491, 0.00884108829498291, 0.008804320335388183, 0.008806624412536622, 0.008778623580932617, 0.008696288108825683, 0.008756352424621581, 0.0086779203414917, 0.008693696022033692, 0.008673760414123535, 0.0087258882522583, 0.008704000473022461, 0.008711711883544922, 0.008771967887878419, 0.009113408088684083, 0.008683199882507325, 0.008751775741577148, 0.008710911750793457]",tokens/s,29212.12764606468,kWh,2.635768096996703e-07,2.906783153199022e-08,1.4802280234868783e-07,4.4066744358034836e-07,tokens/kWh,580936948.5525033,MB,1202.335744,630.12864,0.0,222.298112,184.528384,s,21,10.169293518066407,0.4842520722888765,0.004762281942675445,0.4825516967773438,0.4896417236328125,0.49239239501953125,0.4959256713867187,"[0.4825516967773438, 0.48272607421875, 0.48962738037109377, 0.49239239501953125, 0.48653704833984374, 0.4896417236328125, 0.4887113037109375, 0.48062728881835937, 0.4797857666015625, 0.47930374145507815, 0.4813526611328125, 0.47868646240234375, 0.48649685668945314, 0.48127764892578123, 0.48208526611328123, 0.48090081787109373, 0.4968089904785156, 0.4863446044921875, 0.4827623291015625, 0.48046170043945313, 0.48021176147460937]",tokens/s,130.097533093091,kWh,1.4080092215300685e-05,1.5527927379002416e-06,5.1344020585087585e-06,2.0767287011709686e-05,tokens/kWh,3033617.2444902067,,s,1323,10.159735195636735,0.007679316096475245,0.0001623395081562752,0.007635615825653076,0.007807667064666748,0.007904799890518188,0.008208186264038086,"[0.007541152000427246, 0.007700319766998291, 0.0076250238418579105, 0.007627200126647949, 0.007585536003112793, 0.007609824180603028, 0.007650239944458008, 0.007603392124176025, 0.007580319881439209, 0.007617760181427002, 0.007579648017883301, 0.007595104217529297, 0.007599296092987061, 0.0076121602058410644, 0.007656191825866699, 0.007610367774963379, 0.007623744010925293, 0.007621119976043702, 0.007637440204620361, 0.007581984043121338, 0.0076341118812561035, 0.007616479873657226, 0.007661632061004639, 0.007622911930084229, 0.007602176189422607, 0.007564703941345215, 0.007719776153564453, 0.0076267518997192385, 0.007608352184295654, 0.007757343769073486, 0.007627200126647949, 0.007624703884124756, 0.007620768070220947, 0.007608160018920899, 0.007636159896850586, 0.007686240196228027, 0.00769916820526123, 0.0076958398818969724, 0.007666431903839112, 0.007680928230285645, 0.007764863967895508, 0.007874335765838623, 0.007758048057556153, 0.007652383804321289, 0.007668704032897949, 0.00765340805053711, 0.007739359855651855, 0.007687200069427491, 0.007699423789978027, 0.007718912124633789, 0.007665599822998047, 0.007698783874511719, 0.007673503875732422, 0.007594048023223877, 0.007575551986694336, 0.007605247974395752, 0.007871488094329833, 0.007667712211608887, 0.007609439849853516, 0.007615615844726562, 0.007642911911010742, 0.007706624031066894, 0.007745535850524903, 0.00751804780960083, 0.007632063865661621, 0.0076169600486755374, 0.007631360054016113, 0.007603295803070068, 0.007676767826080322, 0.007614208221435547, 0.007616672039031982, 0.007597951889038086, 0.007614048004150391, 0.007627552032470703, 0.007665215969085693, 0.007756063938140869, 0.0076574721336364745, 0.0076433920860290525, 0.0076583361625671385, 0.007658400058746338, 0.0076631360054016115, 0.007623136043548584, 0.007682079792022705, 0.007636960029602051, 0.00765337610244751, 0.007630591869354248, 0.007694591999053955, 0.007666719913482666, 0.007670752048492432, 0.0076943039894104, 0.007659552097320557, 0.007608320236206055, 0.007697855949401856, 0.007651904106140137, 0.007632895946502686, 0.007628799915313721, 0.007636991977691651, 0.007626175880432129, 0.007700672149658203, 0.0077132158279418946, 0.007683328151702881, 0.007742080211639404, 0.007776319980621338, 0.007727424144744873, 0.007689023971557617, 0.007744671821594239, 0.007616223812103272, 0.007611455917358398, 0.007644095897674561, 0.007663616180419922, 0.007673855781555176, 0.007643136024475097, 0.007653664112091065, 0.007687839984893799, 0.007651360034942627, 0.007770016193389893, 0.007626560211181641, 0.007582272052764892, 0.00760972785949707, 0.007633279800415039, 0.007611936092376709, 0.007625343799591064, 0.007640927791595459, 0.00766323184967041, 0.007666048049926758, 0.007673855781555176, 0.007611904144287109, 0.0077927041053771975, 0.007643551826477051, 0.00762883186340332, 0.007662784099578857, 0.007631680011749267, 0.007603775978088379, 0.0077413439750671385, 0.007680543899536133, 0.007703999996185303, 0.007674752235412597, 0.007718592166900635, 0.007790592193603516, 0.007776319980621338, 0.007855519771575928, 0.007801504135131836, 0.007698304176330566, 0.007677951812744141, 0.007681695938110351, 0.007706975936889649, 0.007765408039093018, 0.007733856201171875, 0.0077619199752807615, 0.007759871959686279, 0.007782400131225586, 0.00773529577255249, 0.007708672046661377, 0.007812320232391357, 0.007707039833068848, 0.007704927921295166, 0.007755104064941406, 0.007829343795776367, 0.007740543842315674, 0.007765279769897461, 0.007694784164428711, 0.007734784126281738, 0.0076839041709899905, 0.007686463832855225, 0.0077535037994384765, 0.007719520092010498, 0.007786240100860596, 0.007722623825073242, 0.0077355198860168456, 0.007979423999786376, 0.008074687957763673, 0.007918144226074219, 0.007796735763549805, 0.007774367809295655, 0.007888959884643556, 0.00788431978225708, 0.007905439853668213, 0.007957791805267334, 0.0077955198287963864, 0.007780352115631104, 0.007809184074401856, 0.00776800012588501, 0.007775519847869873, 0.008122271537780761, 0.0077339839935302735, 0.007741439819335938, 0.007710720062255859, 0.007755775928497314, 0.007802879810333252, 0.007841760158538819, 0.0077291841506958005, 0.007776512145996094, 0.007683839797973633, 0.00773737621307373, 0.007706624031066894, 0.00774121618270874, 0.007734975814819336, 0.007704544067382813, 0.007717088222503662, 0.007870783805847168, 0.0076977920532226565, 0.0076778879165649416, 0.007665631771087647, 0.007731935977935791, 0.007725120067596436, 0.007729087829589844, 0.007706240177154541, 0.0077255358695983885, 0.0077003841400146485, 0.007716864109039307, 0.00777350378036499, 0.007801536083221435, 0.0077814397811889644, 0.0078121919631958005, 0.007761695861816406, 0.007700479984283447, 0.007725120067596436, 0.007716991901397705, 0.007703680038452149, 0.007703296184539795, 0.00766761589050293, 0.007681600093841553, 0.007631167888641357, 0.007667744159698486, 0.007669151782989502, 0.007664415836334228, 0.007718912124633789, 0.00775548791885376, 0.007733823776245118, 0.007736991882324218, 0.007691520214080811, 0.0077012481689453125, 0.00880031967163086, 0.009869312286376953, 0.009905247688293458, 0.007738399982452393, 0.007894911766052247, 0.007857344150543212, 0.00778323221206665, 0.007696032047271728, 0.007735648155212402, 0.007725056171417236, 0.007768191814422607, 0.007703648090362549, 0.007668799877166748, 0.007654655933380127, 0.007659999847412109, 0.007660704135894775, 0.007648096084594726, 0.0076574721336364745, 0.007654592037200928, 0.0077405118942260745, 0.007692287921905518, 0.007661568164825439, 0.007628384113311767, 0.007661471843719483, 0.007672319889068604, 0.007695807933807373, 0.008139552116394043, 0.007678751945495606, 0.007654111862182617, 0.007657120227813721, 0.00774348783493042, 0.007709311962127686, 0.007665664196014404, 0.007648447990417481, 0.007764800071716309, 0.0076201920509338375, 0.007629216194152832, 0.0076340799331665035, 0.007725887775421143, 0.007716896057128907, 0.007644320011138916, 0.007694560050964356, 0.007688543796539307, 0.00764358377456665, 0.007672832012176513, 0.007673823833465576, 0.007684671878814697, 0.007670176029205322, 0.007692192077636719, 0.007651328086853027, 0.007687935829162598, 0.007647200107574463, 0.007639328002929688, 0.007661568164825439, 0.007706560134887695, 0.007729504108428955, 0.007710432052612305, 0.007667327880859375, 0.00775222396850586, 0.007814496040344239, 0.00772873592376709, 0.00781609582901001, 0.007769375801086426, 0.00785264015197754, 0.007722335815429688, 0.007713088035583496, 0.007707104206085205, 0.007793951988220215, 0.0077012162208557125, 0.00765337610244751, 0.0076984319686889645, 0.00783126401901245, 0.007839168071746826, 0.007703519821166992, 0.007725120067596436, 0.007703839778900146, 0.0077309122085571285, 0.007723775863647461, 0.007980447769165039, 0.007748608112335205, 0.0077638721466064455, 0.007704319953918457, 0.007731200218200684, 0.007651167869567871, 0.007687744140625, 0.007674464225769043, 0.007723008155822754, 0.00764089584350586, 0.0077355198860168456, 0.007673920154571533, 0.007714719772338867, 0.007792640209197998, 0.007777344226837158, 0.007750239849090576, 0.0077948799133300785, 0.007841055870056152, 0.007806848049163819, 0.007840767860412597, 0.00788803195953369, 0.007926112174987792, 0.00793228816986084, 0.007962751865386962, 0.00789904022216797, 0.00785584020614624, 0.007788928031921387, 0.00783510398864746, 0.007856671810150146, 0.00781715202331543, 0.007778304100036621, 0.007821375846862794, 0.007755936145782471, 0.007792575836181641, 0.007822368144989014, 0.007814239978790284, 0.007796703815460205, 0.007763775825500488, 0.007745535850524903, 0.007867775917053222, 0.007715456008911133, 0.007718783855438232, 0.007699935913085937, 0.007688864231109619, 0.007730847835540772, 0.007702879905700684, 0.007731008052825928, 0.0077582402229309086, 0.007722303867340088, 0.007735775947570801, 0.00772540807723999, 0.007771711826324463, 0.007793951988220215, 0.007788512229919433, 0.007748000144958496, 0.007700928211212158, 0.007716864109039307, 0.007683775901794434, 0.007691616058349609, 0.007674975872039795, 0.007671743869781494, 0.007763904094696045, 0.007734560012817383, 0.007713503837585449, 0.007710720062255859, 0.007716127872467041, 0.0077339839935302735, 0.007788415908813477, 0.007615839958190918, 0.007761695861816406, 0.007815455913543701, 0.007823296070098877, 0.007750432014465332, 0.007779520034790039, 0.007767007827758789, 0.007808735847473144, 0.00781708812713623, 0.007791071891784668, 0.007738272190093994, 0.007693056106567383, 0.007925600051879882, 0.007817408084869385, 0.007907423973083496, 0.007763264179229737, 0.007767648220062256, 0.007791584014892578, 0.007733248233795166, 0.00769814395904541, 0.007681920051574707, 0.007719327926635742, 0.007759871959686279, 0.007692287921905518, 0.007711999893188477, 0.007717631816864014, 0.007663008213043213, 0.007697247982025147, 0.0077552638053894046, 0.007721216201782227, 0.007696383953094482, 0.007718976020812988, 0.007661759853363037, 0.007693408012390136, 0.007672800064086914, 0.007694015979766846, 0.007688191890716553, 0.007723008155822754, 0.0077021121978759766, 0.007708992004394531, 0.00765715217590332, 0.007765920162200928, 0.007721183776855468, 0.007842080116271972, 0.008282079696655273, 0.008200223922729493, 0.007775296211242675, 0.007721920013427735, 0.007753215789794922, 0.007682559967041015, 0.0076938881874084476, 0.007705056190490723, 0.007745344161987305, 0.007708831787109375, 0.007683167934417725, 0.00780787181854248, 0.007668863773345947, 0.007672736167907715, 0.007640575885772705, 0.0078504958152771, 0.007702527999877929, 0.007659520149230957, 0.007677343845367432, 0.007520864009857178, 0.007653279781341553, 0.0076425280570983885, 0.007659488201141358, 0.007627744197845459, 0.007685887813568115, 0.00767955207824707, 0.007639391899108887, 0.007645343780517578, 0.007694272041320801, 0.007607903957366944, 0.007620480060577393, 0.0076130561828613285, 0.0076512317657470704, 0.007613823890686035, 0.007576511859893799, 0.007618239879608154, 0.007587039947509765, 0.007577792167663574, 0.007602784156799316, 0.00763478422164917, 0.007631008148193359, 0.007632959842681885, 0.007630208015441895, 0.007692448139190674, 0.007661983966827392, 0.007609568119049072, 0.007643936157226562, 0.007632607936859131, 0.007616223812103272, 0.007592512130737305, 0.00760364818572998, 0.007563839912414551, 0.007628575801849365, 0.007613887786865234, 0.007717663764953613, 0.007589888095855713, 0.007616064071655273, 0.007598527908325195, 0.007586016178131104, 0.007620575904846191, 0.007584928035736084, 0.007686816215515137, 0.00759603214263916, 0.007669727802276611, 0.007622176170349121, 0.007629631996154785, 0.00764521598815918, 0.007607967853546143, 0.007632895946502686, 0.00758784008026123, 0.007602176189422607, 0.007569119930267334, 0.007586368083953857, 0.00766323184967041, 0.007622335910797119, 0.00760259199142456, 0.007540736198425293, 0.007586847782135009, 0.007598368167877198, 0.007631552219390869, 0.007669760227203369, 0.007618559837341309, 0.007479680061340332, 0.0075753917694091795, 0.007585984230041504, 0.0075651841163635256, 0.007575488090515137, 0.007557119846343994, 0.007558944225311279, 0.007629024028778076, 0.007583744049072265, 0.0075773119926452635, 0.0077272958755493166, 0.007615615844726562, 0.007620800018310547, 0.007593056201934814, 0.007599967956542969, 0.007572735786437988, 0.007553631782531738, 0.007583744049072265, 0.007559167861938477, 0.0076902399063110355, 0.007590015888214111, 0.007594880104064942, 0.0075820798873901364, 0.007607135772705078, 0.0075829439163208005, 0.0075781760215759275, 0.007587552070617676, 0.0076097922325134276, 0.007579872131347656, 0.007611008167266846, 0.007632736206054687, 0.007624864101409912, 0.007600128173828125, 0.007614463806152344, 0.007624512195587158, 0.007621056079864502, 0.007673600196838379, 0.007706943988800049, 0.007643040180206299, 0.007649087905883789, 0.007665408134460449, 0.0076555838584899906, 0.007634751796722412, 0.007606112003326416, 0.007586080074310303, 0.007623936176300049, 0.007609216213226319, 0.007619775772094726, 0.00758025598526001, 0.00759830379486084, 0.007622719764709473, 0.007638944149017334, 0.007597184181213379, 0.00758787202835083, 0.007586688041687012, 0.007636991977691651, 0.007599199771881103, 0.007580351829528808, 0.007569695949554443, 0.007577631950378418, 0.007601119995117187, 0.007602975845336914, 0.007843552112579345, 0.007479519844055176, 0.007622623920440674, 0.007594367980957031, 0.0076102399826049804, 0.00760422420501709, 0.007606272220611572, 0.007579648017883301, 0.007589920043945313, 0.007564352035522461, 0.0076315197944641115, 0.007614719867706299, 0.007619904041290283, 0.007588543891906738, 0.007603775978088379, 0.007588319778442383, 0.007579616069793701, 0.007599199771881103, 0.007617184162139893, 0.007598368167877198, 0.007614175796508789, 0.0076059517860412595, 0.007621183872222901, 0.007604320049285889, 0.0076102719306945805, 0.007578752040863037, 0.00758022403717041, 0.0075937919616699216, 0.007638976097106933, 0.007581759929656982, 0.007657055854797363, 0.007630784034729004, 0.007610464096069336, 0.00760422420501709, 0.007616415977478027, 0.007602431774139404, 0.00759497594833374, 0.007581408023834228, 0.007589920043945313, 0.007571455955505371, 0.00759932804107666, 0.0075784001350402835, 0.007604544162750244, 0.007618239879608154, 0.007587232112884521, 0.00756547212600708, 0.007567808151245117, 0.007575551986694336, 0.007546879768371582, 0.007718912124633789, 0.007585087776184082, 0.007575551986694336, 0.0075820798873901364, 0.007649504184722901, 0.007603328227996826, 0.007584735870361328, 0.007606495857238769, 0.007571231842041015, 0.007585216045379639, 0.007563680171966553, 0.00777459192276001, 0.007635776042938232, 0.007619071960449219, 0.007587327957153321, 0.007489471912384033, 0.007585824012756348, 0.007569407939910889, 0.007573376178741455, 0.007553055763244629, 0.007571968078613281, 0.0075838398933410645, 0.007526400089263916, 0.007594079971313477, 0.007565536022186279, 0.007581376075744629, 0.007599487781524659, 0.007588384151458741, 0.0075879359245300294, 0.007592000007629394, 0.007559103965759277, 0.007616511821746826, 0.0075913920402526855, 0.007621024131774902, 0.0075996479988098145, 0.0076232638359069824, 0.007624703884124756, 0.007644512176513672, 0.007616384029388428, 0.007610432147979736, 0.007695231914520264, 0.007607615947723388, 0.007622623920440674, 0.00760643196105957, 0.007616479873657226, 0.007625152111053466, 0.007661568164825439, 0.007681280136108398, 0.007727039813995361, 0.007592351913452149, 0.007584159851074219, 0.007616511821746826, 0.007593728065490722, 0.007581471920013428, 0.0075792322158813475, 0.007582592010498047, 0.007589824199676513, 0.0076282558441162105, 0.007584352016448975, 0.0076431999206542965, 0.00755295991897583, 0.007605728149414063, 0.007707168102264404, 0.00757919979095459, 0.009147071838378906, 0.007647007942199707, 0.007987167835235596, 0.007602431774139404, 0.007646336078643799, 0.007579487800598144, 0.007581696033477783, 0.007561888217926025, 0.0076183037757873535, 0.00757206392288208, 0.007601920127868652, 0.007631999969482422, 0.007588768005371093, 0.007614687919616699, 0.00753868818283081, 0.007620128154754638, 0.007612448215484619, 0.007586559772491455, 0.007589183807373047, 0.007586271762847901, 0.007557024002075195, 0.007587264060974121, 0.007586368083953857, 0.007546239852905273, 0.007577824115753174, 0.007600543975830078, 0.007560927867889405, 0.007549439907073975, 0.007585375785827637, 0.007550399780273437, 0.007536960124969482, 0.0075792641639709476, 0.007564000129699707, 0.00757155179977417, 0.007595871925354004, 0.007608704090118408, 0.007570464134216309, 0.00759065580368042, 0.00759219217300415, 0.007677696228027344, 0.0076327681541442875, 0.007585919857025146, 0.007571519851684571, 0.007614399909973145, 0.007628032207489014, 0.0075823040008544925, 0.007544991970062256, 0.007578879833221436, 0.00759884786605835, 0.007616511821746826, 0.007614016056060791, 0.007581344127655029, 0.007543231964111328, 0.007610720157623291, 0.007599167823791504, 0.007603136062622071, 0.007577856063842773, 0.007606175899505616, 0.007580863952636719, 0.0075905599594116215, 0.007590144157409668, 0.007599071979522705, 0.007574304103851319, 0.007580832004547119, 0.007648096084594726, 0.007636672019958496, 0.007565824031829834, 0.007606080055236816, 0.007644864082336425, 0.0076928319931030276, 0.007583199977874756, 0.007613887786865234, 0.007559679985046387, 0.0075779838562011715, 0.007585663795471192, 0.007637407779693604, 0.007548064231872559, 0.007473311901092529, 0.007693215847015381, 0.007611328125, 0.007622303962707519, 0.007637343883514404, 0.0076269440650939945, 0.0076204161643981935, 0.007611711978912353, 0.007622591972351074, 0.00763478422164917, 0.0075970239639282225, 0.007632832050323487, 0.007605631828308105, 0.007642816066741944, 0.008529055595397949, 0.008635552406311035, 0.008239199638366699, 0.00793449592590332, 0.0076902399063110355, 0.007673408031463623, 0.007607840061187744, 0.007637375831604004, 0.007669280052185059, 0.007630144119262695, 0.0076159038543701174, 0.007739488124847412, 0.007649792194366455, 0.007628543853759766, 0.007673791885375977, 0.00762063980102539, 0.007722976207733154, 0.007681280136108398, 0.007604991912841797, 0.007636991977691651, 0.007649343967437744, 0.007653664112091065, 0.007659167766571045, 0.007667168140411377, 0.007639584064483643, 0.0076146240234375, 0.0078887357711792, 0.007849247932434082, 0.008193920135498047, 0.008536479949951171, 0.00766102409362793, 0.007647615909576416, 0.007659743785858155, 0.007664159774780273, 0.0075959677696228025, 0.007629759788513183, 0.007633887767791748, 0.007667456150054932, 0.007648640155792237, 0.007637887954711914, 0.007607456207275391, 0.007622591972351074, 0.007691167831420899, 0.007646592140197754, 0.007615007877349853, 0.007587135791778565, 0.007671999931335449, 0.007634624004364014, 0.007727776050567627, 0.007495327949523926, 0.007616864204406738, 0.007622399806976318, 0.007741695880889893, 0.0076260800361633305, 0.007627423763275147, 0.007658592224121094, 0.007645919799804688, 0.00764089584350586, 0.0076455678939819334, 0.007643136024475097, 0.00758351993560791, 0.007628191947937011, 0.007596415996551514, 0.007672255992889405, 0.007677951812744141, 0.007608320236206055, 0.007620607852935791, 0.007595039844512939, 0.007607264041900635, 0.00763702392578125, 0.007682015895843506, 0.0076308479309082035, 0.007620607852935791, 0.0076382079124450684, 0.007619391918182373, 0.007610367774963379, 0.007645376205444336, 0.007599936008453369, 0.007622496128082275, 0.007679296016693115, 0.007641952037811279, 0.007671807765960693, 0.008210432052612305, 0.007554624080657959, 0.007600575923919678, 0.007619743824005127, 0.0076704959869384765, 0.0076284480094909665, 0.007623136043548584, 0.007619647979736328, 0.007602431774139404, 0.007633567810058594, 0.007607583999633789, 0.007625120162963867, 0.007590240001678467, 0.0075831360816955564, 0.007605855941772461, 0.0076334080696105954, 0.007621439933776855, 0.007630527973175049, 0.00758080005645752, 0.007594944000244141, 0.007636928081512452, 0.007745759963989258, 0.007571231842041015, 0.007610432147979736, 0.00766761589050293, 0.007589983940124512, 0.007612351894378662, 0.00758515214920044, 0.007553023815155029, 0.007576064109802246, 0.007474592208862305, 0.007602784156799316, 0.007615935802459717, 0.007585440158843994, 0.007613344192504883, 0.007588096141815185, 0.00760975980758667, 0.007594304084777832, 0.007591008186340332, 0.007627071857452393, 0.00760259199142456, 0.00759555196762085, 0.007590879917144775, 0.007608384132385254, 0.007579296112060547, 0.007607456207275391, 0.007568448066711426, 0.007694240093231201, 0.007640960216522216, 0.007634943962097168, 0.007616000175476074, 0.007598015785217285, 0.007612991809844971, 0.007640575885772705, 0.0075924482345581055, 0.00758784008026123, 0.007616223812103272, 0.007622943878173828, 0.007618783950805664, 0.008443679809570313, 0.007688159942626953, 0.007663040161132812, 0.007654111862182617, 0.0076358399391174316, 0.007600895881652832, 0.00760038423538208, 0.0076770238876342775, 0.007625631809234619, 0.007577600002288819, 0.007599584102630615, 0.007602464199066162, 0.00759219217300415, 0.007616511821746826, 0.00758739185333252, 0.00760262393951416, 0.007589888095855713, 0.0076528000831604, 0.007633471965789795, 0.007643136024475097, 0.007665247917175293, 0.007622943878173828, 0.007696832180023193, 0.007615520000457764, 0.0076581439971923826, 0.0077621121406555174, 0.007767648220062256, 0.007884255886077881, 0.007647647857666016, 0.0077224960327148436, 0.0077218241691589355, 0.007678207874298096, 0.007644351959228516, 0.00770304012298584, 0.007469727993011475, 0.007600255966186523, 0.007542943954467774, 0.007620351791381836, 0.007583775997161865, 0.007615871906280517, 0.007891520023345948, 0.007618559837341309, 0.007584799766540527, 0.007583744049072265, 0.007602240085601807, 0.007635871887207032, 0.007575551986694336, 0.007591936111450195, 0.007575039863586426, 0.0075905599594116215, 0.00763478422164917, 0.007597472190856934, 0.007590847969055176, 0.0075641279220581056, 0.00762553596496582, 0.0075980801582336424, 0.007571072101593018, 0.007578239917755127, 0.007578495979309082, 0.007590176105499268, 0.007620319843292236, 0.007566207885742187, 0.007541823863983155, 0.007568319797515869, 0.007589888095855713, 0.007587903976440429, 0.007571392059326172, 0.007623968124389649, 0.007577439785003662, 0.00758409595489502, 0.007592480182647705, 0.007579552173614502, 0.007558720111846924, 0.007620448112487793, 0.007596960067749023, 0.007622496128082275, 0.007638720035552979, 0.007622911930084229, 0.007688191890716553, 0.007616511821746826, 0.00764518404006958, 0.007620672225952149, 0.007633920192718506, 0.007593183994293213, 0.007622367858886718, 0.007663616180419922, 0.007696320056915283, 0.007688255786895752, 0.0077266240119934085, 0.007770592212677002, 0.007718912124633789, 0.007731008052825928, 0.007819200038909912, 0.007788640022277832, 0.007784607887268066, 0.007708672046661377, 0.007641088008880615, 0.007574687957763672, 0.007674240112304688, 0.007645408153533936, 0.007598720073699951, 0.0076284799575805666, 0.007604032039642334, 0.0075903677940368655, 0.00758355188369751, 0.007600128173828125, 0.0076614079475402835, 0.007645664215087891, 0.007653247833251953, 0.007628608226776123, 0.007677728176116943, 0.007658751964569091, 0.00766806411743164, 0.007703167915344238, 0.007821311950683594, 0.00786848020553589, 0.007813055992126465, 0.007833600044250488, 0.00788431978225708, 0.007864064216613769, 0.007885663986206055, 0.007868288040161132, 0.00789238405227661, 0.0078787841796875, 0.007868256092071533, 0.007852128028869629, 0.007918111801147461, 0.007917439937591553, 0.007886559963226318, 0.007872928142547607, 0.007989247798919678, 0.007958528041839599, 0.007968768119812012, 0.007940095901489258, 0.007974783897399902, 0.007982687950134277, 0.008073760032653808, 0.008095616340637208, 0.00816579246520996, 0.008064319610595703, 0.008014240264892579, 0.008046175956726074, 0.0080730562210083, 0.008026335716247559, 0.00804207992553711, 0.008055551528930664, 0.00806009578704834, 0.008044992446899414, 0.007952767848968506, 0.008011743545532227, 0.00802620792388916, 0.008052512168884278, 0.008025664329528809, 0.008069727897644043, 0.008006752014160156, 0.007944672107696534, 0.007956799983978271, 0.00797708797454834, 0.007997439861297608, 0.007970816135406494, 0.007845503807067871, 0.007887743949890136, 0.00789308786392212, 0.008013824462890624, 0.007878655910491944, 0.007849984169006348, 0.008056832313537597, 0.007763423919677734, 0.007776800155639648, 0.007759967803955078, 0.007698048114776612, 0.0076813120841979984, 0.00772166395187378, 0.0076455678939819334, 0.007677440166473389, 0.007696576118469238, 0.007719168186187744, 0.007651328086853027, 0.007636288166046143, 0.0076399359703063965, 0.0076080961227417, 0.00762883186340332, 0.007636832237243652, 0.007628960132598877, 0.007627967834472656, 0.007630688190460205, 0.007599071979522705, 0.007605535984039307, 0.007619296073913574, 0.007613791942596436, 0.007581855773925781, 0.007602303981781006, 0.007605631828308105, 0.007617536067962646, 0.008310751914978027, 0.0077990078926086425, 0.007837503910064698, 0.007704256057739257, 0.00794655990600586, 0.008959136009216308, 0.007687007904052735, 0.007688191890716553, 0.007788544178009033, 0.007651328086853027, 0.007658720016479492, 0.007631616115570068, 0.007641280174255371, 0.007628096103668213, 0.007593696117401123, 0.007574336051940918, 0.0075684161186218265, 0.007620672225952149, 0.00764851188659668, 0.007605663776397705, 0.007587615966796875, 0.0075548157691955565, 0.007574240207672119, 0.007581503868103028, 0.007624800205230713, 0.0075706238746643065, 0.0076111359596252445, 0.007585440158843994, 0.007571968078613281, 0.007487584114074707, 0.007585919857025146, 0.007578879833221436, 0.007582464218139648, 0.007601984024047851, 0.007567647933959961, 0.007610144138336182, 0.007622432231903076, 0.0076044478416442875, 0.0075980801582336424, 0.0075980801582336424, 0.007585792064666748, 0.007618656158447265, 0.007665919780731201, 0.007593376159667969, 0.007575104236602783, 0.007602464199066162, 0.007634431838989258, 0.007746463775634766, 0.007571167945861816, 0.007639328002929688, 0.007608320236206055, 0.007591904163360596, 0.007602208137512207, 0.007600128173828125, 0.007623871803283691, 0.007643968105316162, 0.007747583866119385, 0.007733248233795166, 0.007886303901672363, 0.00902182388305664, 0.007679840087890625, 0.007671199798583985, 0.007674784183502197, 0.007624703884124756, 0.007659520149230957, 0.007636991977691651, 0.007644703865051269, 0.007622528076171875, 0.007621024131774902, 0.007630271911621094, 0.007613471984863281, 0.007572864055633545, 0.007588319778442383, 0.007731232166290283, 0.007606112003326416, 0.007612415790557861, 0.007624512195587158, 0.007602367877960205, 0.007677760124206543, 0.007605663776397705, 0.007641407966613769, 0.007594336032867432, 0.007609856128692627, 0.007592576026916504, 0.007620319843292236, 0.007780416011810303, 0.007610591888427734, 0.00791321611404419, 0.0076512317657470704, 0.007658048152923584, 0.007630623817443847, 0.007591072082519531, 0.007565855979919434, 0.00763315200805664, 0.007617568016052246, 0.007638336181640625, 0.007646880149841309, 0.007658559799194336, 0.007615583896636963, 0.007624544143676758, 0.007577600002288819, 0.007818655967712402, 0.007632480144500732, 0.0076605439186096195, 0.007639167785644531, 0.007595839977264404, 0.007650911808013916, 0.007651968002319336, 0.007609600067138672, 0.007606016159057617, 0.0076276159286499026, 0.0077760000228881835, 0.0075840001106262205, 0.007624032020568848, 0.007635615825653076, 0.007606207847595215, 0.007597599983215332, 0.0075865921974182125, 0.0075589118003845214, 0.007577600002288819, 0.007604063987731934, 0.007583456039428711, 0.007586239814758301, 0.007606272220611572, 0.007607744216918945, 0.007594048023223877, 0.007608831882476807, 0.0075980801582336424, 0.0075980801582336424, 0.007577280044555664, 0.007621183872222901, 0.007630591869354248, 0.007579296112060547, 0.007586112022399902, 0.00759987211227417, 0.007592512130737305, 0.007591648101806641, 0.007587776184082031, 0.00758512020111084, 0.00759881591796875, 0.0075797119140625, 0.007593920230865478, 0.0075935997962951664, 0.007632512092590332, 0.007623424053192138, 0.007585792064666748, 0.007616511821746826, 0.007639039993286132, 0.007620704174041748, 0.007798208236694336, 0.0076928000450134275, 0.007634560108184815, 0.007608672142028808, 0.0075980801582336424, 0.007646624088287353, 0.007491583824157715, 0.0075959677696228025, 0.007595808029174805, 0.0076249918937683105, 0.007613887786865234, 0.00756828784942627, 0.007611328125, 0.00763478422164917, 0.0076174077987670894, 0.007571455955505371, 0.0075482878684997555, 0.007604095935821533, 0.007584288120269776, 0.007549280166625976, 0.007628543853759766, 0.007602303981781006, 0.0076080641746520995, 0.007604000091552735, 0.007606751918792724, 0.007617983818054199, 0.0076754879951477055, 0.0076171522140502925, 0.007616864204406738, 0.007618559837341309, 0.0076689600944519045, 0.007642176151275635, 0.007605504035949707, 0.0075944638252258305, 0.00763804817199707, 0.007623040199279786, 0.007673727989196777, 0.007621344089508057, 0.007671807765960693, 0.0076308479309082035, 0.007617919921875, 0.007632800102233887, 0.0076111040115356444, 0.00765337610244751, 0.007594143867492676, 0.007653215885162354, 0.007616511821746826, 0.0076406078338623045, 0.007573984146118164, 0.007622655868530274, 0.007639039993286132, 0.007658912181854248, 0.0076184959411621095, 0.007602272033691407, 0.007626399993896484, 0.007594912052154541, 0.007701727867126465, 0.007608863830566406, 0.007589407920837402, 0.007580639839172364, 0.007624447822570801, 0.007579808235168457, 0.007609407901763916, 0.007594783782958984, 0.007576831817626953, 0.007609375953674316, 0.00764463996887207, 0.0076167678833007815, 0.007700032234191895]",tokens/s,130.21992940998916,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2163.679232,2192.50688,0.0,1814.03648,1724.34432,s,1,8.6876279296875,8.6876279296875,0.0,8.6876279296875,8.6876279296875,8.6876279296875,8.6876279296875,[8.6876279296875],,kWh,5.2378508583327246e-05,5.770573031547885e-06,1.6856124595995836e-05,7.500520621087097e-05,,MB,2234.568704,2337.210368,0.0,1929.37984,1887.281152,s,10,1.687504837036133,0.16875048370361329,0.0007658928780544979,0.16883905792236328,0.16966636962890624,0.16974974212646485,0.16981644012451172,"[0.1671366424560547, 0.16897462463378907, 0.16917030334472657, 0.16802627563476563, 0.16818873596191405, 0.16919760131835937, 0.1687034912109375, 0.16964784240722655, 0.16983311462402345, 0.16862620544433593]",tokens/s,1517.0326886270047,kWh,5.101365707686887e-06,5.625916986270001e-07,3.3957594215861072e-06,9.059716827899995e-06,tokens/kWh,28256953.816882126,MB,2242.064384,2484.011008,0.0,2076.18048,1946.88,s,10,21.033904541015623,2.1033904541015622,0.014815617817911382,2.100609619140625,2.11775634765625,2.1223167724609375,2.1259651123046877,"[2.096475830078125, 2.126877197265625, 2.116742919921875, 2.073525146484375, 2.09481396484375, 2.11563916015625, 2.116057861328125, 2.102658935546875, 2.09255322265625, 2.098560302734375]",tokens/s,29.95164301385483,kWh,6.14890224543972e-05,6.78207965365139e-06,3.169151864821403e-05,9.99626207562626e-05,tokens/kWh,630235.5772925561,,s,630,21.030927616119378,0.03338242478749109,0.0005420377273182946,0.033304639816284176,0.033925838470458984,0.03415963306427002,0.03517127418518068,"[0.03385971069335938, 0.033560577392578124, 0.03357491302490234, 0.033230846405029296, 0.034095104217529294, 0.03323878479003906, 0.03341337585449219, 0.033593505859375, 0.033570655822753905, 0.03295363235473633, 0.033733345031738284, 0.033013343811035156, 0.032943649291992186, 0.032912254333496094, 0.03292144012451172, 0.03291971206665039, 0.033044063568115234, 0.032995742797851564, 0.03378364944458008, 0.03377577590942383, 0.033650688171386715, 0.03292131042480469, 0.03308915328979492, 0.03289155197143555, 0.033114017486572264, 0.033036384582519535, 0.03295379257202148, 0.032776447296142576, 0.032903488159179685, 0.03292940902709961, 0.03299161529541016, 0.033124351501464845, 0.033293663024902345, 0.03300214385986328, 0.0331038703918457, 0.03307900619506836, 0.0329587516784668, 0.03311174392700195, 0.033005889892578126, 0.03298099136352539, 0.033090847015380856, 0.03294486236572266, 0.03301580810546875, 0.03337830352783203, 0.03301715087890625, 0.03285881423950195, 0.032829536437988284, 0.033850528717041015, 0.03352652740478516, 0.0332344970703125, 0.03326556777954102, 0.032995872497558594, 0.03349903869628906, 0.03322889709472656, 0.033060256958007815, 0.03326620864868164, 0.03399481582641602, 0.03362403106689453, 0.03370374298095703, 0.033810272216796874, 0.03382720184326172, 0.034293022155761715, 0.03377020645141601, 0.03381862258911133, 0.03363577651977539, 0.03324371337890625, 0.032933887481689454, 0.032774303436279295, 0.03368921661376953, 0.03720828628540039, 0.03580089569091797, 0.03385567855834961, 0.03385139083862305, 0.03459036636352539, 0.03438595199584961, 0.03406867218017578, 0.03383219146728516, 0.0340937614440918, 0.033742111206054685, 0.03396006393432617, 0.035255073547363285, 0.033924606323242186, 0.03343001556396484, 0.03358310317993164, 0.033274974822998044, 0.03343049621582031, 0.0337509765625, 0.03369574356079102, 0.03310953521728516, 0.033247711181640625, 0.03285606384277344, 0.03278643035888672, 0.03297644805908203, 0.03341094589233398, 0.03296723175048828, 0.03373875045776367, 0.03348889541625977, 0.03366214370727539, 0.033770305633544925, 0.03431398391723633, 0.03384345626831055, 0.03416064071655273, 0.033808383941650394, 0.03448384094238281, 0.033982463836669925, 0.03373708724975586, 0.03339491271972656, 0.03378972625732422, 0.03367279815673828, 0.033614238739013674, 0.03377971267700195, 0.03365599822998047, 0.03329926300048828, 0.03371615982055664, 0.033453376770019534, 0.03355519866943359, 0.03371177673339844, 0.034154624938964845, 0.034449630737304685, 0.03386777496337891, 0.03343155288696289, 0.03338444900512695, 0.03341862487792969, 0.03345996856689453, 0.03323936080932617, 0.03337273788452148, 0.03368521499633789, 0.03354431915283203, 0.03430825424194336, 0.03375059127807617, 0.03346886444091797, 0.033474529266357425, 0.03363174438476563, 0.03320876693725586, 0.033451904296875, 0.033191776275634764, 0.033188224792480465, 0.0332861442565918, 0.0332421760559082, 0.03330495834350586, 0.033464897155761716, 0.033388511657714844, 0.03366060638427734, 0.03333769607543945, 0.033521503448486326, 0.033161376953125, 0.03338854217529297, 0.033639904022216796, 0.03352617645263672, 0.03398665618896484, 0.033947681427001955, 0.03381452941894531, 0.03406972885131836, 0.034613407135009766, 0.034587265014648434, 0.033834720611572264, 0.033706272125244144, 0.03392451095581055, 0.03428208160400391, 0.03386777496337891, 0.033625278472900394, 0.03346688079833984, 0.033492767333984375, 0.033304222106933595, 0.033532798767089846, 0.03349462509155274, 0.03746815872192383, 0.0339439697265625, 0.03410124969482422, 0.03377356719970703, 0.0337606086730957, 0.033868446350097656, 0.033705215454101566, 0.0337599983215332, 0.03341516876220703, 0.03320732879638672, 0.033557472229003904, 0.033500831604003904, 0.033206623077392576, 0.033150558471679685, 0.03313910293579102, 0.03287161636352539, 0.032889503479003906, 0.03310198211669922, 0.03272496032714844, 0.03287817764282226, 0.03278006362915039, 0.03288336181640625, 0.03340220642089844, 0.03318198394775391, 0.03299881744384765, 0.03312428665161133, 0.03287926483154297, 0.03282124710083008, 0.03284172821044922, 0.032858081817626957, 0.03300102233886719, 0.03290358352661133, 0.03283062362670899, 0.03333987045288086, 0.0328974723815918, 0.03282124710083008, 0.032655361175537106, 0.03282473754882813, 0.03273936080932617, 0.03260883331298828, 0.03278438568115234, 0.03302387237548828, 0.03278451156616211, 0.032753662109375, 0.032732673645019535, 0.032700927734375, 0.03294003295898437, 0.03287788772583008, 0.03329683303833008, 0.03305472183227539, 0.032868606567382816, 0.03276201629638672, 0.03290476989746094, 0.032866241455078125, 0.03268803024291992, 0.03280704116821289, 0.03270892715454102, 0.03266479873657226, 0.032897823333740236, 0.032860160827636715, 0.03278598403930664, 0.033753536224365235, 0.033355777740478515, 0.032812801361083985, 0.032798816680908206, 0.03283779144287109, 0.0325766716003418, 0.03320304107666015, 0.03269222259521484, 0.03264716720581055, 0.032830463409423825, 0.03281808090209961, 0.03263702392578125, 0.03293366241455078, 0.03273926544189453, 0.033585281372070314, 0.032747264862060546, 0.03314729690551758, 0.03314387130737305, 0.03361264038085938, 0.03289712142944336, 0.0328623046875, 0.03276335906982422, 0.03292963027954102, 0.03275414276123047, 0.03306918334960938, 0.034050239562988284, 0.03365017700195312, 0.033632606506347654, 0.03326755142211914, 0.03302431869506836, 0.03311001586914063, 0.03337625503540039, 0.03315273666381836, 0.03313897705078125, 0.03321446228027344, 0.03322985458374023, 0.03344688034057617, 0.03330841445922852, 0.03292505645751953, 0.03315801620483398, 0.03324313735961914, 0.033576961517333984, 0.03328409576416016, 0.03374694442749023, 0.033081344604492184, 0.033073150634765625, 0.03316940689086914, 0.03321651077270508, 0.03348889541625977, 0.03354214477539062, 0.0332677116394043, 0.033638111114501955, 0.033562145233154296, 0.03360015869140625, 0.033635520935058595, 0.03366118240356445, 0.03379008102416992, 0.03347715377807617, 0.03341107177734375, 0.033288192749023435, 0.03308544158935547, 0.03387587356567383, 0.03290940856933594, 0.03309721755981445, 0.03306752014160156, 0.03285811233520508, 0.03290521621704102, 0.033089183807373045, 0.0333043212890625, 0.033454689025878906, 0.03304025650024414, 0.0329686393737793, 0.032913600921630856, 0.03331071853637695, 0.033056766510009765, 0.03309568023681641, 0.03284313583374023, 0.032891521453857424, 0.03292979049682617, 0.032968894958496094, 0.03317695999145508, 0.03322515106201172, 0.03304652786254883, 0.03292505645751953, 0.03304307174682617, 0.03320012664794922, 0.03283967971801758, 0.03299916839599609, 0.033716350555419924, 0.03355420684814453, 0.03374028778076172, 0.03377020645141601, 0.033377281188964845, 0.03347967910766601, 0.03354828643798828, 0.033406368255615236, 0.033385055541992184, 0.033058815002441407, 0.03316873550415039, 0.03397216033935547, 0.033495777130126955, 0.033057022094726565, 0.03311999893188477, 0.032890689849853515, 0.032823486328125, 0.03287449645996094, 0.03291446304321289, 0.03304735946655273, 0.03329782485961914, 0.03356953430175781, 0.03359539031982422, 0.034549758911132815, 0.03382476806640625, 0.033462272644042966, 0.03362972640991211, 0.033513824462890626, 0.03347049713134766, 0.03353174209594727, 0.0334637451171875, 0.033057342529296874, 0.03319206237792969, 0.033285919189453124, 0.03392351913452148, 0.037513118743896484, 0.03350262451171875, 0.03330928039550781, 0.03320012664794922, 0.033058815002441407, 0.03306086349487305, 0.03359699249267578, 0.03366249465942383, 0.03359017562866211, 0.03569868850708008, 0.03471769714355469, 0.03386368179321289, 0.03354214477539062, 0.0336629753112793, 0.03361705780029297, 0.03377558517456054, 0.03364537429809571, 0.034116992950439455, 0.03341139221191406, 0.0332353286743164, 0.03340268707275391, 0.03365292739868164, 0.033685504913330076, 0.033438720703125, 0.033291263580322264, 0.03330355072021484, 0.03333017730712891, 0.0336732177734375, 0.03399318313598633, 0.03402873611450195, 0.034085025787353514, 0.03452399826049805, 0.034170047760009765, 0.03451372909545898, 0.034245662689208985, 0.03372540664672852, 0.03393692779541015, 0.03344966506958008, 0.03336272048950195, 0.033664798736572264, 0.033495201110839846, 0.032940097808837894, 0.03482777786254883, 0.03350783920288086, 0.03280806350708008, 0.03309862518310547, 0.03299049758911133, 0.032809791564941404, 0.03305011367797851, 0.032855518341064455, 0.032779201507568356, 0.03310960006713867, 0.03358143997192383, 0.03342448043823242, 0.033923553466796874, 0.03364255905151367, 0.034142623901367186, 0.03404390335083008, 0.033883201599121095, 0.033684417724609374, 0.0337367057800293, 0.034162017822265626, 0.033317569732666016, 0.03331273651123047, 0.03359743881225586, 0.03336761474609375, 0.033398880004882815, 0.03318204879760742, 0.03297484970092773, 0.033266944885253905, 0.03320908737182617, 0.03306646347045898, 0.03337648010253906, 0.03314448165893555, 0.03336259078979492, 0.03365644836425781, 0.03355481719970703, 0.033372161865234375, 0.03390614318847656, 0.033617790222167966, 0.033581729888916015, 0.03376451110839844, 0.03436835098266602, 0.035823776245117185, 0.03371769714355469, 0.033347103118896486, 0.03326569747924805, 0.03351023864746094, 0.033367649078369144, 0.033055137634277344, 0.03310182571411133, 0.03496611022949219, 0.03398451232910156, 0.03394355010986328, 0.034095104217529294, 0.03377577590942383, 0.033883998870849606, 0.03415840148925781, 0.03369184112548828, 0.03364799880981445, 0.033479297637939456, 0.0332578239440918, 0.03310147094726563, 0.03318374252319336, 0.03325337600708008, 0.03328409576416016, 0.033067008972167966, 0.0328007698059082, 0.0329051513671875, 0.0329420166015625, 0.03274665451049805, 0.03314527893066406, 0.0331063346862793, 0.033242271423339846, 0.03375151824951172, 0.03375155258178711, 0.03426537704467773, 0.03389616012573242, 0.033799232482910155, 0.033616703033447264, 0.03368889617919922, 0.034030399322509765, 0.03352883148193359, 0.03343215942382813, 0.033283809661865234, 0.033149631500244144, 0.03405974578857422, 0.03387241744995117, 0.0333496322631836, 0.033105247497558596, 0.03296527862548828, 0.03279199981689453, 0.03287823867797852, 0.03285903930664062, 0.03285308837890625, 0.03309241485595703, 0.033447776794433594, 0.03357708740234375, 0.03310172653198242, 0.032917728424072264, 0.03273689651489258, 0.032944190979003904, 0.033113792419433595, 0.03277423858642578, 0.03305292892456055, 0.03291164779663086, 0.03301580810546875, 0.03326566314697266, 0.033081344604492184, 0.03301375961303711, 0.03326678466796875, 0.03349382400512695, 0.033380447387695314, 0.033562625885009766, 0.03363616180419922, 0.03343993759155273, 0.03353571319580078, 0.03390288162231445, 0.03304560089111328, 0.03294739151000976, 0.03288444900512695, 0.032737281799316405, 0.03295830535888672, 0.03291561508178711, 0.03273932647705078, 0.034062335968017575, 0.03370729446411133, 0.033757919311523436, 0.03357491302490234, 0.03354009628295898, 0.033352832794189456, 0.03361667251586914, 0.03428691101074219, 0.03369859313964844, 0.03370598220825195, 0.03339820861816406, 0.033016319274902346, 0.03312236785888672, 0.033337215423583984, 0.033969280242919925, 0.033299102783203124, 0.03359737777709961, 0.03309609603881836, 0.03316326522827148, 0.03309113693237305, 0.03281894302368164, 0.03297548675537109, 0.03333740615844726, 0.03352576065063476, 0.03347171020507812, 0.033484928131103514, 0.033232864379882814, 0.03335443115234375, 0.03313638305664063, 0.03288691329956055, 0.033177310943603516, 0.03299343872070312, 0.03284137725830078, 0.032860767364501955, 0.03266559982299805, 0.032702465057373044, 0.032859233856201174, 0.033364414215087894, 0.032920032501220706, 0.03307724761962891, 0.032860160827636715, 0.03280883026123047, 0.033236576080322267, 0.033003456115722654, 0.03274163055419922, 0.03289116668701172, 0.032873950958251956, 0.03356118392944336, 0.03304857635498047, 0.03289907073974609, 0.03273318481445313, 0.03281305694580078, 0.03365718460083008, 0.03350153732299805, 0.03354828643798828, 0.033462272644042966, 0.03361753463745117, 0.03348691177368164, 0.03340934371948242, 0.03301324844360352, 0.03316787338256836, 0.03316940689086914, 0.03322995376586914, 0.03366182327270508, 0.033355777740478515, 0.03299926376342773, 0.03310812759399414, 0.03298918533325195, 0.0330015983581543, 0.03304947280883789, 0.03323801422119141, 0.03365683364868164, 0.03399270248413086, 0.03379609680175781, 0.03347433471679687, 0.03352348709106445, 0.033752670288085936, 0.03345632171630859, 0.03367388916015625, 0.033726463317871096, 0.033465343475341795, 0.03341209411621094, 0.03336355209350586, 0.03355430221557617, 0.03329692840576172, 0.0344535026550293, 0.03321392059326172, 0.033102367401123045, 0.03326976013183594, 0.03285606384277344, 0.03290521621704102, 0.033314208984375, 0.03288275146484375, 0.03388022232055664, 0.033936958312988284, 0.03324169540405274, 0.033056991577148434, 0.03277983856201172, 0.03275411224365234, 0.03304243087768555, 0.03285948944091797, 0.032627201080322264, 0.033406238555908206, 0.03302652740478516, 0.0327982063293457, 0.032934814453125, 0.032903167724609376, 0.03273523330688476, 0.03288444900512695, 0.0328768310546875, 0.03390259170532227, 0.033056575775146486, 0.03292793655395508, 0.03297206497192383, 0.03476095962524414]",tokens/s,29.955882664782205,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4359.479296,4562.28864,0.0,4183.81824,4182.069248,s,1,10.2638212890625,10.2638212890625,0.0,10.2638212890625,10.2638212890625,10.2638212890625,10.2638212890625,[10.2638212890625],,kWh,9.306532716250094e-05,1.0258485197407685e-05,2.925252340199992e-05,0.00013257633576190854,,MB,1650.229248,4715.380736,0.0,4307.550208,4281.174016,s,10,4.167554779052734,0.41675547790527345,0.006229591291880606,0.41788583374023436,0.42322588806152345,0.42383643951416017,0.42432488067626956,"[0.40105911254882814, 0.4190422668457031, 0.4153018798828125, 0.41954266357421877, 0.42015341186523436, 0.4149872131347656, 0.4167294006347656, 0.4244469909667969, 0.41320162963867185, 0.4230902099609375]",tokens/s,614.2690704072463,kWh,1.2115698900666626e-05,1.3361515482894504e-06,8.055739777919988e-06,2.1507590226876064e-05,tokens/kWh,11902774.662318991,MB,1658.617856,4730.0608,0.0,4322.230272,4281.176576,s,10,25.517523681640625,2.551752368164063,0.01786764099154408,2.5614346923828126,2.569007763671875,2.569834814453125,2.5704964550781253,"[2.522132080078125, 2.530519775390625, 2.5293154296875, 2.564641357421875, 2.541203125, 2.568823974609375, 2.567356689453125, 2.570661865234375, 2.560109130859375, 2.56276025390625]",tokens/s,24.68891605078721,kWh,7.458373593766659e-05,8.226687555871763e-06,4.913483375228005e-05,0.00013194525724581844,tokens/kWh,477470.7429053618,,s,630,25.51400104522709,0.04049841435750325,0.0006892768181309524,0.04049379348754883,0.04097646636962891,0.04123205833435059,0.043241351547241215,"[0.042024959564208986, 0.04011318588256836, 0.04326089477539063, 0.03966345596313477, 0.039863967895507814, 0.0402474250793457, 0.04042176055908203, 0.04027190399169922, 0.040433631896972654, 0.040605342864990235, 0.040393054962158205, 0.04018175888061523, 0.04040703964233398, 0.03999948883056641, 0.040253440856933595, 0.040062110900878904, 0.039905502319335935, 0.04002195358276367, 0.03972166442871094, 0.04012236785888672, 0.04004044723510742, 0.03980054473876953, 0.040242881774902345, 0.04012502288818359, 0.03992575836181641, 0.0397242546081543, 0.03997113418579101, 0.039577438354492185, 0.03962944030761719, 0.03973952102661133, 0.03973516845703125, 0.0397946891784668, 0.03964518356323242, 0.03965068817138672, 0.03959228897094726, 0.039540321350097656, 0.03948409652709961, 0.03997459030151367, 0.03957587051391601, 0.03965651321411133, 0.03950073623657226, 0.039591934204101564, 0.039573184967041014, 0.03990572738647461, 0.03992563247680664, 0.03967967987060547, 0.039696704864501955, 0.03959958267211914, 0.04003894424438476, 0.040003360748291014, 0.040433887481689454, 0.039711872100830076, 0.0398078727722168, 0.03953184127807617, 0.04002835083007812, 0.03998307037353516, 0.040083999633789065, 0.04035174560546875, 0.04012214279174805, 0.04027391815185547, 0.04023654556274414, 0.040282848358154294, 0.04009296035766602, 0.040704254150390626, 0.040509151458740233, 0.0403482551574707, 0.04049846267700195, 0.03994214248657227, 0.040106689453125, 0.03988275146484375, 0.040298526763916015, 0.04028006362915039, 0.04077772903442383, 0.04047052764892578, 0.04070707321166992, 0.04032553482055664, 0.04060425567626953, 0.040548351287841795, 0.04070352172851562, 0.04056111907958984, 0.04061199951171875, 0.04062006378173828, 0.041121505737304685, 0.04030620956420899, 0.040804000854492185, 0.0402276496887207, 0.04024124908447266, 0.04010332870483398, 0.040393310546875, 0.0407459831237793, 0.04041011047363281, 0.040449344635009765, 0.040462753295898435, 0.0397540168762207, 0.040031742095947266, 0.03975628662109375, 0.04025753784179688, 0.03969843292236328, 0.04002816009521484, 0.04012236785888672, 0.03965951919555664, 0.03975715255737305, 0.03980380630493164, 0.03998080062866211, 0.04007689666748047, 0.03982963180541992, 0.039870750427246096, 0.03970048141479492, 0.03947468948364258, 0.04010847854614258, 0.040107486724853515, 0.04016803359985351, 0.04023295974731445, 0.04002121734619141, 0.040126976013183595, 0.04000390243530273, 0.04004246520996094, 0.0396693115234375, 0.039635391235351564, 0.03999276733398437, 0.04019257736206055, 0.0398721923828125, 0.0401447982788086, 0.039532001495361326, 0.03940879821777344, 0.039409439086914064, 0.04073046493530273, 0.03975702285766602, 0.039475326538085935, 0.04128441619873047, 0.0404637451171875, 0.03932223892211914, 0.03946700668334961, 0.039202495574951174, 0.039591392517089846, 0.039813983917236326, 0.04002921676635742, 0.04005542373657227, 0.03995846557617187, 0.04023484802246094, 0.040005760192871095, 0.04001327896118164, 0.039917888641357424, 0.039834270477294924, 0.03988595199584961, 0.039785343170166014, 0.0404400634765625, 0.041285377502441406, 0.040118270874023435, 0.039808158874511716, 0.039919742584228514, 0.03991328048706055, 0.04567337417602539, 0.040304641723632816, 0.039800670623779295, 0.03969244766235352, 0.03932160186767578, 0.03912908935546875, 0.039900672912597655, 0.04657526397705078, 0.039264865875244144, 0.0393177604675293, 0.04019200134277344, 0.03984588623046875, 0.0400096321105957, 0.03952239990234375, 0.039785823822021484, 0.04115635299682617, 0.03956828689575195, 0.039516353607177736, 0.03970790481567383, 0.03966172790527344, 0.039492000579833986, 0.039522014617919925, 0.0395041618347168, 0.03970265579223633, 0.039933792114257814, 0.040108062744140624, 0.040211967468261715, 0.040968097686767575, 0.04028886413574219, 0.04033126449584961, 0.04007753753662109, 0.04011110305786133, 0.04055324935913086, 0.04012441635131836, 0.040081409454345705, 0.04003424072265625, 0.03970054244995117, 0.04093801498413086, 0.04028460693359375, 0.04024300765991211, 0.04011644744873047, 0.04018694305419922, 0.040270751953125, 0.03990323257446289, 0.03978230285644531, 0.039792736053466796, 0.04064051055908203, 0.04005887985229492, 0.039948287963867186, 0.0407347183227539, 0.04444160079956055, 0.04019200134277344, 0.04010598373413086, 0.0405362548828125, 0.04045395278930664, 0.04122544097900391, 0.03994499206542969, 0.040182785034179686, 0.04040192031860351, 0.04061187362670898, 0.04299929428100586, 0.04352867126464844, 0.04187660980224609, 0.0406146240234375, 0.04064636611938476, 0.040658432006835936, 0.0412108154296875, 0.041166847229003906, 0.041850879669189454, 0.040337120056152344, 0.040735008239746094, 0.040461345672607424, 0.040503326416015624, 0.04103238296508789, 0.040874305725097655, 0.04072959899902344, 0.041140960693359374, 0.040607391357421876, 0.04076201629638672, 0.04060355377197265, 0.040953857421875, 0.04088134384155274, 0.04049798583984375, 0.040736766815185545, 0.04036422348022461, 0.04033107376098633, 0.04022502517700195, 0.040253150939941404, 0.04043328094482422, 0.040446369171142575, 0.04045004653930664, 0.040591327667236325, 0.04041350555419922, 0.04055420684814453, 0.0404029426574707, 0.040556640625, 0.04063606262207031, 0.04078208160400391, 0.04067737579345703, 0.04074905776977539, 0.040993152618408205, 0.040267105102539065, 0.040677089691162106, 0.040541118621826175, 0.04066304016113281, 0.04079001617431641, 0.04042956924438477, 0.040130561828613284, 0.040861473083496094, 0.04024956893920899, 0.040321025848388675, 0.041261409759521486, 0.040438526153564455, 0.04029872131347656, 0.04083097457885742, 0.04050348663330078, 0.040588993072509766, 0.04054713439941406, 0.04048896026611328, 0.040529918670654294, 0.04034764862060547, 0.04040230560302734, 0.03992214584350586, 0.04055177688598633, 0.04027475357055664, 0.04007302474975586, 0.04020659255981445, 0.0401673583984375, 0.04013187026977539, 0.04001603317260742, 0.03988323211669922, 0.04004079818725586, 0.03968547058105469, 0.03974185562133789, 0.03946700668334961, 0.04005062484741211, 0.039884864807128904, 0.04023859024047852, 0.04039487838745117, 0.040110462188720705, 0.04026777648925781, 0.039847934722900394, 0.039826816558837894, 0.04026841735839844, 0.03987865447998047, 0.040097793579101565, 0.03988431930541992, 0.040243679046630856, 0.04003830337524414, 0.040115936279296875, 0.0399284782409668, 0.04013347244262695, 0.04031372833251953, 0.040922496795654295, 0.041091552734375, 0.04050307083129883, 0.040441505432128905, 0.040471038818359374, 0.04086806488037109, 0.04055449676513672, 0.04061980819702148, 0.04087968063354492, 0.040618656158447265, 0.04086083221435547, 0.040065696716308594, 0.03993414306640625, 0.04018096160888672, 0.04012521743774414, 0.04038444900512695, 0.04040620803833008, 0.040341793060302736, 0.045468257904052733, 0.04045209503173828, 0.040476673126220705, 0.04062822341918945, 0.04071014404296875, 0.04084326553344726, 0.04077772903442383, 0.04048073577880859, 0.0410747184753418, 0.04045318222045898, 0.040465087890625, 0.040495361328125, 0.04018380737304687, 0.04093648147583008, 0.0404694709777832, 0.040755199432373046, 0.04066099166870117, 0.04092927932739258, 0.04071219253540039, 0.04099071884155273, 0.04063846588134765, 0.04103366470336914, 0.041692607879638674, 0.04270758438110352, 0.040582782745361326, 0.04059324645996094, 0.04081699371337891, 0.04092252731323242, 0.0409505615234375, 0.040755199432373046, 0.040613887786865234, 0.04044521713256836, 0.040557281494140625, 0.040632320404052735, 0.04039680099487305, 0.04026367950439453, 0.04042342376708984, 0.04038057708740234, 0.04048374557495117, 0.04132863998413086, 0.04171846389770508, 0.04063257598876953, 0.040288257598876956, 0.04029974365234375, 0.04049148941040039, 0.040876415252685545, 0.04066841506958008, 0.0405203857421875, 0.04052195358276367, 0.04049283218383789, 0.040675296783447265, 0.04049513626098633, 0.04307353591918945, 0.04070604705810547, 0.04050534439086914, 0.041286209106445315, 0.04079430389404297, 0.04042956924438477, 0.040564289093017576, 0.040454593658447266, 0.04069481658935547, 0.04123747253417969, 0.04067327880859375, 0.04066227340698242, 0.040661758422851565, 0.04071219253540039, 0.040531967163085936, 0.040771583557128906, 0.04076339340209961, 0.04057452774047852, 0.04083324813842774, 0.04036608123779297, 0.040761566162109374, 0.04085145568847656, 0.04090265655517578, 0.04053606414794922, 0.040716289520263675, 0.04140351867675781, 0.040711040496826174, 0.04049452972412109, 0.040878654479980465, 0.04050515365600586, 0.04068694305419922, 0.04121481704711914, 0.04067907333374023, 0.04095564651489258, 0.04065955352783203, 0.04048191833496094, 0.04035625457763672, 0.040112865447998046, 0.04045529556274414, 0.04029481506347656, 0.0401901741027832, 0.040120319366455076, 0.042261695861816405, 0.04118815994262695, 0.040703998565673825, 0.04083011245727539, 0.040866207122802735, 0.04109151840209961, 0.040927230834960936, 0.04061363220214844, 0.04043955230712891, 0.040600223541259764, 0.04149059295654297, 0.04192540740966797, 0.040659358978271484, 0.04067103958129883, 0.0404568977355957, 0.04040723037719727, 0.041459518432617186, 0.040417278289794925, 0.04211036682128906, 0.04025324630737305, 0.040453983306884767, 0.040584129333496095, 0.040279998779296874, 0.04032928085327148, 0.041011199951171876, 0.04057292938232422, 0.04053401565551758, 0.04046233749389649, 0.040513439178466795, 0.04066624069213867, 0.04059849548339844, 0.04085139083862305, 0.040888031005859374, 0.04066361618041992, 0.0406976318359375, 0.04059340667724609, 0.040785152435302736, 0.04108774566650391, 0.040959999084472655, 0.040720382690429685, 0.04087398529052735, 0.04064166259765625, 0.04088102340698242, 0.04099689483642578, 0.04111289596557617, 0.04089267349243164, 0.0409194564819336, 0.04091289520263672, 0.04089414215087891, 0.04086201477050781, 0.041003009796142575, 0.04095974349975586, 0.0408328971862793, 0.04053833770751953, 0.04146777725219727, 0.041299968719482424, 0.040859935760498046, 0.040642177581787106, 0.04096160125732422, 0.04098239898681641, 0.04087612915039063, 0.040901214599609374, 0.04087628936767578, 0.04067440032958984, 0.04102159881591797, 0.04100787353515625, 0.040682880401611325, 0.0411440315246582, 0.0408353271484375, 0.04090275192260742, 0.04055507278442383, 0.040607135772705076, 0.040664833068847654, 0.04050825500488281, 0.040757438659667966, 0.040637889862060544, 0.040431297302246094, 0.040766239166259766, 0.04056054306030273, 0.040521728515625, 0.040687168121337894, 0.040656768798828125, 0.04076192092895508, 0.040777057647705076, 0.04083164978027344, 0.04072857666015625, 0.04076275253295898, 0.0411800651550293, 0.04095180892944336, 0.041164798736572264, 0.040827999114990236, 0.04074383926391602, 0.04084870529174805, 0.040936126708984374, 0.0413076171875, 0.04063286590576172, 0.04082688140869141, 0.04075843048095703, 0.0408009262084961, 0.04059126281738281, 0.040449951171875, 0.040524158477783206, 0.04086579132080078, 0.04072412872314453, 0.04056860733032226, 0.040434238433837894, 0.040584224700927735, 0.04021343994140625, 0.04022012710571289, 0.04013520050048828, 0.04042345428466797, 0.04028985595703125, 0.04036447906494141, 0.041060352325439455, 0.04061798477172852, 0.040787967681884765, 0.04058854293823242, 0.040395519256591794, 0.04040438461303711, 0.04085411071777344, 0.04084659194946289, 0.040604415893554686, 0.040648704528808595, 0.040959648132324215, 0.04114182281494141, 0.040632705688476566, 0.04064457702636719, 0.040542655944824216, 0.04064371109008789, 0.04067007827758789, 0.04115184020996094, 0.04079478454589844, 0.04066838455200195, 0.040527713775634765, 0.04026179122924805, 0.040007774353027346, 0.04040547180175781, 0.040089824676513675, 0.04039475250244141, 0.040493057250976565, 0.04032441711425781, 0.04032166290283203, 0.0404337272644043, 0.041748191833496096, 0.04050582504272461, 0.04025904083251953, 0.04022716903686523, 0.040578784942626955, 0.04053942489624023, 0.040575424194335935, 0.04448969650268555, 0.040820735931396485, 0.040941505432128905, 0.040975807189941406, 0.040683231353759765, 0.040810657501220704, 0.0406638069152832, 0.040771808624267575, 0.04049795150756836, 0.04012134552001953, 0.040037887573242184, 0.04006067276000977, 0.040174335479736326, 0.04319350433349609, 0.04014166259765625, 0.040654399871826175, 0.04043001556396485, 0.04048896026611328, 0.04030668640136719, 0.040785472869873045, 0.039983329772949217, 0.04008940887451172, 0.040527584075927735, 0.04015897750854492, 0.040331199645996095, 0.04055955123901367, 0.04012777709960937, 0.040497791290283205, 0.04046454238891602, 0.0405948486328125, 0.04027862548828125, 0.04098867034912109, 0.040966304779052734, 0.040610782623291014, 0.0405533447265625, 0.040818496704101564, 0.0406447982788086, 0.04087603378295898, 0.040656192779541016, 0.04070060729980469, 0.04071353530883789, 0.04074367904663086, 0.040642143249511715, 0.04065734481811523, 0.04063449478149414, 0.04049059295654297, 0.04028435134887695, 0.04091494369506836, 0.04066918563842774, 0.04058236694335938, 0.040618751525878904, 0.040654880523681644, 0.04057088088989258, 0.040591358184814456, 0.04057907104492187, 0.040509441375732425, 0.04055859375, 0.040499198913574216, 0.04066918563842774, 0.040648414611816404, 0.0406715202331543, 0.04090060806274414, 0.04107468795776367]",tokens/s,24.69232477035801,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1479.483392,1324.285952,0.0,945.815552,943.480832,s,1,8.6265927734375,8.6265927734375,0.0,8.6265927734375,8.6265927734375,8.6265927734375,8.6265927734375,[8.6265927734375],,kWh,3.7683497074990176e-05,4.149340682252034e-06,1.2175009740003628e-05,5.4007847497245837e-05,,MB,1564.012544,1487.863808,0.0,1080.03328,1046.51776,s,10,0.8680324630737304,0.08680324630737304,0.0006278083832273644,0.08660831832885743,0.08742984313964844,0.0878890167236328,0.08825635559082032,"[0.08834819030761719, 0.08587276458740234, 0.08732780456542968, 0.08653225708007813, 0.08689542388916016, 0.08658422088623047, 0.08647846221923829, 0.08643663787841797, 0.08663241577148438, 0.08692428588867188]",tokens/s,2949.1984561671334,kWh,2.7090827212576394e-06,2.987616894364072e-07,1.794060077222214e-06,4.801904487916261e-06,tokens/kWh,53312180.749161184,MB,1567.834112,1573.84704,0.0,1166.016512,1082.822656,s,10,15.914975585937501,1.5914975585937499,0.01135133220021555,1.58663525390625,1.6098780395507812,1.611726983642578,1.6132061389160157,"[1.5799921875, 1.613575927734375, 1.58172900390625, 1.5866124267578126, 1.59975390625, 1.6094671630859374, 1.5866580810546875, 1.582243408203125, 1.5903753662109374, 1.584568115234375]",tokens/s,39.58535761479076,kWh,4.6502290346242e-05,5.128885789732272e-06,2.1754403205977757e-05,7.338557934195202e-05,tokens/kWh,858479.2893225149,,s,630,15.912596687316864,0.025258089979868087,0.00046688073172256815,0.02511844730377197,0.025812412071228027,0.02593710994720459,0.02718702049255372,"[0.025034719467163086, 0.024907808303833007, 0.024838144302368165, 0.025008127212524413, 0.02488035202026367, 0.0248306884765625, 0.02478291130065918, 0.025324607849121095, 0.025113536834716798, 0.024991743087768553, 0.02496905517578125, 0.025039007186889648, 0.025044767379760743, 0.025059551239013673, 0.025648128509521483, 0.025131776809692384, 0.02503638458251953, 0.024948543548583984, 0.02489036750793457, 0.02489139175415039, 0.025034400939941408, 0.02510812759399414, 0.025182367324829102, 0.025321855545043945, 0.025200639724731445, 0.025067359924316406, 0.024954944610595702, 0.025072895050048827, 0.02503971290588379, 0.025040895462036132, 0.024928255081176756, 0.024930240631103516, 0.025098304748535156, 0.024954879760742187, 0.024938495635986328, 0.02514364814758301, 0.025018016815185548, 0.02494259262084961, 0.024954879760742187, 0.024958240509033204, 0.024918752670288084, 0.025057279586791992, 0.024899423599243162, 0.025041023254394532, 0.0249866886138916, 0.025082048416137696, 0.024988319396972655, 0.02512499237060547, 0.025110080718994142, 0.027017120361328126, 0.02555926322937012, 0.02500377655029297, 0.024928352355957032, 0.02496928024291992, 0.025163679122924804, 0.025143808364868164, 0.025118719100952147, 0.025092096328735353, 0.024989280700683594, 0.025563552856445314, 0.02497331237792969, 0.024944799423217773, 0.024831071853637695, 0.02504515266418457, 0.02492985534667969, 0.02506604766845703, 0.02511359977722168, 0.024988447189331055, 0.025075807571411132, 0.024991424560546874, 0.02512928009033203, 0.025149440765380858, 0.025065599441528322, 0.025064863204956055, 0.025038368225097658, 0.02518111991882324, 0.025296672821044922, 0.025329023361206054, 0.025458560943603516, 0.02570953559875488, 0.025644319534301758, 0.02564579200744629, 0.025659391403198242, 0.025868064880371095, 0.025870559692382812, 0.02560646438598633, 0.025517087936401367, 0.025543327331542968, 0.02554265594482422, 0.025857791900634766, 0.02614076805114746, 0.025762176513671874, 0.025982240676879882, 0.026845312118530272, 0.025885055541992188, 0.025980928421020507, 0.025977855682373048, 0.026009952545166016, 0.02598979187011719, 0.02595840072631836, 0.026200063705444337, 0.025925792694091798, 0.025945375442504883, 0.025894720077514647, 0.025886463165283202, 0.02568294334411621, 0.025758016586303712, 0.025810176849365235, 0.02565190315246582, 0.025452287673950195, 0.0257126407623291, 0.029024255752563476, 0.025606143951416017, 0.025350112915039063, 0.02531059265136719, 0.025352863311767577, 0.025362432479858397, 0.025382495880126952, 0.025399711608886717, 0.02550579261779785, 0.02539241600036621, 0.02544476890563965, 0.025430303573608398, 0.025444095611572265, 0.025325855255126952, 0.025179647445678712, 0.02539084815979004, 0.025233087539672853, 0.02516169548034668, 0.025106752395629883, 0.02521072006225586, 0.025090240478515626, 0.024993791580200195, 0.024963071823120117, 0.024869056701660158, 0.024831520080566407, 0.0249715518951416, 0.02495065689086914, 0.025054336547851563, 0.025242528915405273, 0.025345535278320314, 0.024920223236083984, 0.025040895462036132, 0.024726015090942383, 0.02497488021850586, 0.02477712059020996, 0.02488319969177246, 0.024841567993164063, 0.025072288513183594, 0.024809471130371095, 0.02491596794128418, 0.024825344085693358, 0.024951295852661134, 0.02479292869567871, 0.02481577682495117, 0.024796735763549804, 0.024842079162597657, 0.02481939125061035, 0.024941471099853514, 0.025402463912963868, 0.025666463851928712, 0.02488684844970703, 0.025008352279663085, 0.02474367904663086, 0.02497177505493164, 0.024999519348144532, 0.02503718376159668, 0.02476995277404785, 0.025167871475219726, 0.025025119781494142, 0.025328927993774415, 0.025344991683959962, 0.025577215194702147, 0.025436159133911132, 0.02553446388244629, 0.025482240676879882, 0.025600608825683595, 0.02544476890563965, 0.02550783920288086, 0.025241600036621094, 0.02554252815246582, 0.02533788871765137, 0.02530518341064453, 0.025028608322143556, 0.025152896881103514, 0.024932992935180663, 0.025255935668945313, 0.02512073516845703, 0.025476640701293945, 0.025128608703613282, 0.025056991577148437, 0.0250118408203125, 0.024964384078979492, 0.024946367263793946, 0.02500521659851074, 0.02510220718383789, 0.025455615997314454, 0.02520377540588379, 0.02523027229309082, 0.02487500762939453, 0.02509823989868164, 0.024999359130859374, 0.02502876853942871, 0.02515190315246582, 0.02488319969177246, 0.024844608306884765, 0.024989248275756836, 0.025072799682617188, 0.026908992767333984, 0.025842336654663085, 0.025084928512573244, 0.024902111053466798, 0.024891839981079102, 0.024878623962402344, 0.024973888397216797, 0.024964832305908204, 0.02484662437438965, 0.024862016677856445, 0.024950815200805665, 0.024883392333984376, 0.024816095352172853, 0.02498150444030762, 0.024954879760742187, 0.024877056121826172, 0.025038848876953124, 0.02485215950012207, 0.02921504020690918, 0.025169919967651368, 0.025318559646606446, 0.025064287185668947, 0.025574623107910158, 0.02497817611694336, 0.02527440071105957, 0.02488319969177246, 0.025444351196289062, 0.025257823944091796, 0.02506675148010254, 0.025123136520385742, 0.025166400909423826, 0.025045024871826173, 0.024977407455444335, 0.025083904266357423, 0.025174016952514647, 0.02500320053100586, 0.024950944900512695, 0.024939168930053712, 0.02521513557434082, 0.025284671783447267, 0.025455615997314454, 0.02527027130126953, 0.025545503616333006, 0.02532681655883789, 0.025987071990966795, 0.027271167755126953, 0.025520128250122072, 0.02533580780029297, 0.02533580780029297, 0.025298688888549806, 0.025176319122314453, 0.02519196891784668, 0.025247871398925783, 0.02591155242919922, 0.025763999938964843, 0.025853023529052735, 0.02573766326904297, 0.02585775947570801, 0.025801824569702148, 0.02584307289123535, 0.025835615158081054, 0.02587455940246582, 0.025843679428100588, 0.025784351348876952, 0.025806367874145506, 0.02584828758239746, 0.025843584060668945, 0.025843839645385742, 0.02559916877746582, 0.025612447738647463, 0.02540595245361328, 0.025315488815307617, 0.025293983459472657, 0.02538582420349121, 0.025254240036010744, 0.025355199813842773, 0.025237375259399415, 0.02511296081542969, 0.024868448257446288, 0.02510732841491699, 0.02513100814819336, 0.025051136016845704, 0.025049087524414062, 0.024958976745605467, 0.02503628730773926, 0.024993791580200195, 0.024918527603149415, 0.024922111511230468, 0.024900928497314453, 0.024953535079956055, 0.024901216506958007, 0.02500783920288086, 0.024883712768554687, 0.024952255249023437, 0.024791744232177733, 0.024956256866455077, 0.024902368545532228, 0.025132095336914063, 0.025125215530395508, 0.025335615158081054, 0.025115392684936524, 0.025878559112548827, 0.02568191909790039, 0.025593856811523437, 0.025266176223754884, 0.025395200729370116, 0.025288896560668947, 0.025277952194213867, 0.025433984756469727, 0.025473535537719725, 0.025283967971801758, 0.02520537567138672, 0.025049087524414062, 0.02510233688354492, 0.024973119735717773, 0.024968608856201172, 0.024916383743286134, 0.02497983932495117, 0.024987167358398437, 0.025274208068847656, 0.02499612808227539, 0.02508220863342285, 0.025595903396606445, 0.025273408889770508, 0.02525279998779297, 0.025016319274902343, 0.025122207641601564, 0.02485103988647461, 0.024952096939086912, 0.02486502456665039, 0.024994239807128907, 0.025068767547607423, 0.025146240234375, 0.025216896057128905, 0.025699392318725586, 0.025668607711791993, 0.025471071243286132, 0.025800607681274415, 0.025632768630981444, 0.02570035171508789, 0.025544704437255858, 0.02552217674255371, 0.025413856506347657, 0.02531011199951172, 0.025415775299072265, 0.026667808532714842, 0.025905311584472655, 0.028718847274780274, 0.026013792037963866, 0.025927007675170897, 0.026058496475219725, 0.025862367630004882, 0.025950624465942384, 0.02586038398742676, 0.025952255249023438, 0.025790239334106447, 0.025864416122436524, 0.025897151947021486, 0.025986112594604493, 0.025973440170288086, 0.0259748477935791, 0.025820735931396485, 0.02582271957397461, 0.025811487197875977, 0.02584003257751465, 0.025825279235839844, 0.025705663681030274, 0.025468896865844727, 0.025496416091918946, 0.025460960388183594, 0.025593568801879883, 0.025286943435668945, 0.02511257553100586, 0.02510438346862793, 0.025519807815551757, 0.025184095382690428, 0.025186784744262697, 0.025030399322509767, 0.024968992233276366, 0.024956512451171874, 0.024942943572998047, 0.025133600234985353, 0.02489753532409668, 0.02493235206604004, 0.024889055252075194, 0.02497532844543457, 0.024912448883056642, 0.025267967224121092, 0.025432064056396485, 0.02546892738342285, 0.025269823074340822, 0.025422271728515626, 0.025399007797241212, 0.025540895462036133, 0.025484928131103514, 0.025510271072387694, 0.025446399688720703, 0.02547235107421875, 0.025473279953002928, 0.02543657684326172, 0.025413631439208984, 0.02534726333618164, 0.025170751571655273, 0.024986656188964843, 0.025094432830810545, 0.025237279891967772, 0.025238079071044923, 0.025215328216552733, 0.024983552932739257, 0.025252960205078126, 0.025133983612060547, 0.025067520141601563, 0.024956384658813478, 0.025049631118774413, 0.024913919448852538, 0.025012224197387696, 0.024966495513916016, 0.025251935958862305, 0.025432640075683594, 0.025341951370239257, 0.02555904006958008, 0.0252192325592041, 0.025057024002075195, 0.025023679733276367, 0.024959903717041015, 0.024975360870361327, 0.02502364730834961, 0.025166688919067384, 0.024993791580200195, 0.024978752136230468, 0.025032928466796875, 0.025118175506591796, 0.025002143859863282, 0.025002464294433594, 0.02494803237915039, 0.025016191482543946, 0.02490665626525879, 0.02505904006958008, 0.025028736114501952, 0.0248240966796875, 0.024945632934570312, 0.02497737693786621, 0.024879903793334962, 0.025087871551513673, 0.025245824813842774, 0.025284608840942382, 0.024879007339477538, 0.025077407836914062, 0.02486534309387207, 0.025294527053833008, 0.025136480331420897, 0.02502454376220703, 0.02484876823425293, 0.024973760604858397, 0.02485862350463867, 0.024930303573608398, 0.024904767990112306, 0.02552707290649414, 0.024989280700683594, 0.024999807357788086, 0.026184640884399413, 0.02498739242553711, 0.0249007682800293, 0.025129215240478515, 0.02500003242492676, 0.025012319564819335, 0.02501696014404297, 0.025196319580078126, 0.02490777587890625, 0.025202688217163087, 0.024927295684814454, 0.025918239593505858, 0.024882720947265624, 0.02519126319885254, 0.024849216461181642, 0.02496406364440918, 0.02489753532409668, 0.02513715171813965, 0.027610271453857423, 0.025205120086669922, 0.025077407836914062, 0.025137983322143554, 0.024983104705810548, 0.02550239944458008, 0.02491916847229004, 0.025215103149414064, 0.02491993522644043, 0.025080448150634767, 0.024995840072631836, 0.025292448043823242, 0.02506515121459961, 0.02504361534118652, 0.024991743087768553, 0.025200895309448242, 0.024874752044677734, 0.02507776069641113, 0.02506435203552246, 0.025008127212524413, 0.024863935470581053, 0.024953664779663084, 0.025200639724731445, 0.02560153579711914, 0.024908287048339844, 0.025113983154296873, 0.02520083236694336, 0.02570230484008789, 0.02519094467163086, 0.025116895675659178, 0.02490083122253418, 0.025108863830566406, 0.02497926330566406, 0.025082239151000975, 0.024936704635620116, 0.025229055404663085, 0.025057024002075195, 0.02509644889831543, 0.024872671127319335, 0.025034624099731444, 0.024891807556152345, 0.024989696502685548, 0.025092063903808595, 0.02535958480834961, 0.025359167098999023, 0.025411584854125976, 0.025417472839355467, 0.025909503936767577, 0.02604412841796875, 0.02563039970397949, 0.02549177551269531, 0.02575103950500488, 0.025547552108764648, 0.025618080139160157, 0.025317216873168947, 0.02571683120727539, 0.025082271575927736, 0.0255729923248291, 0.02520307159423828, 0.025501407623291016, 0.02523311996459961, 0.025352640151977537, 0.025105600357055665, 0.025228096008300782, 0.025045120239257812, 0.025218496322631834, 0.025078399658203125, 0.02546886444091797, 0.025231327056884766, 0.02531724739074707, 0.025186752319335936, 0.025519840240478514, 0.025464832305908205, 0.025769983291625977, 0.025044544219970703, 0.025201087951660157, 0.024848384857177733, 0.024971263885498047, 0.02481295967102051, 0.02505379295349121, 0.0248668155670166, 0.025025535583496093, 0.025011167526245118, 0.024815616607666017, 0.024932319641113282, 0.024842336654663087, 0.024940479278564454, 0.024895488739013674, 0.024915552139282225, 0.024816032409667968, 0.02481705665588379, 0.02507632064819336, 0.024964927673339844, 0.024950944900512695, 0.024994911193847655, 0.024959936141967773, 0.025161727905273438, 0.02789990425109863, 0.025964544296264647, 0.02520579147338867, 0.02514633560180664, 0.025165824890136718, 0.024997888565063478, 0.025156639099121095, 0.025123807907104494, 0.02503014373779297, 0.025022111892700195, 0.0248919677734375, 0.02502176094055176, 0.024839136123657228, 0.025034751892089844, 0.02495052719116211, 0.025131263732910157, 0.024845951080322264, 0.025371007919311524, 0.026218496322631835, 0.025219072341918947, 0.025310880661010744, 0.024999935150146483, 0.02489904022216797, 0.025031391143798827, 0.024898015975952148, 0.024921791076660156, 0.025159679412841796, 0.02530873680114746, 0.02725641632080078, 0.025234272003173828, 0.024887584686279298, 0.025093215942382813, 0.024875648498535158, 0.025057279586791992, 0.02513043212890625, 0.024945215225219728, 0.024856512069702148, 0.02499180793762207, 0.024862592697143554, 0.025022592544555664, 0.025220191955566407, 0.02549852752685547, 0.02510643196105957, 0.025133056640625, 0.02519478416442871, 0.0249946231842041, 0.02509212875366211]",tokens/s,39.59127553972007,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3130.662912,4369.350656,0.0,3990.880256,3908.719616,s,1,10.236123046875,10.236123046875,0.0,10.236123046875,10.236123046875,10.236123046875,10.236123046875,[10.236123046875],,kWh,9.583114598749489e-05,1.056201819548879e-05,3.11369693540009e-05,0.00013753013353698458,,MB,3038.547968,4736.352256,0.0,4328.521728,4275.211264,s,10,2.2004496765136716,0.22004496765136716,0.0017202474563007386,0.22012940979003906,0.22196122589111328,0.22231882400512695,0.22260490249633788,"[0.21974082946777343, 0.21711001586914064, 0.21838937377929687, 0.21789814758300782, 0.22267642211914063, 0.22020541381835937, 0.22084617614746094, 0.22164813232421876, 0.22005340576171875, 0.2218817596435547]",tokens/s,1163.3985668129387,kWh,6.732886124338178e-06,7.425168213095678e-07,4.457497252863894e-06,1.193290019851164e-05,tokens/kWh,21453292.63978343,MB,3045.90848,4738.449408,0.0,4330.61888,4275.213824,s,10,33.801979003906254,3.3801979003906255,0.032645405571140824,3.3820683593750003,3.4161689453125,3.4249503173828124,3.4319754150390627,"[3.38634912109375, 3.319181640625, 3.352884765625, 3.361162353515625, 3.3530166015625, 3.414217529296875, 3.37778759765625, 3.395310546875, 3.433731689453125, 3.408337158203125]",tokens/s,18.637961994094944,kWh,9.909167303066185e-05,1.0930001910182555e-05,5.104018603413411e-05,0.0001610618609749785,tokens/kWh,391154.0548372731,,s,630,33.79927659225466,0.05364964538453117,0.0010443882622783844,0.05361448097229004,0.05456036338806153,0.054943120002746586,0.05713659645080567,"[0.05480243301391602, 0.05358102416992187, 0.053728031158447265, 0.053386878967285153, 0.05339993667602539, 0.0534466552734375, 0.053437599182128905, 0.05309638214111328, 0.053158111572265625, 0.05378947067260742, 0.05319260787963867, 0.05398527908325195, 0.05310665512084961, 0.052916255950927735, 0.052574207305908206, 0.052795391082763675, 0.05241142272949219, 0.05253152084350586, 0.05255644989013672, 0.05395225524902344, 0.05399343872070313, 0.05269942474365234, 0.05240627288818359, 0.05375907135009766, 0.05234985733032226, 0.05235004806518555, 0.05232665634155274, 0.05287184143066406, 0.05320243072509766, 0.05378623962402344, 0.05387148666381836, 0.05370265579223633, 0.05356460952758789, 0.05338604736328125, 0.05316198348999023, 0.053712417602539066, 0.05423283386230469, 0.0544222412109375, 0.05471641540527344, 0.05437827301025391, 0.05440902328491211, 0.053658016204833986, 0.054441761016845704, 0.053739105224609375, 0.05370124816894531, 0.06377881622314453, 0.054335487365722655, 0.05368012619018555, 0.05439692687988281, 0.05446870422363281, 0.0536143684387207, 0.0535777587890625, 0.05308339309692383, 0.05554399871826172, 0.05357020950317383, 0.05351580810546875, 0.05327052688598633, 0.05431343841552734, 0.054075393676757816, 0.054130687713623046, 0.054017822265625, 0.05412681579589844, 0.05391360092163086, 0.0547993278503418, 0.05399027252197266, 0.053617855072021485, 0.054067134857177734, 0.05375475311279297, 0.05353014373779297, 0.05260131072998047, 0.052326400756835936, 0.05258444976806641, 0.05231814575195313, 0.052240478515625, 0.05203945541381836, 0.05205180740356445, 0.05215881729125976, 0.05219446563720703, 0.05219955062866211, 0.05222598266601562, 0.0520497932434082, 0.052155296325683595, 0.05246895980834961, 0.05209276962280274, 0.05224281692504883, 0.05212326431274414, 0.052026016235351566, 0.05203936004638672, 0.052095584869384766, 0.05211328125, 0.05256969451904297, 0.05288399887084961, 0.052119327545166017, 0.05207241439819336, 0.055653663635253904, 0.05628598403930664, 0.05245561599731445, 0.052325439453125, 0.0523922233581543, 0.05236086273193359, 0.05240224075317383, 0.05297657775878906, 0.052319297790527346, 0.05215078353881836, 0.052418846130371094, 0.05224204635620117, 0.05223040008544922, 0.05216899108886719, 0.052974689483642576, 0.05280828857421875, 0.052746337890625, 0.052480224609375, 0.052506622314453126, 0.0522608642578125, 0.052221630096435545, 0.05275174331665039, 0.05282271957397461, 0.05284377670288086, 0.05246054458618164, 0.05275155258178711, 0.05264851379394531, 0.05244099044799805, 0.052564319610595704, 0.05306163024902344, 0.05315523147583008, 0.05326704025268555, 0.05422288131713867, 0.05315785598754883, 0.052926464080810545, 0.05264384078979492, 0.05261011123657226, 0.05288032150268555, 0.05364726257324219, 0.053588062286376956, 0.05414425659179688, 0.053564159393310544, 0.05352352142333985, 0.05328377532958985, 0.05302377700805664, 0.053220096588134765, 0.05322476959228516, 0.052911006927490234, 0.052587966918945316, 0.05282787322998047, 0.05284745788574219, 0.053174049377441406, 0.053166305541992184, 0.05323980712890625, 0.053741214752197265, 0.05294457626342773, 0.05263836669921875, 0.057360511779785156, 0.05283785629272461, 0.052854976654052734, 0.05238560104370117, 0.052455711364746097, 0.05264191818237305, 0.05219126510620117, 0.052195297241210935, 0.05267660903930664, 0.053542911529541014, 0.05321932983398438, 0.054080894470214844, 0.053460895538330076, 0.05340643310546875, 0.05302403259277344, 0.05317705535888672, 0.05312102508544922, 0.05313049697875977, 0.053166561126708985, 0.05286921691894531, 0.0528089599609375, 0.05349846267700195, 0.053534976959228514, 0.053491809844970706, 0.05464678573608398, 0.05489049530029297, 0.053629150390625, 0.05337680053710937, 0.053181854248046875, 0.05314982223510742, 0.05329072189331055, 0.053004257202148436, 0.05283900833129883, 0.052312255859375, 0.052453182220458985, 0.05285808181762695, 0.05335363388061523, 0.05276633453369141, 0.053014816284179686, 0.05275033569335937, 0.05291747283935547, 0.05288768005371094, 0.052937374114990235, 0.053233409881591795, 0.05321139144897461, 0.05363622283935547, 0.053539710998535155, 0.053501953125, 0.05447407913208008, 0.05378319931030273, 0.05381324768066406, 0.05372313690185547, 0.053855712890625, 0.05365011215209961, 0.05393718338012695, 0.05410268783569336, 0.05383411026000977, 0.053114654541015625, 0.052873374938964844, 0.052709217071533206, 0.052776607513427734, 0.052943199157714844, 0.05279081726074219, 0.05257577514648438, 0.05285577774047852, 0.052559009552001955, 0.052589374542236327, 0.052340862274169925, 0.05268191909790039, 0.05230662536621094, 0.058383647918701174, 0.05350230407714844, 0.053082496643066406, 0.053073566436767576, 0.05334460830688476, 0.0530882568359375, 0.05266236877441406, 0.05338723373413086, 0.05539158248901367, 0.05363772964477539, 0.05295849609375, 0.05274867248535156, 0.05369071960449219, 0.052776641845703125, 0.052865345001220705, 0.06351049423217774, 0.05330742263793945, 0.052776958465576174, 0.05253049468994141, 0.052746849060058595, 0.05250672149658203, 0.05286297607421875, 0.05277196884155273, 0.05285366439819336, 0.052571582794189456, 0.05265375900268555, 0.0525013427734375, 0.05266620635986328, 0.05246787261962891, 0.05311689758300781, 0.05257206344604492, 0.05323356628417969, 0.05270732879638672, 0.05248803329467774, 0.052455585479736326, 0.052432415008544925, 0.05242723083496094, 0.05248409652709961, 0.05254143905639649, 0.05251862335205078, 0.05250393676757813, 0.05237164688110352, 0.05238224029541016, 0.05292665481567383, 0.053370174407958985, 0.05408633422851562, 0.05387667083740234, 0.05377750396728516, 0.053212127685546874, 0.052916255950927735, 0.05769823837280273, 0.05325417709350586, 0.05283625411987305, 0.052516960144042966, 0.052585601806640625, 0.05281881713867188, 0.05350297546386719, 0.0535843505859375, 0.0533837776184082, 0.05354079818725586, 0.05336812973022461, 0.0533633918762207, 0.05304707336425781, 0.05301475143432617, 0.05306163024902344, 0.05318143844604492, 0.052946174621582034, 0.05290927886962891, 0.05265462493896484, 0.052739456176757814, 0.053451393127441404, 0.05321664047241211, 0.05294963073730469, 0.052836353302001954, 0.05279948806762695, 0.052727008819580076, 0.052625823974609375, 0.05322351837158203, 0.05362054443359375, 0.05343484878540039, 0.05328700637817383, 0.053423969268798825, 0.0535115852355957, 0.05415388870239258, 0.053287071228027345, 0.05335228729248047, 0.05337484741210938, 0.05406943893432617, 0.05361971282958984, 0.05344966506958008, 0.05361692810058594, 0.05404848098754883, 0.053907455444335936, 0.05401804733276367, 0.0551649284362793, 0.05720873641967773, 0.055070110321044925, 0.05451232147216797, 0.055285598754882814, 0.05459743881225586, 0.05475363159179687, 0.054314590454101565, 0.054290752410888675, 0.0546038703918457, 0.05451718521118164, 0.05411078262329101, 0.054296577453613284, 0.054085601806640626, 0.05408476638793945, 0.05424342346191406, 0.05364556884765625, 0.05339926528930664, 0.05332064056396484, 0.05364108657836914, 0.05361868667602539, 0.05426287841796875, 0.05459241485595703, 0.054095775604248046, 0.054331489562988285, 0.054507518768310545, 0.05428377532958984, 0.053690879821777344, 0.05373299026489258, 0.05389503860473633, 0.05368057632446289, 0.053682239532470706, 0.05369356918334961, 0.05329110336303711, 0.053352222442626954, 0.05645004653930664, 0.05396684646606445, 0.05361459350585938, 0.052924415588378904, 0.05524684906005859, 0.05341129684448242, 0.053148193359375, 0.05316812896728516, 0.05402627182006836, 0.05388284683227539, 0.0543719367980957, 0.053967262268066404, 0.055332160949707034, 0.054155967712402345, 0.054216064453125, 0.054325630187988284, 0.054311038970947266, 0.05353279876708984, 0.05370022583007812, 0.05370425415039062, 0.053598880767822266, 0.05382364654541016, 0.05408563232421875, 0.054986751556396485, 0.054919361114501956, 0.053802814483642575, 0.05419766235351563, 0.05519011306762695, 0.054475616455078125, 0.05407299041748047, 0.05416585540771485, 0.053819393157958986, 0.05390950393676758, 0.053370208740234376, 0.05365731048583984, 0.052894271850585935, 0.05316755294799805, 0.05318041610717773, 0.05298067092895508, 0.05260902404785156, 0.05287526321411133, 0.05296537780761719, 0.05282524871826172, 0.053016704559326173, 0.05273225784301758, 0.05531292724609375, 0.05382128143310547, 0.05349980926513672, 0.053673439025878907, 0.05392601776123047, 0.05335897445678711, 0.05333414459228516, 0.05353267288208008, 0.05703472137451172, 0.05397414398193359, 0.053963615417480466, 0.05384304046630859, 0.05471942520141602, 0.05372214508056641, 0.053700702667236325, 0.05388092803955078, 0.053570369720458984, 0.05368012619018555, 0.053198623657226565, 0.05335881423950195, 0.05337519836425781, 0.053403167724609374, 0.0535340805053711, 0.05433433532714844, 0.05351222229003906, 0.053165950775146485, 0.05329446411132813, 0.05303104019165039, 0.05296758270263672, 0.05301248168945313, 0.05300883102416992, 0.05416454315185547, 0.05375481414794922, 0.05336441421508789, 0.05393644714355469, 0.0537784309387207, 0.05390729522705078, 0.053235870361328125, 0.05349577713012695, 0.05358969497680664, 0.05365184020996094, 0.054067073822021486, 0.05369865417480469, 0.053609664916992185, 0.05351916885375976, 0.053303295135498044, 0.05445817565917969, 0.05360655975341797, 0.05398940658569336, 0.053823486328125, 0.056296863555908204, 0.053725791931152345, 0.053583038330078124, 0.05364406585693359, 0.05339116668701172, 0.053655776977539066, 0.05349286270141602, 0.05282905578613281, 0.05293875122070312, 0.052884544372558594, 0.053205951690673825, 0.053703807830810545, 0.0538710708618164, 0.05379219055175781, 0.05412550354003906, 0.05394761657714844, 0.05360108947753906, 0.05417977523803711, 0.05408291244506836, 0.053891807556152346, 0.05382070541381836, 0.05356003189086914, 0.053585918426513675, 0.05346259307861328, 0.054109825134277346, 0.053957439422607424, 0.05406924819946289, 0.05398332977294922, 0.05420431900024414, 0.053846080780029296, 0.05406508636474609, 0.053661697387695315, 0.0538869743347168, 0.05381891250610352, 0.053750240325927734, 0.05360547256469726, 0.054317440032958984, 0.05370729446411133, 0.05378262329101562, 0.054206207275390626, 0.05376015853881836, 0.05360435104370117, 0.05353267288208008, 0.05413273620605469, 0.054177310943603514, 0.05401033782958985, 0.05406105422973633, 0.053907455444335936, 0.053905406951904294, 0.05391769790649414, 0.05420851135253906, 0.05396480178833008, 0.056586238861083986, 0.053553150177001956, 0.05374156951904297, 0.05399347305297852, 0.053982273101806644, 0.05411718368530274, 0.053759391784667966, 0.054728382110595705, 0.054092353820800784, 0.05455974578857422, 0.05424630355834961, 0.05455062484741211, 0.0545054702758789, 0.055119873046875, 0.05449728012084961, 0.05472003173828125, 0.05480495834350586, 0.05450688171386719, 0.05454092788696289, 0.05405491256713867, 0.05400985717773438, 0.05432633590698242, 0.054047679901123045, 0.05411840057373047, 0.05493465423583985, 0.054558624267578126, 0.05595849609375, 0.055512863159179686, 0.05478422546386719, 0.05467750549316406, 0.05546303939819336, 0.05455756759643555, 0.05460508728027344, 0.05447679901123047, 0.05418819046020508, 0.05427872085571289, 0.05456592178344727, 0.0542852783203125, 0.05427123260498047, 0.054427745819091794, 0.05469046401977539, 0.05466726303100586, 0.054816959381103515, 0.05472172927856445, 0.05469862365722656, 0.05461814498901367, 0.054790111541748045, 0.054540287017822264, 0.054766719818115234, 0.05428627014160156, 0.056070240020751956, 0.05419865417480469, 0.0541065902709961, 0.05545574569702148, 0.054093822479248044, 0.054235038757324217, 0.05436380767822266, 0.05433152008056641, 0.054322784423828124, 0.054366943359375, 0.0540263671875, 0.05409747314453125, 0.05416172790527344, 0.05374156951904297, 0.05396835327148437, 0.053551361083984374, 0.054130817413330076, 0.054353664398193356, 0.05412918472290039, 0.054155136108398436, 0.055801246643066404, 0.05488259124755859, 0.053911872863769535, 0.05345619201660156, 0.0531297607421875, 0.05495004653930664, 0.054165695190429686, 0.0539824333190918, 0.054852096557617185, 0.05717820739746094, 0.05409817504882813, 0.054021472930908206, 0.054118080139160155, 0.054459102630615236, 0.053866497039794924, 0.053389022827148434, 0.053285152435302734, 0.05335881423950195, 0.053430049896240235, 0.053661697387695315, 0.05384396743774414, 0.05350809478759765, 0.053424095153808596, 0.0535142707824707, 0.05340678405761719, 0.053339073181152344, 0.05369651031494141, 0.054214656829833986, 0.05489254379272461, 0.05402563095092774, 0.05417417526245117, 0.054042369842529296, 0.05391580963134766, 0.053953758239746095, 0.053926719665527346, 0.0537081298828125, 0.05396128082275391, 0.05381148910522461, 0.05386614227294922, 0.05403596878051758, 0.05417407989501953, 0.05408979034423828, 0.05416387176513672, 0.05414912033081055, 0.05408723068237305, 0.054067615509033204, 0.05381289672851562, 0.053924095153808596, 0.05397862243652344, 0.054037120819091795, 0.05407660675048828, 0.05345158386230469, 0.053208927154541015, 0.05412665557861328, 0.05431856155395508, 0.056236671447753905, 0.05455462265014648, 0.054489086151123044, 0.05464995193481445, 0.05432352066040039, 0.054245281219482425, 0.05433529663085938, 0.054293376922607425]",tokens/s,18.639452187103004,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1015.31648,867.106816,0.0,488.636416,482.553856,s,1,8.1101240234375,8.1101240234375,0.0,8.1101240234375,8.1101240234375,8.1101240234375,8.1101240234375,[8.1101240234375],,kWh,2.895301289996344e-05,3.1790054850173666e-06,8.45528454196387e-06,4.0587302926944674e-05,,MB,1269.141504,1024.393216,0.0,616.562688,581.925888,s,10,0.3571811180114745,0.03571811180114746,0.00024128084346798723,0.03568595123291016,0.0359390510559082,0.03610627670288086,0.036240057220458984,"[0.03627350234985351, 0.035744129180908205, 0.03564166259765625, 0.035899871826171874, 0.03551369476318359, 0.035539169311523434, 0.035438591003417966, 0.035730239868164065, 0.03549836730957031, 0.03590188980102539]",tokens/s,7167.232171320319,kWh,1.1306078822072933e-06,1.246858942938572e-07,7.525145445251156e-07,2.007808321026266e-06,tokens/kWh,127502210.9028559,MB,1302.872064,1039.07328,0.0,631.242752,597.192192,s,10,17.16676220703125,1.7166762207031248,0.010543409282530449,1.7162196655273436,1.7303883056640625,1.73304150390625,1.7351640625,"[1.7196602783203125, 1.7171500244140625, 1.706823974609375, 1.715289306640625, 1.6963521728515625, 1.7125345458984376, 1.7297987060546876, 1.7356947021484375, 1.71229150390625, 1.7211669921875]",tokens/s,36.69882488044026,kWh,5.007109421654183e-05,5.522389019015717e-06,1.920022363987546e-05,7.479370687543301e-05,tokens/kWh,842316.8556804501,,s,630,17.161206449508658,0.027240010237315346,0.0005188733701594881,0.02709940814971924,0.02779759616851807,0.02792737331390381,0.029234470233917238,"[0.027675647735595704, 0.027025407791137695, 0.0269354248046875, 0.026990463256835937, 0.026808319091796876, 0.02679110336303711, 0.027115488052368165, 0.02669843292236328, 0.027004255294799804, 0.026748863220214844, 0.02677235221862793, 0.026746879577636717, 0.02672230339050293, 0.02672960090637207, 0.026765727996826173, 0.026963808059692382, 0.026674816131591797, 0.026856447219848634, 0.02701692771911621, 0.027678272247314454, 0.02742755126953125, 0.027883487701416014, 0.02728940773010254, 0.027338783264160157, 0.026982175827026368, 0.026990655899047852, 0.027154399871826173, 0.027189151763916015, 0.027142623901367187, 0.027060224533081056, 0.027894880294799803, 0.027333696365356444, 0.02734467124938965, 0.02705824089050293, 0.027176895141601563, 0.027256128311157226, 0.027377792358398437, 0.02716476821899414, 0.02755638313293457, 0.027098592758178712, 0.027162815093994142, 0.02699299240112305, 0.02683679962158203, 0.027074880599975586, 0.026856416702270507, 0.027110111236572264, 0.027228511810302735, 0.027453279495239256, 0.028520448684692383, 0.02765945625305176, 0.02781279945373535, 0.027895360946655273, 0.027851295471191407, 0.02776006317138672, 0.02769875144958496, 0.027867712020874024, 0.02777110481262207, 0.027824447631835936, 0.02787705612182617, 0.028055168151855468, 0.02783475112915039, 0.02780521583557129, 0.02767305564880371, 0.027300031661987304, 0.027410240173339845, 0.027450559616088867, 0.026913600921630858, 0.026771455764770507, 0.026738687515258788, 0.02669068717956543, 0.026907520294189455, 0.027125759124755858, 0.02925542449951172, 0.027385696411132813, 0.026886528015136718, 0.026765344619750976, 0.026617855072021485, 0.026617855072021485, 0.026732255935668945, 0.02668367958068848, 0.026648832321166993, 0.026676095962524415, 0.02670476722717285, 0.02675302314758301, 0.02669935989379883, 0.02660598373413086, 0.02654207992553711, 0.030242816925048828, 0.02679311943054199, 0.026712352752685548, 0.026828704833984376, 0.027150976181030274, 0.027025440216064452, 0.02676736068725586, 0.026937280654907227, 0.02674643135070801, 0.026735103607177735, 0.026891807556152343, 0.029122848510742188, 0.028404224395751954, 0.027829023361206056, 0.027661216735839843, 0.02795315170288086, 0.02754256057739258, 0.027666559219360353, 0.02746985626220703, 0.027693056106567384, 0.02756211280822754, 0.027625696182250976, 0.02750326347351074, 0.02740947151184082, 0.027459455490112306, 0.027582847595214843, 0.027478527069091797, 0.027495744705200196, 0.02770207977294922, 0.02740559959411621, 0.027372127532958986, 0.027262975692749023, 0.027067935943603516, 0.02691440010070801, 0.027276159286499024, 0.026869760513305665, 0.02711507225036621, 0.027474111557006835, 0.027023616790771483, 0.02651817512512207, 0.02729759979248047, 0.028847360610961915, 0.02705299186706543, 0.027455039978027344, 0.02675059127807617, 0.02670675277709961, 0.026613759994506835, 0.026705919265747072, 0.026695680618286134, 0.02674844741821289, 0.026632671356201173, 0.02663417625427246, 0.0267194881439209, 0.026714912414550783, 0.026703903198242188, 0.026847232818603517, 0.02680169677734375, 0.02704636764526367, 0.02693120002746582, 0.026697727203369142, 0.026681343078613282, 0.02672982406616211, 0.026655391693115236, 0.02672640037536621, 0.02676940727233887, 0.026771455764770507, 0.026730495452880858, 0.027097215652465822, 0.02721900749206543, 0.027067071914672853, 0.026871583938598634, 0.02689039993286133, 0.02694588851928711, 0.026729791641235352, 0.02679043197631836, 0.02678374481201172, 0.027236320495605468, 0.027318336486816405, 0.027922655105590822, 0.02765817642211914, 0.02728563117980957, 0.026850368499755858, 0.02679772758483887, 0.026897119522094726, 0.02688640022277832, 0.0269434871673584, 0.02719875144958496, 0.0270097599029541, 0.027060224533081056, 0.027112480163574218, 0.027169759750366212, 0.02699673652648926, 0.027122976303100586, 0.027654111862182616, 0.027797567367553712, 0.027767488479614258, 0.02777292823791504, 0.027893024444580077, 0.027717472076416016, 0.027693952560424805, 0.027822080612182616, 0.028004352569580077, 0.02742425537109375, 0.027900415420532225, 0.02795644760131836, 0.028986143112182616, 0.028463104248046874, 0.02780963134765625, 0.02746384048461914, 0.027288959503173827, 0.027361343383789063, 0.02714681625366211, 0.026908672332763672, 0.02733260726928711, 0.0269550724029541, 0.026812864303588868, 0.026800384521484377, 0.026720256805419923, 0.026617984771728515, 0.026877824783325194, 0.026742271423339844, 0.028252447128295898, 0.027893695831298828, 0.026915103912353515, 0.026877952575683595, 0.026728288650512695, 0.026975519180297853, 0.026800512313842773, 0.02674265670776367, 0.026993280410766603, 0.027254783630371093, 0.027046176910400392, 0.02710905647277832, 0.027419872283935547, 0.027667264938354492, 0.027735359191894533, 0.027765439987182616, 0.02755075263977051, 0.02761440086364746, 0.027372928619384767, 0.027461151123046874, 0.027072608947753905, 0.0268953914642334, 0.02672368049621582, 0.026970527648925782, 0.02714419174194336, 0.02699673652648926, 0.02691196823120117, 0.026907424926757812, 0.026807872772216797, 0.028145183563232423, 0.02776678466796875, 0.027022239685058593, 0.026944831848144533, 0.026700031280517577, 0.027038143157958983, 0.02727071952819824, 0.027234752655029296, 0.027164512634277344, 0.026893728256225585, 0.026655487060546875, 0.026953727722167968, 0.026986495971679687, 0.02686502456665039, 0.026943552017211915, 0.026728448867797853, 0.02694704055786133, 0.026892831802368164, 0.026705919265747072, 0.026816511154174806, 0.026591232299804687, 0.02667241668701172, 0.02681248092651367, 0.026739360809326172, 0.026641759872436523, 0.026762079238891602, 0.02673206329345703, 0.02664886474609375, 0.026695680618286134, 0.02654003143310547, 0.026660863876342773, 0.02664019203186035, 0.026763456344604492, 0.02702035140991211, 0.02685638427734375, 0.026736640930175783, 0.026818559646606444, 0.026828351974487304, 0.026823104858398436, 0.0268155517578125, 0.02691516876220703, 0.027329120635986328, 0.02701312065124512, 0.027061824798583985, 0.026874271392822266, 0.02689846420288086, 0.027017215728759765, 0.027691007614135742, 0.027550783157348633, 0.027323328018188476, 0.02734694480895996, 0.027467967987060547, 0.02806937599182129, 0.02718137550354004, 0.027092992782592775, 0.027209728240966798, 0.026801376342773436, 0.026874975204467775, 0.027419551849365235, 0.02711222457885742, 0.026906272888183595, 0.026806623458862304, 0.026763263702392577, 0.026881311416625978, 0.02697075271606445, 0.027066463470458983, 0.027076608657836915, 0.026849023818969725, 0.026689056396484376, 0.0265960636138916, 0.02690662384033203, 0.026607328414916993, 0.026753183364868163, 0.02665894317626953, 0.026884096145629883, 0.02680143928527832, 0.026740671157836914, 0.02675996780395508, 0.026572799682617186, 0.02689023971557617, 0.02685923194885254, 0.026876192092895507, 0.026771455764770507, 0.026799488067626952, 0.026780288696289064, 0.026832767486572266, 0.026984575271606446, 0.02704150390625, 0.02704207992553711, 0.027140096664428712, 0.027099136352539063, 0.027183103561401366, 0.027020896911621094, 0.027090431213378906, 0.027034528732299806, 0.02702739143371582, 0.026974271774291993, 0.02688204765319824, 0.026955360412597655, 0.027198911666870117, 0.0270284481048584, 0.02695599937438965, 0.027191072463989257, 0.02724870491027832, 0.02709907150268555, 0.02693939208984375, 0.027181055068969725, 0.027170976638793944, 0.02739616012573242, 0.02768160057067871, 0.02782512092590332, 0.02775859260559082, 0.027550848007202148, 0.027509248733520508, 0.02762326431274414, 0.02776915168762207, 0.0277926082611084, 0.0275382080078125, 0.027506879806518555, 0.027514911651611327, 0.027639455795288086, 0.027627647399902342, 0.02737993621826172, 0.02717081642150879, 0.02733670425415039, 0.02750828742980957, 0.027148767471313475, 0.027076576232910158, 0.027256799697875977, 0.02745756721496582, 0.02735923194885254, 0.027586559295654296, 0.027131168365478516, 0.026825439453125, 0.026828800201416016, 0.026736480712890625, 0.02685148811340332, 0.026959871292114256, 0.026947519302368165, 0.027002784729003908, 0.026846752166748047, 0.026652063369750977, 0.027099679946899415, 0.02711155128479004, 0.02717750358581543, 0.026966976165771483, 0.02694003105163574, 0.026951776504516602, 0.02674518394470215, 0.027111072540283204, 0.027107328414916993, 0.026789888381958008, 0.026836448669433594, 0.026787679672241212, 0.026834911346435546, 0.02684364891052246, 0.02695395278930664, 0.027218975067138673, 0.027685855865478514, 0.027789024353027342, 0.027781408309936525, 0.02851430320739746, 0.029511680603027345, 0.030965856552124024, 0.027887840270996094, 0.027801279067993165, 0.027985664367675783, 0.02776835250854492, 0.027732032775878906, 0.02792723274230957, 0.02804230308532715, 0.027927488327026368, 0.027728927612304687, 0.0275567684173584, 0.02750048065185547, 0.02722831916809082, 0.027073600769042968, 0.02691600036621094, 0.027563680648803712, 0.0268657283782959, 0.026861568450927735, 0.027056127548217773, 0.026986431121826172, 0.0273918399810791, 0.027475872039794923, 0.027519296646118165, 0.02794495964050293, 0.02756387138366699, 0.027594911575317384, 0.02753657531738281, 0.027451616287231445, 0.027570783615112306, 0.027578464508056642, 0.02762656021118164, 0.027341663360595705, 0.02739200019836426, 0.02794495964050293, 0.02750771141052246, 0.027405311584472656, 0.027289600372314454, 0.02737299156188965, 0.02699673652648926, 0.026806272506713868, 0.02713043212890625, 0.0267509765625, 0.027082752227783204, 0.029183168411254883, 0.0270601921081543, 0.027363616943359374, 0.0274366397857666, 0.02719228744506836, 0.029930944442749022, 0.02762588882446289, 0.02756537628173828, 0.02751148796081543, 0.027473440170288087, 0.02741926383972168, 0.027438432693481445, 0.027830944061279297, 0.027522079467773436, 0.027829216003417968, 0.03010918426513672, 0.027225631713867188, 0.027374080657958984, 0.027316640853881836, 0.027043903350830078, 0.02692131233215332, 0.026938175201416014, 0.027058944702148438, 0.027060096740722656, 0.027314207077026368, 0.0273635196685791, 0.027230207443237304, 0.027236255645751953, 0.0275313606262207, 0.027582592010498046, 0.02736729621887207, 0.027313695907592774, 0.027101984024047853, 0.026998464584350585, 0.02705161666870117, 0.027558559417724608, 0.02771718406677246, 0.027769023895263673, 0.02766819190979004, 0.027689407348632813, 0.027718591690063476, 0.027810720443725585, 0.027717695236206055, 0.027776960372924805, 0.027758848190307616, 0.027612415313720703, 0.029039295196533203, 0.027840320587158202, 0.02787705612182617, 0.028334400177001954, 0.027736320495605468, 0.028142784118652345, 0.027413055419921874, 0.02750048065185547, 0.02742073631286621, 0.02735923194885254, 0.027064096450805663, 0.026884319305419922, 0.026790176391601563, 0.0267238712310791, 0.02691004753112793, 0.02657148742675781, 0.026826751708984374, 0.026949567794799803, 0.02719526481628418, 0.027017375946044923, 0.027109632492065428, 0.027217695236206055, 0.027428863525390625, 0.02742032051086426, 0.02733296012878418, 0.027085056304931642, 0.026963104248046876, 0.026813024520874022, 0.027033599853515625, 0.027056127548217773, 0.027170047760009766, 0.026650623321533205, 0.0268374080657959, 0.026980703353881835, 0.02696396827697754, 0.02694144058227539, 0.026957151412963867, 0.027073183059692384, 0.027066368103027344, 0.0276081600189209, 0.02755471992492676, 0.027613183975219727, 0.027027231216430664, 0.02705135917663574, 0.02694009590148926, 0.026818016052246093, 0.026875776290893556, 0.027091392517089842, 0.026954143524169923, 0.02703385543823242, 0.02754979133605957, 0.02772777557373047, 0.02759561538696289, 0.02744403266906738, 0.027399904251098634, 0.02721232032775879, 0.02703548812866211, 0.027013023376464843, 0.02701683235168457, 0.027056608200073242, 0.02689580726623535, 0.02703971290588379, 0.026884864807128907, 0.027742271423339845, 0.0286176643371582, 0.027697792053222658, 0.02693734359741211, 0.0267675838470459, 0.02695577621459961, 0.02735103988647461, 0.027060159683227537, 0.027136064529418944, 0.02730803108215332, 0.02735206413269043, 0.027282560348510742, 0.027322240829467773, 0.02780486488342285, 0.027283807754516602, 0.02771011161804199, 0.029391136169433594, 0.027764799118041993, 0.0275742073059082, 0.027507743835449218, 0.027846656799316406, 0.02774457550048828, 0.02774700736999512, 0.02823779106140137, 0.027706432342529296, 0.027570783615112306, 0.027691423416137697, 0.02768492889404297, 0.027774911880493164, 0.027725759506225585, 0.027731327056884764, 0.027709440231323244, 0.02770800018310547, 0.02782316780090332, 0.027797855377197266, 0.027777727127075196, 0.02766431999206543, 0.027555200576782228, 0.02740902328491211, 0.027363391876220704, 0.027571231842041015, 0.027827104568481444, 0.02751804733276367, 0.027197856903076172, 0.026838592529296875, 0.026887104034423827, 0.026804224014282226, 0.026740480422973632, 0.027191551208496093, 0.02691276741027832, 0.0271646728515625, 0.02663827133178711, 0.026941024780273437, 0.02692553520202637, 0.026850688934326173, 0.027027168273925782, 0.02703657531738281, 0.026921247482299803, 0.026982112884521483, 0.02694265556335449, 0.027019968032836916, 0.02691494369506836, 0.02677299118041992, 0.026746368408203124, 0.026895360946655275, 0.026952863693237305, 0.02690287971496582, 0.02679612731933594, 0.026759103775024416, 0.02676140785217285, 0.026656063079833984, 0.027648992538452148, 0.027062271118164064, 0.027082752227783204, 0.02720774459838867, 0.026980287551879884, 0.026990591049194337, 0.027250303268432616]",tokens/s,36.710705733514274,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1550.159872,1546.584064,0.0,1168.113664,1154.613248,s,1,8.27979296875,8.27979296875,0.0,8.27979296875,8.27979296875,8.27979296875,8.27979296875,[8.27979296875],,kWh,3.8902994579211734e-05,4.283699223596933e-06,1.2075565215999284e-05,5.526225901880795e-05,,MB,1575.97696,1796.145152,0.0,1388.314624,1334.065152,s,10,0.8412918395996094,0.08412918395996094,0.0007119479755757377,0.08390304183959961,0.08507656478881835,0.08525823402404785,0.08540356941223144,"[0.08503619384765625, 0.08369420623779297, 0.08369564819335938, 0.08416291046142578, 0.08543990325927735, 0.08339462280273438, 0.08411043548583984, 0.08362022399902344, 0.08324924468994141, 0.08488845062255859]",tokens/s,3042.939298232542,kWh,2.653956018371059e-06,2.9267770883412247e-07,1.7620014096001901e-06,4.708635136805371e-06,tokens/kWh,54368196.42255955,MB,1581.961216,1796.145152,0.0,1388.314624,1372.847616,s,10,17.37643688964844,1.7376436889648434,0.009143634783886863,1.7364484252929686,1.7483337524414062,1.749937811279297,1.7512210583496093,"[1.7515418701171874, 1.7352840576171875, 1.7309111328125, 1.73761279296875, 1.746974609375, 1.7248458251953125, 1.7445201416015625, 1.731179443359375, 1.747977294921875, 1.7255897216796876]",tokens/s,36.25599448269549,kWh,5.164052417496419e-05,5.69565426621506e-06,2.3607491108198663e-05,8.094366954937792e-05,tokens/kWh,778319.0501583107,,s,630,17.37390524673463,0.027577627375769236,0.0005843101071754203,0.027507391929626465,0.027891829681396484,0.02806671571731567,0.0305977998161316,"[0.02744688034057617, 0.027542240142822267, 0.027391679763793947, 0.027338495254516603, 0.027394304275512694, 0.027510143280029296, 0.02774412727355957, 0.028047264099121092, 0.027975744247436523, 0.028023584365844727, 0.02783203125, 0.02764041519165039, 0.027354816436767578, 0.027553056716918944, 0.027392160415649413, 0.027251007080078125, 0.02729350471496582, 0.027663040161132812, 0.027280799865722655, 0.027158464431762695, 0.02718160057067871, 0.02736934471130371, 0.02735875129699707, 0.027222400665283204, 0.02726076889038086, 0.027504255294799804, 0.027748319625854494, 0.02782480049133301, 0.027906047821044923, 0.02817401695251465, 0.02766640090942383, 0.02770569610595703, 0.027873279571533204, 0.027684223175048827, 0.02768739128112793, 0.027562143325805664, 0.027461631774902344, 0.027249727249145508, 0.027368383407592775, 0.027613279342651367, 0.027570079803466797, 0.027463232040405273, 0.030377824783325194, 0.030687648773193358, 0.027514944076538084, 0.02774233627319336, 0.027842048645019532, 0.027688671112060546, 0.027419424057006835, 0.027529119491577148, 0.028219615936279297, 0.02777280044555664, 0.02780771255493164, 0.027934751510620116, 0.02787139129638672, 0.027628639221191405, 0.02781216049194336, 0.028502336502075197, 0.03155548858642578, 0.028169471740722655, 0.028111839294433595, 0.028016639709472657, 0.02780364799499512, 0.028270591735839845, 0.02795635223388672, 0.027689855575561525, 0.027375423431396484, 0.027332000732421875, 0.02770800018310547, 0.02742291259765625, 0.027520959854125976, 0.027319488525390626, 0.027671424865722657, 0.02774393653869629, 0.027729856491088868, 0.027932159423828123, 0.027845279693603515, 0.02774825668334961, 0.027760896682739258, 0.027927711486816408, 0.027953216552734375, 0.027796512603759767, 0.027754304885864257, 0.028010496139526365, 0.027889535903930663, 0.027828351974487305, 0.029272064208984375, 0.02766582489013672, 0.02737609672546387, 0.027611263275146486, 0.027940383911132814, 0.027848255157470702, 0.027715808868408204, 0.027713632583618163, 0.027318111419677736, 0.027294303894042967, 0.027156639099121093, 0.027294784545898437, 0.027395008087158203, 0.027288896560668945, 0.02716089630126953, 0.027351551055908203, 0.02754547119140625, 0.027701248168945314, 0.02757529640197754, 0.027325439453125, 0.027793407440185547, 0.027276800155639647, 0.027162464141845703, 0.027128480911254884, 0.027149568557739256, 0.027185920715332032, 0.027277151107788087, 0.02727494430541992, 0.02713852882385254, 0.02766339111328125, 0.027360000610351563, 0.027189664840698242, 0.02715452766418457, 0.02749228858947754, 0.027162687301635742, 0.027177888870239256, 0.027146080017089843, 0.027343839645385743, 0.02711347198486328, 0.02713145637512207, 0.027199840545654295, 0.02720796775817871, 0.027146495819091798, 0.02705449676513672, 0.02712892723083496, 0.027619840621948243, 0.027969600677490235, 0.02736742401123047, 0.02759814453125, 0.027533824920654298, 0.027459264755249024, 0.027490816116333007, 0.027940927505493166, 0.02767967987060547, 0.02770227241516113, 0.027672576904296874, 0.027613407135009767, 0.02735424041748047, 0.02735171127319336, 0.027223039627075195, 0.02736387252807617, 0.027382240295410157, 0.027278944015502928, 0.027298208236694335, 0.02753126335144043, 0.0273670711517334, 0.027104864120483397, 0.027136768341064453, 0.02711756706237793, 0.027154560089111327, 0.027324287414550782, 0.027483232498168947, 0.02769977569580078, 0.027646303176879883, 0.027561695098876952, 0.0276911678314209, 0.027713567733764648, 0.027639904022216798, 0.027627519607543945, 0.028248064041137694, 0.027700704574584963, 0.02777756881713867, 0.02773811149597168, 0.027499839782714842, 0.027407039642333986, 0.02740019226074219, 0.027232255935668945, 0.027495967864990235, 0.027400672912597655, 0.027332256317138672, 0.02727462387084961, 0.02724265670776367, 0.027281824111938476, 0.027531328201293944, 0.027557247161865233, 0.02743087959289551, 0.02739852714538574, 0.027291999816894532, 0.027361568450927735, 0.027550975799560548, 0.027820543289184572, 0.027627904891967772, 0.027599903106689454, 0.034813377380371095, 0.030335424423217773, 0.02773414421081543, 0.02781119918823242, 0.027711328506469728, 0.02734160041809082, 0.02754560089111328, 0.02743427276611328, 0.02763644790649414, 0.027543552398681642, 0.027549440383911133, 0.02788582420349121, 0.027933792114257814, 0.02727619171142578, 0.027056127548217773, 0.02718720054626465, 0.027197439193725585, 0.027133504867553712, 0.027103263854980467, 0.027147743225097658, 0.02748044776916504, 0.027572799682617187, 0.02750464057922363, 0.0275599365234375, 0.027695104598999022, 0.027674207687377928, 0.027625759124755858, 0.027618431091308595, 0.027620256423950194, 0.027655679702758788, 0.02792508888244629, 0.027828479766845705, 0.027564863204956054, 0.027343807220458986, 0.02733670425415039, 0.027362943649291992, 0.027525503158569335, 0.027535360336303712, 0.027327999114990235, 0.027181215286254883, 0.0275579833984375, 0.02727961540222168, 0.02723744010925293, 0.02721683120727539, 0.027379711151123046, 0.027395103454589845, 0.02728611183166504, 0.027443584442138673, 0.027606399536132812, 0.027376256942749023, 0.027131904602050783, 0.027125759124755858, 0.027076608657836915, 0.027119583129882812, 0.02709712028503418, 0.0272523193359375, 0.027213472366333008, 0.027144479751586913, 0.02719340705871582, 0.02718351936340332, 0.027492416381835937, 0.027183040618896485, 0.02706211280822754, 0.027336671829223634, 0.027383968353271483, 0.027332447052001954, 0.027287551879882813, 0.02792428779602051, 0.027515071868896485, 0.027686912536621092, 0.027592704772949218, 0.02753126335144043, 0.027662336349487306, 0.02758438491821289, 0.027588544845581056, 0.02743724822998047, 0.027228160858154295, 0.02727529525756836, 0.027235519409179686, 0.027346847534179687, 0.027366336822509767, 0.027385791778564452, 0.02773302459716797, 0.027433855056762695, 0.027275360107421875, 0.027250688552856447, 0.02734284782409668, 0.0275283203125, 0.027458431243896485, 0.027546720504760744, 0.027558015823364257, 0.027820543289184572, 0.027795743942260743, 0.02776054382324219, 0.027705440521240233, 0.027852800369262694, 0.0276889591217041, 0.02778112030029297, 0.027780351638793947, 0.02780771255493164, 0.02813007926940918, 0.02800041580200195, 0.02783856010437012, 0.028280479431152344, 0.027625312805175783, 0.02767487907409668, 0.02786463928222656, 0.028055328369140625, 0.031793792724609374, 0.02823139190673828, 0.028168415069580077, 0.02781180763244629, 0.027799392700195314, 0.02796076774597168, 0.027663040161132812, 0.027586111068725584, 0.02747078323364258, 0.027651744842529295, 0.027834367752075196, 0.028092159271240234, 0.028800416946411132, 0.027644767761230468, 0.027522592544555663, 0.02751740837097168, 0.02742793655395508, 0.0274564151763916, 0.02793008041381836, 0.02778780746459961, 0.0278220157623291, 0.027701311111450196, 0.02904870414733887, 0.028043392181396485, 0.027730976104736328, 0.027683807373046876, 0.0272936954498291, 0.027297183990478514, 0.027711328506469728, 0.027443103790283203, 0.027470687866210937, 0.027410432815551757, 0.02812518310546875, 0.027258880615234377, 0.027310176849365233, 0.027154207229614258, 0.027158912658691407, 0.027177919387817384, 0.02721705627441406, 0.02722371292114258, 0.02735103988647461, 0.02773196792602539, 0.02755414390563965, 0.028116640090942384, 0.027380863189697267, 0.027170848846435548, 0.027225343704223633, 0.027170303344726563, 0.02710537528991699, 0.027242496490478517, 0.02706572723388672, 0.02694806480407715, 0.027130016326904295, 0.027149568557739256, 0.027129823684692383, 0.027196191787719728, 0.027133024215698243, 0.02716963195800781, 0.027162687301635742, 0.026989919662475586, 0.02699929618835449, 0.027343008041381837, 0.02727731132507324, 0.027391328811645507, 0.027247264862060548, 0.027373472213745118, 0.027136095046997072, 0.0273787841796875, 0.02712668800354004, 0.027084800720214845, 0.02708652877807617, 0.027224384307861327, 0.027248640060424805, 0.02735923194885254, 0.027203359603881837, 0.02729315185546875, 0.027294464111328125, 0.02727244758605957, 0.027327232360839844, 0.027354143142700196, 0.02742927932739258, 0.027840511322021484, 0.027897504806518553, 0.027805152893066405, 0.027687583923339844, 0.027649471282958984, 0.027597856521606446, 0.02747939109802246, 0.027280960083007812, 0.030124319076538085, 0.03112931251525879, 0.02784752082824707, 0.027588703155517577, 0.02722601509094238, 0.027422208786010743, 0.02721628761291504, 0.02721187210083008, 0.027197439193725585, 0.02751283264160156, 0.02720358467102051, 0.027296928405761717, 0.027429727554321288, 0.02767568016052246, 0.027529184341430663, 0.027431711196899414, 0.031690975189208985, 0.02743270492553711, 0.02753500747680664, 0.027303680419921875, 0.02726905632019043, 0.027169696807861327, 0.02729145622253418, 0.027166912078857422, 0.027543552398681642, 0.02718720054626465, 0.028479488372802734, 0.028022655487060545, 0.027353216171264648, 0.027271167755126953, 0.027215871810913086, 0.027328672409057616, 0.02718499183654785, 0.02714419174194336, 0.027172864913940428, 0.027190528869628906, 0.0272043514251709, 0.027256832122802735, 0.02752921676635742, 0.027655359268188476, 0.02774684715270996, 0.027746591567993164, 0.027996160507202147, 0.02766374397277832, 0.02758924865722656, 0.028657663345336915, 0.028096736907958983, 0.02772559928894043, 0.027613183975219727, 0.02737583923339844, 0.027414304733276367, 0.027465728759765624, 0.0279815673828125, 0.027876928329467775, 0.027451391220092772, 0.02740671920776367, 0.027331008911132812, 0.027336128234863283, 0.02713632011413574, 0.02720191955566406, 0.027546911239624022, 0.027682527542114258, 0.02745142364501953, 0.027564512252807618, 0.027494911193847657, 0.027660287857055665, 0.027553056716918944, 0.027698976516723633, 0.02774726486206055, 0.027615200042724608, 0.027624832153320313, 0.027830400466918946, 0.02755756759643555, 0.027253183364868164, 0.027168895721435545, 0.02723823928833008, 0.02756390380859375, 0.02743328094482422, 0.027273056030273437, 0.027347360610961914, 0.02727231979370117, 0.02729203224182129, 0.027277824401855468, 0.0272957763671875, 0.027232032775878906, 0.027162464141845703, 0.027207839965820314, 0.027607231140136718, 0.027578367233276366, 0.027451583862304688, 0.027394176483154297, 0.027282495498657227, 0.027128448486328127, 0.02734694480895996, 0.02758246421813965, 0.027125280380249025, 0.027115327835083008, 0.027160863876342773, 0.02726335906982422, 0.027624671936035156, 0.027613792419433594, 0.0276297607421875, 0.027611135482788086, 0.027596799850463868, 0.027711488723754882, 0.027881824493408203, 0.027810911178588867, 0.027800128936767577, 0.027838464736938476, 0.027711807250976564, 0.02753299140930176, 0.027613183975219727, 0.027762208938598633, 0.027736127853393553, 0.027658655166625978, 0.027432960510253908, 0.027457279205322267, 0.027402303695678712, 0.027676416397094727, 0.027568639755249022, 0.027568063735961913, 0.0278240966796875, 0.02770844841003418, 0.027755552291870118, 0.027766143798828125, 0.02780985641479492, 0.027813568115234374, 0.02781648063659668, 0.027860799789428712, 0.027939327239990236, 0.027774335861206055, 0.027816511154174803, 0.02783033561706543, 0.02767692756652832, 0.02761292839050293, 0.027522464752197266, 0.02746940803527832, 0.027827135086059572, 0.02785446357727051, 0.02772982406616211, 0.027635936737060548, 0.02741689682006836, 0.027635711669921875, 0.02748441505432129, 0.027465471267700194, 0.027414527893066407, 0.027418367385864256, 0.027381887435913087, 0.027361408233642578, 0.02761238479614258, 0.027884319305419923, 0.027799232482910156, 0.03082271957397461, 0.02790777587890625, 0.027733919143676757, 0.027654111862182616, 0.02759702491760254, 0.02754787254333496, 0.027490304946899413, 0.027479232788085936, 0.027601760864257814, 0.02737353515625, 0.02737766456604004, 0.028261856079101564, 0.027679136276245117, 0.027515008926391603, 0.027674591064453125, 0.027639423370361328, 0.027693632125854493, 0.027464960098266603, 0.02768070411682129, 0.027873952865600585, 0.02785420799255371, 0.027816576004028322, 0.027821855545043947, 0.02787743949890137, 0.027764896392822265, 0.027983871459960938, 0.028076032638549804, 0.027930784225463866, 0.027755680084228514, 0.028053951263427735, 0.02767638397216797, 0.02727507209777832, 0.027118047714233397, 0.027181055068969725, 0.027445119857788088, 0.02712384033203125, 0.027043840408325196, 0.027038751602172853, 0.0271759033203125, 0.02707571220397949, 0.02731916809082031, 0.027675872802734376, 0.027575071334838868, 0.027387136459350585, 0.027534080505371095, 0.027679840087890626, 0.02746460723876953, 0.02756233596801758, 0.027547296524047853, 0.027494400024414063, 0.027551744461059572, 0.027420352935791016, 0.027719999313354494, 0.027483455657958983, 0.02741881561279297, 0.027261215209960936, 0.0271889591217041, 0.027181568145751952, 0.02738092803955078, 0.027329599380493164, 0.027523839950561523, 0.027302783966064455, 0.027183040618896485, 0.02720502471923828, 0.027013919830322267, 0.027051584243774414, 0.027575904846191407, 0.02701705551147461, 0.027087936401367186, 0.027398080825805665, 0.02733670425415039, 0.027347232818603514, 0.027090015411376952, 0.02702195167541504, 0.027150655746459963, 0.027204832077026366, 0.027399967193603516, 0.02741881561279297, 0.027542015075683594, 0.02748579216003418, 0.027676864624023436, 0.027891199111938478, 0.027728607177734375, 0.027584512710571288, 0.027729471206665038, 0.027791807174682617, 0.027590112686157228, 0.0273306884765625, 0.027281824111938476, 0.027371519088745116, 0.027414751052856446, 0.027215072631835938]",tokens/s,36.26127753392732,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,6655.082496,7525.564416,0.0,7147.094016,7138.9184,s,1,11.4468486328125,11.4468486328125,0.0,11.4468486328125,11.4468486328125,11.4468486328125,11.4468486328125,[11.4468486328125],,kWh,0.00013003639331662573,1.4336647016102484e-05,4.1124477344001154e-05,0.00018549751767672935,,MB,1625.870336,8226.013184,0.0,7818.182656,7724.300288,s,10,7.336798767089843,0.7336798767089843,0.0035879936953573557,0.7331307373046876,0.738969970703125,0.7389776611328125,0.7389838134765625,"[0.7282850341796875, 0.7327047729492188, 0.7337333984375, 0.7297399291992187, 0.7321773071289063, 0.7389853515625, 0.7308643798828125, 0.7335567016601563, 0.7377836303710937, 0.73896826171875]",tokens/s,348.92602090753945,kWh,2.140277468333428e-05,2.3603525715006946e-06,1.4255666166428291e-05,3.801879342126327e-05,tokens/kWh,6733511.954559388,MB,1630.072832,8372.813824,0.0,7964.983296,7904.605696,s,10,33.90838354492187,3.3908383544921876,0.007679516201068551,3.3904208984375,3.3990643310546873,3.403143005371094,3.4064059448242188,"[3.3854130859375, 3.38075048828125, 3.381290283203125, 3.38770849609375, 3.39313330078125, 3.387211181640625, 3.393403564453125, 3.4072216796875, 3.398157958984375, 3.394093505859375]",tokens/s,18.579476050970552,kWh,9.931091062333278e-05,1.0954522001515274e-05,6.582875901217209e-05,0.00017609419163702012,tokens/kWh,357763.0778978832,,s,630,33.90527875900272,0.05381790279206775,0.0005654044508975231,0.05381379127502441,0.05440257415771484,0.05455391273498535,0.05594764579772949,"[0.05611299133300781, 0.053880992889404296, 0.05342822265625, 0.05351820755004883, 0.05314691162109375, 0.05317647933959961, 0.053160640716552736, 0.053153377532958984, 0.053032993316650394, 0.05321094512939453, 0.053161663055419923, 0.05333638381958008, 0.053318046569824216, 0.05323996734619141, 0.05317529678344726, 0.053384193420410155, 0.05373302459716797, 0.053844032287597654, 0.05401833724975586, 0.05368832015991211, 0.05341593551635742, 0.05325743865966797, 0.053365184783935544, 0.05333555221557617, 0.05295600128173828, 0.052822017669677736, 0.05313299179077149, 0.0533837776184082, 0.05339263916015625, 0.05397318267822265, 0.05323775863647461, 0.05382992172241211, 0.05375897598266602, 0.05351663970947266, 0.05375151824951172, 0.054241249084472656, 0.05422396850585937, 0.0548263053894043, 0.05408396911621094, 0.05415340805053711, 0.054077438354492184, 0.05498275375366211, 0.05415248107910156, 0.05370329666137695, 0.05385420989990235, 0.05383782577514649, 0.05377024078369141, 0.053787742614746094, 0.053738399505615236, 0.05371001434326172, 0.0536748161315918, 0.05411840057373047, 0.05400140762329102, 0.05425791931152344, 0.0541921272277832, 0.05413683319091797, 0.05404796981811524, 0.05425027084350586, 0.054077438354492184, 0.05399087905883789, 0.05386431884765625, 0.05376432037353516, 0.05376396942138672, 0.055836673736572265, 0.05372518539428711, 0.052795391082763675, 0.05276790237426758, 0.05287129592895508, 0.05279305648803711, 0.05271446228027344, 0.05332156753540039, 0.05257247924804687, 0.0529222412109375, 0.05278950500488281, 0.0530296630859375, 0.05296966552734375, 0.05299894332885742, 0.05333401489257812, 0.05296899032592774, 0.05343280029296875, 0.053663326263427735, 0.05406908798217774, 0.05383430480957031, 0.0536473617553711, 0.053155742645263675, 0.05300848007202148, 0.053200897216796876, 0.05327465438842773, 0.05329302215576172, 0.053430015563964844, 0.05338684844970703, 0.05336099243164062, 0.05326812744140625, 0.05316259384155273, 0.05310879898071289, 0.053166305541992184, 0.053227294921875, 0.05355091094970703, 0.05404691314697266, 0.05405491256713867, 0.05412659072875976, 0.053907455444335936, 0.054093822479248044, 0.05402828979492187, 0.05418188858032227, 0.05397734451293945, 0.05406623840332031, 0.05396960067749024, 0.054023361206054686, 0.05376412963867187, 0.05404956817626953, 0.05395014572143555, 0.05394464111328125, 0.05408963012695313, 0.054138271331787106, 0.05435385513305664, 0.05435878372192383, 0.05438259124755859, 0.05442764663696289, 0.054497344970703125, 0.054408321380615236, 0.05416182327270508, 0.05418844985961914, 0.054255615234375, 0.05422396850585937, 0.05409913635253906, 0.055831550598144535, 0.053561599731445315, 0.05307727813720703, 0.054111774444580076, 0.05269583892822265, 0.05293462371826172, 0.05297571182250976, 0.05296752166748047, 0.05295727920532227, 0.053079967498779294, 0.05301808166503906, 0.05309468841552734, 0.05310284805297852, 0.05460297775268555, 0.05323651123046875, 0.05348966217041016, 0.053905406951904294, 0.054212608337402345, 0.054451839447021484, 0.05402377700805664, 0.05388540649414063, 0.05383737564086914, 0.053424896240234374, 0.05322111892700195, 0.05330051040649414, 0.05329315185546875, 0.0535313606262207, 0.053633182525634766, 0.0534835205078125, 0.05326812744140625, 0.053388927459716795, 0.05364809417724609, 0.053495807647705076, 0.05328076934814453, 0.05363916778564453, 0.05374697494506836, 0.05408358383178711, 0.05425980758666992, 0.05374630355834961, 0.053645057678222655, 0.05356553649902344, 0.053387168884277345, 0.05344076919555664, 0.05347686386108398, 0.05352294540405273, 0.053575328826904293, 0.053776737213134765, 0.053507713317871096, 0.05383411026000977, 0.05371599960327148, 0.05346403121948242, 0.05394847869873047, 0.054136768341064456, 0.054083072662353515, 0.05426764678955078, 0.05389564895629883, 0.05394255828857422, 0.053927776336669925, 0.05400592041015625, 0.053778270721435546, 0.05367004776000977, 0.05450038528442383, 0.05442390441894531, 0.05584681701660156, 0.05393743896484375, 0.053304065704345704, 0.05313328170776367, 0.05301571273803711, 0.052882080078125, 0.05322108840942383, 0.0531578254699707, 0.053430049896240235, 0.05337961578369141, 0.053228897094726564, 0.053484416961669924, 0.053337150573730466, 0.053306304931640625, 0.05360220718383789, 0.053749439239501956, 0.05350883102416992, 0.054189918518066406, 0.0542674560546875, 0.05396918487548828, 0.05364115142822266, 0.05350003051757812, 0.05351212692260742, 0.053386302947998045, 0.05340665435791016, 0.05326633453369141, 0.05356083297729492, 0.05334451293945312, 0.05316198348999023, 0.0533342399597168, 0.05373142242431641, 0.05348150253295898, 0.05347731018066406, 0.053620800018310544, 0.0542658576965332, 0.054178878784179686, 0.05430163192749023, 0.054405311584472656, 0.054198078155517575, 0.05398732757568359, 0.05390335845947265, 0.05380300903320313, 0.05375385665893555, 0.05384969711303711, 0.05397315216064453, 0.05376230239868164, 0.053687328338623046, 0.05358396911621094, 0.05391024017333984, 0.053928001403808594, 0.05402550506591797, 0.05411923217773437, 0.05403363037109375, 0.05426051330566406, 0.05432729721069336, 0.05399552154541016, 0.05423513412475586, 0.05408768081665039, 0.05405641555786133, 0.05410460662841797, 0.05429452896118164, 0.05409584045410156, 0.05392585754394531, 0.05595532989501953, 0.05365760040283203, 0.05325241470336914, 0.05307769775390625, 0.05297119903564453, 0.05309001541137695, 0.05329062271118164, 0.05300118255615234, 0.05289100646972656, 0.05306022262573242, 0.05335039901733398, 0.05313324737548828, 0.053112895965576175, 0.05339302444458008, 0.05358537673950195, 0.05384284973144531, 0.053512191772460936, 0.05361772918701172, 0.05440198516845703, 0.05422444915771484, 0.053959102630615235, 0.05382963180541992, 0.0539521598815918, 0.05363542556762695, 0.053628929138183595, 0.05368012619018555, 0.053567134857177734, 0.05324835205078125, 0.053337791442871096, 0.05337436676025391, 0.05357046508789062, 0.05360416030883789, 0.05379296112060547, 0.053688385009765624, 0.053813182830810546, 0.0538524169921875, 0.05434771347045898, 0.05513353729248047, 0.054058910369873044, 0.05408415985107422, 0.05431267166137695, 0.054265792846679685, 0.05405926513671875, 0.05396694564819336, 0.054195873260498045, 0.054978240966796876, 0.05427676773071289, 0.05438172912597656, 0.05418684768676758, 0.054079486846923826, 0.05403209686279297, 0.054230655670166016, 0.05409804916381836, 0.054419361114501956, 0.05455500793457031, 0.05400191879272461, 0.05412163162231445, 0.05437116622924805, 0.05419375991821289, 0.05435193634033203, 0.053719390869140626, 0.05370032119750977, 0.053782817840576175, 0.055956001281738284, 0.05390739059448242, 0.05332489776611328, 0.05314236831665039, 0.05325727844238281, 0.052891647338867184, 0.05310976028442383, 0.05306959915161133, 0.05299836730957031, 0.05306163024902344, 0.053174270629882815, 0.05324595260620117, 0.05304441452026367, 0.05298204803466797, 0.05297411346435547, 0.05347430419921875, 0.05337395095825195, 0.05447270584106445, 0.05457715225219727, 0.054042625427246097, 0.053710369110107424, 0.05345283126831055, 0.05335452651977539, 0.053432735443115234, 0.0535327033996582, 0.053563358306884766, 0.053753215789794924, 0.05339152145385742, 0.053553184509277346, 0.05332012939453125, 0.05380028915405274, 0.05324652862548828, 0.05332592010498047, 0.053901409149169924, 0.054029247283935544, 0.054130687713623046, 0.05436038589477539, 0.05435433578491211, 0.05405721664428711, 0.05420851135253906, 0.05392895889282227, 0.05381836700439453, 0.05380201721191406, 0.053691360473632814, 0.053556350708007815, 0.05367184066772461, 0.05392892837524414, 0.05383168029785156, 0.05415321731567383, 0.05424332809448242, 0.05381289672851562, 0.05360838317871094, 0.054429534912109376, 0.05429043197631836, 0.05421318435668945, 0.05398518371582031, 0.054074623107910155, 0.054131553649902346, 0.054091102600097654, 0.054356639862060546, 0.05427571105957031, 0.0542558708190918, 0.054102142333984374, 0.0559288330078125, 0.05385820770263672, 0.053222721099853515, 0.05309491348266602, 0.05277315139770508, 0.052853824615478516, 0.05314646530151367, 0.053607872009277344, 0.05389583969116211, 0.05325551986694336, 0.05311929702758789, 0.052996383666992185, 0.05317228698730469, 0.053359966278076175, 0.053523265838623046, 0.05382332611083984, 0.05400105667114258, 0.0544134407043457, 0.054436321258544924, 0.05422694396972656, 0.05398463821411133, 0.05367078399658203, 0.05357900619506836, 0.053363201141357425, 0.05368368148803711, 0.05378473663330078, 0.053763774871826174, 0.05398835372924805, 0.053657279968261716, 0.053348575592041016, 0.053617984771728515, 0.053827808380126956, 0.053712478637695314, 0.05395654296875, 0.054046497344970704, 0.0542171516418457, 0.054086143493652344, 0.054306880950927734, 0.05415020751953125, 0.05387887954711914, 0.05382774353027344, 0.053666431427001955, 0.053652767181396485, 0.053768287658691405, 0.054088321685791016, 0.05367327880859375, 0.053578304290771483, 0.05386419296264648, 0.05371244812011719, 0.053623809814453124, 0.05387795257568359, 0.0541927375793457, 0.05445840072631836, 0.05429673767089844, 0.05448483276367187, 0.054335487365722655, 0.05440499114990234, 0.05456908798217774, 0.05461196899414063, 0.05432499313354492, 0.05426131057739258, 0.054206336975097656, 0.05426262283325195, 0.05608272171020508, 0.05599296188354492, 0.05338447952270508, 0.05328287887573242, 0.05346345520019531, 0.05339775848388672, 0.05349548721313477, 0.05371526336669922, 0.05325766372680664, 0.05372518539428711, 0.05346566390991211, 0.053300769805908206, 0.05323209762573242, 0.05357920074462891, 0.05373001480102539, 0.05406067276000977, 0.05382908630371094, 0.05459020614624024, 0.05444384002685547, 0.05385388946533203, 0.0535577278137207, 0.053620704650878905, 0.05355116653442383, 0.05370470428466797, 0.053664768218994144, 0.053482177734375, 0.053571903228759765, 0.05354214477539063, 0.05342284774780273, 0.05344412612915039, 0.05364579010009766, 0.053830848693847654, 0.053949249267578124, 0.054046718597412106, 0.05433731079101563, 0.054542560577392575, 0.05435391998291016, 0.05442559814453125, 0.05430025482177735, 0.05418841552734375, 0.05392508697509766, 0.05410079956054688, 0.053975040435791016, 0.05392998504638672, 0.05402828979492187, 0.054220703125, 0.054114398956298826, 0.05429452896118164, 0.05429270553588867, 0.054335262298583986, 0.05560054397583008, 0.054376609802246095, 0.05448704147338867, 0.05459807968139648, 0.055027233123779294, 0.05473459243774414, 0.05448751831054687, 0.054649089813232424, 0.055657951354980466, 0.05415990447998047, 0.054368160247802735, 0.054275775909423826, 0.054171871185302735, 0.05671475219726563, 0.053736000061035155, 0.05292230224609375, 0.0531701774597168, 0.05304729461669922, 0.053429439544677736, 0.05330940628051758, 0.05310345458984375, 0.053303295135498044, 0.05359779357910156, 0.05357968139648438, 0.053451263427734375, 0.05326182556152344, 0.05356390380859375, 0.054216896057128906, 0.05382486343383789, 0.05388336181640625, 0.054468608856201174, 0.05472489547729492, 0.0543639030456543, 0.05397705459594727, 0.0535654411315918, 0.0543559684753418, 0.053370880126953124, 0.05328464126586914, 0.053714942932128903, 0.05349964904785156, 0.05340822219848633, 0.05331087875366211, 0.0533551025390625, 0.05355491256713867, 0.053575969696044924, 0.05349286270141602, 0.054012767791748045, 0.05397711944580078, 0.05437772750854492, 0.054702049255371095, 0.05416812896728516, 0.05410220718383789, 0.05417372894287109, 0.054091392517089845, 0.05392012786865234, 0.05402995300292969, 0.05370076751708985, 0.05391996765136719, 0.053975040435791016, 0.05414297485351562, 0.053800960540771485, 0.054056129455566405, 0.053881664276123044, 0.05406105422973633, 0.054152671813964846, 0.05448252868652344, 0.05434771347045898, 0.05452012634277344, 0.05445292663574219, 0.054263168334960935, 0.05446025466918945, 0.054399871826171876, 0.05451715087890625, 0.05441558456420898, 0.05436419296264648, 0.05425356674194336, 0.05651865768432617, 0.053999359130859376, 0.053321983337402346, 0.05293670272827149, 0.05309030532836914, 0.05311283111572265, 0.05294694519042969, 0.053191745758056644, 0.052975616455078124, 0.05312156677246094, 0.05312067031860351, 0.05310950469970703, 0.05310796737670898, 0.053063934326171874, 0.053254657745361325, 0.05329084777832031, 0.053418048858642576, 0.0542242546081543, 0.05418819046020508, 0.054059585571289065, 0.054093822479248044, 0.053645313262939455, 0.053399742126464846, 0.053294910430908206, 0.053198848724365234, 0.05324390411376953, 0.05440230560302734, 0.05384236907958984, 0.05368454360961914, 0.05381439971923828, 0.05382819366455078, 0.05372041702270508, 0.0536945915222168, 0.054946624755859375, 0.05453823852539062, 0.054254657745361326, 0.054182910919189455, 0.05414495849609375, 0.05451776123046875, 0.054128929138183596, 0.05431881713867188, 0.05413849639892578, 0.05387712097167969, 0.053833728790283204, 0.053956382751464846, 0.054204193115234375, 0.05396688079833984, 0.05393654251098633, 0.05376409530639648, 0.053700607299804685, 0.05428220748901367, 0.05416988754272461, 0.05454207992553711, 0.05455257415771484, 0.05427609634399414, 0.05433865737915039, 0.054559070587158205, 0.05477024078369141, 0.05437971115112305, 0.053898048400878903, 0.053771488189697264, 0.053992225646972654, 0.053945343017578126]",tokens/s,18.581177417180783,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,11033.808896,12225.282048,0.0,11846.811648,11814.785024,s,1,14.5446103515625,14.5446103515625,0.0,14.5446103515625,14.5446103515625,14.5446103515625,14.5446103515625,[14.5446103515625],,kWh,0.00021487755945832606,2.3695210668930488e-05,6.834561023200125e-05,0.0003069183803592578,,MB,2022.076416,13158.514688,0.0,12750.68416,12641.86368,s,10,13.2707529296875,1.32707529296875,0.004613391175216123,1.3276134033203126,1.3325015625,1.3327134887695313,1.3328830297851564,"[1.3182591552734375, 1.320686279296875, 1.3242823486328126, 1.3263507080078125, 1.327857666015625, 1.327369140625, 1.3311546630859374, 1.3294130859375, 1.3324544677734376, 1.3329254150390626]",tokens/s,192.90540736939806,kWh,3.879431871291899e-05,4.277520867026141e-06,2.583313177760005e-05,6.890497135754518e-05,tokens/kWh,3715261.6851348225,MB,2031.058944,13452.115968,0.0,13044.28544,12933.698048,s,10,51.7908515625,5.17908515625,0.0033790050210793367,5.179365478515625,5.182355322265625,5.183464282226563,5.1843514501953125,"[5.17328857421875, 5.17350244140625, 5.180556640625, 5.18149169921875, 5.17875244140625, 5.1845732421875, 5.17892431640625, 5.179806640625, 5.18210888671875, 5.1778466796875]",tokens/s,12.164310510317263,kWh,0.00015175458123291267,1.6740703577205495e-05,0.00010082966399700065,0.0002693249488071188,tokens/kWh,233918.17311777684,,s,630,51.7864017944336,0.08220063776894221,0.0006728557495673679,0.08218268966674805,0.0829502426147461,0.08315069580078124,0.08478962242126466,"[0.08456169891357422, 0.08118694305419921, 0.0809335708618164, 0.08112547302246094, 0.081331298828125, 0.08113040161132813, 0.08157324981689452, 0.08140009307861328, 0.08107599639892578, 0.08123206329345703, 0.08104338836669922, 0.08238925170898438, 0.08226630401611328, 0.08149759674072266, 0.08128543853759766, 0.08151007843017578, 0.08170700836181641, 0.08183654022216796, 0.08158793640136719, 0.08168067169189454, 0.0815288314819336, 0.08181542205810546, 0.08185049438476563, 0.08251590728759765, 0.08246041870117188, 0.08226956939697265, 0.0819352035522461, 0.0818644790649414, 0.08150169372558594, 0.08195273590087891, 0.08182870483398437, 0.08192924499511718, 0.08211759948730468, 0.08198143768310547, 0.08286790466308594, 0.0826289291381836, 0.08233369445800781, 0.08279654693603515, 0.08241356658935547, 0.08198934173583984, 0.0816701431274414, 0.08215580749511718, 0.08205913543701172, 0.08215347290039063, 0.08213497924804687, 0.0822458267211914, 0.08282112121582032, 0.08289222717285157, 0.08270697784423828, 0.08245629119873046, 0.08238111877441406, 0.08225993347167969, 0.08217552185058594, 0.08244271850585938, 0.08250736236572266, 0.08216006469726563, 0.0827734375, 0.08324147033691406, 0.08261449432373047, 0.08304009246826172, 0.08304029083251953, 0.08295011138916016, 0.08299343872070312, 0.08484451293945312, 0.08114192199707031, 0.08077705383300782, 0.08111411285400391, 0.08118521881103516, 0.0814986572265625, 0.08141417694091797, 0.0813786849975586, 0.08126643371582032, 0.08130035400390626, 0.08130355072021485, 0.08237670135498047, 0.08220262145996093, 0.08201824188232422, 0.081417724609375, 0.0818304672241211, 0.08153199768066406, 0.08151952362060547, 0.08189734649658204, 0.0816388168334961, 0.08149884796142579, 0.08167628479003906, 0.08162051391601563, 0.0826024627685547, 0.08236838531494141, 0.08186688232421875, 0.08182988739013672, 0.08182755279541015, 0.081580322265625, 0.08187699127197266, 0.08182374572753906, 0.08189520263671875, 0.08232777404785156, 0.08200396728515626, 0.08189542388916016, 0.08255795288085938, 0.08274813079833984, 0.08241728210449219, 0.08252073669433593, 0.08244429016113282, 0.08223305511474609, 0.08203907012939453, 0.08229478454589843, 0.08244124603271484, 0.08242063903808594, 0.08230944061279297, 0.08251980590820313, 0.08273824310302734, 0.08283232116699218, 0.08245862579345703, 0.08251779174804688, 0.08242825317382813, 0.08244937896728516, 0.08250790405273438, 0.0821987533569336, 0.08248582458496094, 0.08269414520263672, 0.08270662689208984, 0.08253187561035157, 0.08252444458007813, 0.08313037109375, 0.08298249816894532, 0.08253817749023437, 0.08465523529052735, 0.08204377746582031, 0.08077507019042969, 0.08093705749511719, 0.08088124847412109, 0.08112374114990234, 0.08086287689208985, 0.08139571380615235, 0.08155990600585937, 0.08210431671142578, 0.08147353363037109, 0.08223856353759766, 0.08271878051757812, 0.08185123443603516, 0.08148588562011719, 0.08150579071044922, 0.08186720275878906, 0.08174387359619141, 0.08148377227783203, 0.08144895935058594, 0.08198054504394531, 0.08347843170166015, 0.08198652648925782, 0.08210173034667968, 0.08278275299072266, 0.08242176055908203, 0.08164886474609374, 0.08170909118652343, 0.08212274932861328, 0.08173158264160156, 0.0817174072265625, 0.08179158020019531, 0.08216575622558593, 0.08243814086914063, 0.08205020904541016, 0.08352239990234375, 0.0828579864501953, 0.08253798675537109, 0.0835076446533203, 0.08181929779052734, 0.0818017578125, 0.08212057495117188, 0.08244217681884766, 0.08219638061523438, 0.0827455062866211, 0.08250342559814453, 0.0825631332397461, 0.08255302429199218, 0.08273625946044921, 0.08281568145751952, 0.08270246124267579, 0.08237267303466797, 0.08246886444091797, 0.08223334503173828, 0.0822518081665039, 0.08298697662353516, 0.08249523162841797, 0.08284515380859375, 0.08301849365234375, 0.08254847717285156, 0.08295862579345703, 0.08314675140380859, 0.08303945922851562, 0.08564070129394531, 0.0817197723388672, 0.08130928039550782, 0.08110041809082032, 0.08120531463623047, 0.08152451324462891, 0.08156047821044922, 0.08136252593994141, 0.08132653045654296, 0.08151449584960938, 0.08196915435791016, 0.08253440093994141, 0.08262860870361328, 0.0821363525390625, 0.08163555145263672, 0.08159190368652344, 0.08138758087158203, 0.08176726531982421, 0.0819603500366211, 0.08150486755371093, 0.08162268829345704, 0.08195516967773438, 0.08213206481933594, 0.08279952239990235, 0.08348262023925782, 0.08202854156494141, 0.08175926208496094, 0.081602783203125, 0.08274969482421875, 0.08165769958496094, 0.08208860778808594, 0.081829345703125, 0.08170317077636718, 0.08169296264648437, 0.08224358367919922, 0.08291709136962891, 0.08272697448730469, 0.08272509002685546, 0.08242166137695313, 0.08204502105712891, 0.08237260437011719, 0.08187289428710938, 0.08218214416503906, 0.08293990325927734, 0.08205862426757812, 0.08225247955322265, 0.08280262756347656, 0.08320800018310547, 0.08283891296386718, 0.08276051330566406, 0.08260995483398438, 0.08249366760253907, 0.08250163269042969, 0.08198925018310547, 0.08207807922363282, 0.08246451568603516, 0.0825483169555664, 0.08261698913574218, 0.08277986907958984, 0.08313065338134766, 0.08315494537353516, 0.08300051116943359, 0.08282943725585938, 0.08430976104736328, 0.08145331573486328, 0.08131075286865234, 0.08116937255859374, 0.08137145233154297, 0.08132563018798829, 0.08124758148193359, 0.08143456268310546, 0.08127369689941406, 0.08173363494873047, 0.08184422302246094, 0.08260403442382812, 0.08267916870117188, 0.08202454376220703, 0.08160924530029297, 0.08175564575195313, 0.08200863647460938, 0.08173152160644531, 0.0817623062133789, 0.08203395080566406, 0.08158806610107422, 0.08210521697998047, 0.08193638610839844, 0.08247673797607422, 0.08239750671386718, 0.08206540679931641, 0.08244818878173828, 0.08202873229980469, 0.08204902648925781, 0.08206902313232421, 0.08169500732421875, 0.0817476806640625, 0.08174639892578126, 0.08196505737304688, 0.08234544372558594, 0.08310838317871094, 0.08235401916503907, 0.0821126708984375, 0.08207360076904296, 0.08216925048828125, 0.08241359710693359, 0.08243462371826171, 0.08224931335449219, 0.08221737670898438, 0.08222310638427735, 0.08236441802978516, 0.08252210998535156, 0.08313037109375, 0.08240850830078125, 0.08220767974853516, 0.08265318298339844, 0.08250572967529297, 0.08241356658935547, 0.08283545684814453, 0.08260179138183593, 0.0823154525756836, 0.08251126098632812, 0.08239369964599609, 0.08302950286865235, 0.08290560150146484, 0.08275750732421874, 0.0829024658203125, 0.08312818908691406, 0.08488285064697265, 0.08145967864990235, 0.08114112091064453, 0.08113228607177735, 0.08113740539550782, 0.08137583923339843, 0.08147727966308593, 0.08218624114990235, 0.08142633819580078, 0.08161698913574218, 0.08174092864990234, 0.08263542175292969, 0.08244175720214844, 0.08150640106201172, 0.08192060852050781, 0.08129955291748046, 0.08168418884277344, 0.08160176086425781, 0.08171209716796875, 0.0817786865234375, 0.08165769958496094, 0.08188915252685547, 0.08218038177490235, 0.08271257781982422, 0.08263680267333984, 0.08208726501464844, 0.0823340835571289, 0.08169049835205078, 0.08273347473144531, 0.08163123321533203, 0.08192819213867188, 0.08212070465087891, 0.08179507446289062, 0.08164761352539063, 0.08220877075195313, 0.08296428680419922, 0.08287660980224609, 0.08283055877685547, 0.08254544067382813, 0.08378982543945312, 0.08177664184570313, 0.08222310638427735, 0.0823334732055664, 0.08236377716064452, 0.0832807388305664, 0.08240284729003906, 0.08249945831298829, 0.08280534362792968, 0.08390799713134765, 0.08256540679931641, 0.08284976196289062, 0.08248953247070312, 0.08232489776611328, 0.08253311920166016, 0.08247500610351563, 0.08255257415771484, 0.08242201232910157, 0.08299110412597656, 0.0827166748046875, 0.08308700561523437, 0.08287785339355469, 0.08322870635986328, 0.08301046752929687, 0.08494649505615234, 0.08111558532714844, 0.08106393432617187, 0.08129535675048828, 0.08134630584716797, 0.08140585327148438, 0.08137503814697265, 0.08191849517822265, 0.08171520233154297, 0.08159366607666016, 0.08136978912353515, 0.082766845703125, 0.08230726623535156, 0.08142665863037109, 0.0815785903930664, 0.08183715057373046, 0.08228752136230469, 0.08149811553955078, 0.08172748565673828, 0.08190086364746094, 0.08189730834960937, 0.08181539154052735, 0.08202957153320313, 0.08205449676513672, 0.0823539810180664, 0.08214611053466797, 0.08145513916015625, 0.08202649688720703, 0.08159404754638672, 0.08161721801757812, 0.08208589172363281, 0.08199922943115234, 0.08182848358154297, 0.08205721282958985, 0.0823595199584961, 0.08224553680419922, 0.08246975708007813, 0.08216371154785156, 0.08232959747314453, 0.08229388427734376, 0.08196153259277343, 0.08201859283447266, 0.08204070281982422, 0.0823493423461914, 0.0825107192993164, 0.08253833770751953, 0.08293007659912109, 0.08293555450439453, 0.08276188659667968, 0.08380598449707032, 0.08232262420654297, 0.08286707305908203, 0.08302591705322265, 0.08254054260253907, 0.08250367736816407, 0.0827448959350586, 0.08258019256591796, 0.08300240325927734, 0.08283837127685546, 0.0827042236328125, 0.08281855773925781, 0.08268569946289063, 0.08271334075927735, 0.08514598083496094, 0.08145766448974609, 0.08181705474853515, 0.08107881927490235, 0.0809881591796875, 0.08148953247070312, 0.08115238189697266, 0.08198553466796875, 0.08171056365966797, 0.08157647705078125, 0.08152012634277343, 0.08245024108886718, 0.08248390197753906, 0.08180518341064454, 0.08127910614013673, 0.0811167984008789, 0.081795166015625, 0.08169606781005859, 0.08203772735595703, 0.08164886474609374, 0.08187369537353516, 0.08174787139892578, 0.08213085174560547, 0.08208198547363281, 0.08276985931396484, 0.0824730224609375, 0.08204287719726562, 0.08161849975585937, 0.08148831939697265, 0.08260835266113281, 0.08229456329345704, 0.0817413101196289, 0.08151062774658203, 0.08187522888183593, 0.08218550109863282, 0.08298159790039063, 0.08280678558349609, 0.0828960952758789, 0.08226448059082031, 0.08251216125488281, 0.08187709045410156, 0.0819768295288086, 0.08228486633300781, 0.08198572540283203, 0.08177565002441406, 0.08192921447753906, 0.08257532501220703, 0.08283340454101562, 0.08325939178466797, 0.08313446044921875, 0.08275353240966797, 0.08253254699707031, 0.08250962829589843, 0.08279417419433593, 0.0825367660522461, 0.08274739074707031, 0.08273715209960937, 0.08243814086914063, 0.08271459197998046, 0.08297270202636718, 0.08304434967041016, 0.08290509033203125, 0.08289881896972656, 0.08501558685302735, 0.08155235290527343, 0.08131123352050781, 0.08128972625732422, 0.08149942779541015, 0.08159101104736328, 0.08147763061523437, 0.08116944122314453, 0.08162989044189453, 0.08223977661132813, 0.08157727813720703, 0.08225247955322265, 0.08317942047119141, 0.08195696258544923, 0.08169446563720703, 0.08156963348388672, 0.08179737854003906, 0.08172354888916016, 0.0816332778930664, 0.08131584167480468, 0.08195481872558594, 0.08195891571044922, 0.08179859161376953, 0.08230729675292969, 0.08319747161865235, 0.08211743927001953, 0.08221638488769531, 0.08191238403320313, 0.08189952087402344, 0.08166339111328125, 0.08199858856201171, 0.08172322845458985, 0.08180531311035157, 0.08196812438964844, 0.08308617401123047, 0.08262159729003907, 0.08290201568603515, 0.08241129302978516, 0.08182009887695313, 0.08171907043457032, 0.0817305908203125, 0.08220976257324218, 0.08226988983154297, 0.08207520294189453, 0.0821456298828125, 0.08252867126464844, 0.08285779571533203, 0.08265734100341797, 0.08322265625, 0.08275107574462891, 0.0833653793334961, 0.08231005096435547, 0.08253145599365234, 0.08270323181152343, 0.08246476745605469, 0.08242585754394531, 0.08281702423095703, 0.08272589111328126, 0.08315392303466797, 0.08317250823974609, 0.08328278350830078, 0.08295142364501953, 0.08283417510986328, 0.08503270721435546, 0.08157587432861328, 0.08065376281738282, 0.08089485168457031, 0.08105939483642578, 0.08132621002197266, 0.081498046875, 0.08180770874023438, 0.08136502075195312, 0.08162918090820312, 0.08191795349121093, 0.08256092834472656, 0.08274931335449219, 0.08222509002685546, 0.08142671966552735, 0.08163533020019531, 0.08132608032226563, 0.08155296325683593, 0.08157341003417969, 0.08151856231689453, 0.08160966491699219, 0.08197119903564454, 0.08218323516845703, 0.08241862487792968, 0.08223919677734375, 0.08195475006103516, 0.08160460662841797, 0.08144512176513671, 0.08178083038330078, 0.08185855865478515, 0.08167584228515624, 0.08186515045166015, 0.08194595336914062, 0.08188118743896484, 0.08228691101074219, 0.08284540557861328, 0.08266973114013672, 0.08301376342773438, 0.08234175872802735, 0.08224601745605468, 0.08197119903564454, 0.08232575988769532, 0.0823325424194336, 0.08237667083740234, 0.08240630340576172, 0.08217362976074219, 0.08256336212158204, 0.08278828430175782, 0.08294111633300781, 0.08277289581298829, 0.08269004821777344, 0.0824393310546875, 0.08260694122314453, 0.08225318145751953, 0.08226265716552734, 0.08320956420898437, 0.08234358215332031, 0.08263577270507813, 0.08284569549560547, 0.08293593597412109, 0.08332028961181641, 0.08305500793457031, 0.08299858856201171]",tokens/s,12.165355733746255,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.747328,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3546.959872,4490.985472,0.0,4112.515072,3975.832064,s,1,9.7631337890625,9.7631337890625,0.0,9.7631337890625,9.7631337890625,9.7631337890625,9.7631337890625,[9.7631337890625],,kWh,8.129022722916588e-05,8.9596226684968e-06,2.642113224800009e-05,0.00011667098214566276,,MB,1479.933952,4688.11776,0.0,4280.287232,4101.544448,s,10,3.15926431274414,0.315926431274414,0.0010109459895497204,0.31538465881347655,0.31761447753906247,0.3176444427490234,0.3176684149169922,"[0.315052734375, 0.3151170043945313, 0.3151600952148437, 0.315346923828125, 0.317018798828125, 0.3176744079589844, 0.3156051330566406, 0.31542239379882814, 0.31525900268554685, 0.3176078186035156]",tokens/s,810.3152337312296,kWh,9.566036410752691e-06,1.0549664938000431e-06,6.345775685935489e-06,1.696677859048822e-05,tokens/kWh,15088309.111519653,MB,1483.948032,4698.60352,0.0,4290.772992,4101.547008,s,10,26.534383544921877,2.6534383544921876,0.00979076944631834,2.656810791015625,2.6639314453125,2.6643930908203126,2.6647624072265628,"[2.65593896484375, 2.64122021484375, 2.6576826171875, 2.63843017578125, 2.662216064453125, 2.653052978515625, 2.664854736328125, 2.638652587890625, 2.663828857421875, 2.65850634765625]",tokens/s,23.742778833863987,kWh,7.725948908341414e-05,8.521821049239766e-06,4.6889766902464504e-05,0.0001326710770351184,tokens/kWh,474858.58566840255,,s,630,26.531919879913332,0.04211415853954497,0.0007032065951850768,0.0420089111328125,0.04267241172790527,0.043027694702148435,0.045596519966125494,"[0.04278915023803711, 0.04206681442260742, 0.04215881729125977, 0.041765056610107425, 0.04200646209716797, 0.041768959045410156, 0.04157968139648437, 0.04172854232788086, 0.04170783996582031, 0.042733089447021484, 0.042060062408447264, 0.04209683227539063, 0.04245651245117187, 0.04193337631225586, 0.04279417419433594, 0.0428337287902832, 0.04258838272094727, 0.04172470474243164, 0.041697025299072266, 0.041691390991210935, 0.04191823959350586, 0.04171798324584961, 0.041616737365722654, 0.04158531188964844, 0.041543678283691404, 0.041744384765625, 0.041984001159667966, 0.04177920150756836, 0.04248076629638672, 0.04215283203125, 0.04303833770751953, 0.042307647705078125, 0.04219731140136719, 0.04278451156616211, 0.0425945930480957, 0.04255292892456055, 0.04247916793823242, 0.04233299255371094, 0.04220927810668945, 0.04241804885864258, 0.04294259262084961, 0.042123264312744144, 0.04210073471069336, 0.04207382583618164, 0.042313823699951174, 0.042172672271728516, 0.041960609436035155, 0.042229759216308595, 0.04238620758056641, 0.04293632125854492, 0.04228300857543945, 0.04245491027832031, 0.04188787078857422, 0.041836544036865236, 0.04183244705200195, 0.042196990966796875, 0.04196761703491211, 0.042493694305419924, 0.041865470886230466, 0.04200815963745117, 0.042098209381103514, 0.04193574523925781, 0.041981952667236325, 0.043318336486816406, 0.04226144027709961, 0.042842113494873046, 0.04219084930419922, 0.04202905654907227, 0.041850879669189454, 0.04170751953125, 0.041883647918701174, 0.042004417419433594, 0.04399929428100586, 0.04259340667724609, 0.04164412689208984, 0.04230390548706055, 0.042022911071777344, 0.04160345458984375, 0.04134108734130859, 0.04156156921386719, 0.04254348754882813, 0.04160038375854492, 0.04141120147705078, 0.041560062408447264, 0.04135696029663086, 0.04141091156005859, 0.041312255859375, 0.04151910400390625, 0.041272544860839845, 0.041685791015625, 0.04166656112670898, 0.041419998168945316, 0.04153408050537109, 0.041429153442382814, 0.04143299102783203, 0.04235651016235351, 0.04188726425170899, 0.04252751922607422, 0.041404415130615234, 0.041670654296875, 0.04206585693359375, 0.04196307373046875, 0.04193468856811523, 0.04189977645874023, 0.042740638732910154, 0.041957279205322266, 0.042109024047851565, 0.04217647933959961, 0.041842655181884764, 0.041682048797607424, 0.041683902740478514, 0.04154560089111328, 0.04179776000976562, 0.04188147354125977, 0.0419117431640625, 0.042050113677978514, 0.041807456970214846, 0.04187804794311523, 0.042205184936523435, 0.041928382873535154, 0.04147004699707031, 0.042215648651123046, 0.041711616516113284, 0.042032192230224606, 0.04232233428955078, 0.04202665710449219, 0.04261273574829102, 0.04176668930053711, 0.04155619049072266, 0.041472000122070314, 0.04195052719116211, 0.04135977554321289, 0.04144319915771484, 0.04129548645019531, 0.04130607986450195, 0.041997150421142576, 0.04146726226806641, 0.04385830307006836, 0.04354790496826172, 0.04158768081665039, 0.04199161529541016, 0.041551422119140625, 0.04158566284179688, 0.0418485107421875, 0.04220489501953125, 0.042151519775390625, 0.042038272857666016, 0.04165222549438476, 0.041688095092773436, 0.041618400573730466, 0.04149452972412109, 0.04158591842651367, 0.0415030403137207, 0.04190579223632813, 0.041874496459960935, 0.04324121475219726, 0.04225843048095703, 0.04231167984008789, 0.0451932144165039, 0.044146015167236326, 0.04240991973876953, 0.04321148681640625, 0.04193075180053711, 0.04225228881835937, 0.04235171127319336, 0.04194393539428711, 0.04182422256469726, 0.04172969436645508, 0.04223235321044922, 0.04216726303100586, 0.041847713470458986, 0.041799678802490234, 0.0425984001159668, 0.04170492935180664, 0.04182684707641601, 0.04183564758300781, 0.042019710540771485, 0.042712158203125, 0.042062366485595706, 0.041997726440429685, 0.04227139282226562, 0.04191648101806641, 0.041974014282226565, 0.04225609588623047, 0.04277008056640625, 0.04225497436523438, 0.04223385620117188, 0.045540447235107424, 0.04266662216186524, 0.04313587188720703, 0.04221952056884765, 0.041777153015136716, 0.041404415130615234, 0.04160710525512695, 0.0414035530090332, 0.041522079467773435, 0.04151295852661133, 0.041809921264648435, 0.04171139144897461, 0.04170774459838867, 0.04159718322753906, 0.04234419250488281, 0.04123209762573242, 0.04177129745483398, 0.041828353881835936, 0.04153097534179687, 0.04159280014038086, 0.04189046478271485, 0.04231964874267578, 0.0415968017578125, 0.04160019302368164, 0.041807903289794925, 0.041540512084960936, 0.0418939208984375, 0.04157350540161133, 0.04244259262084961, 0.04160742568969727, 0.04168576049804688, 0.04177305603027344, 0.04160841751098633, 0.041812767028808595, 0.04204339218139649, 0.04207846450805664, 0.04221011352539063, 0.04168307113647461, 0.04265011215209961, 0.04200019073486328, 0.042106624603271484, 0.04224691009521484, 0.042430526733398435, 0.04336172866821289, 0.04214822387695313, 0.04281692886352539, 0.04170576095581055, 0.04202950286865234, 0.04162355041503906, 0.04156620788574219, 0.04196147155761719, 0.04227811050415039, 0.04228931045532226, 0.04173068618774414, 0.04155596923828125, 0.04144057464599609, 0.04128780746459961, 0.041476032257080075, 0.04126076889038086, 0.0418436164855957, 0.041842369079589846, 0.042420543670654294, 0.04168035125732422, 0.041490974426269533, 0.04207001495361328, 0.04563478469848633, 0.04276035308837891, 0.04221401596069336, 0.04200966262817383, 0.04215478515625, 0.04230364990234375, 0.042633216857910154, 0.04267212677001953, 0.04249190521240234, 0.042055137634277345, 0.0422116813659668, 0.04330720138549805, 0.04294655990600586, 0.04249302291870117, 0.042042270660400394, 0.042780670166015625, 0.04268185424804687, 0.042608158111572265, 0.042574817657470704, 0.04175667190551758, 0.04232806396484375, 0.04200147247314453, 0.042557472229003905, 0.04238748931884766, 0.04212428665161133, 0.04202073669433594, 0.041647838592529296, 0.0417239990234375, 0.041888031005859375, 0.042127201080322266, 0.04255136108398438, 0.04272332763671875, 0.042570785522460936, 0.042183361053466796, 0.04266214370727539, 0.043644161224365235, 0.04295872116088867, 0.04245967864990234, 0.04193852615356446, 0.04267497634887695, 0.041783199310302735, 0.0422213134765625, 0.04228335952758789, 0.04266131210327148, 0.04202348709106445, 0.04144332885742188, 0.04207206344604492, 0.041605121612548826, 0.04159436798095703, 0.04171404647827148, 0.041729248046875, 0.041781536102294924, 0.04173068618774414, 0.04179753494262695, 0.041439327239990234, 0.041702816009521484, 0.041644416809082034, 0.04145993423461914, 0.0423436164855957, 0.04179167938232422, 0.042496448516845704, 0.041597057342529296, 0.041549888610839844, 0.04286054229736328, 0.04206796646118164, 0.04148588943481445, 0.041080638885498046, 0.041500831604003904, 0.04118985748291016, 0.041613311767578126, 0.041475807189941406, 0.04125699234008789, 0.04136166381835937, 0.04112998580932617, 0.04170326232910156, 0.04156947326660156, 0.042218463897705075, 0.04132767868041992, 0.04133369445800781, 0.0417628173828125, 0.04204339218139649, 0.04165017700195312, 0.04196352005004883, 0.04196352005004883, 0.04230963134765625, 0.042040863037109376, 0.041827838897705076, 0.041947166442871095, 0.04191638565063477, 0.04197679901123047, 0.0419060173034668, 0.04190326309204102, 0.04208051300048828, 0.04199523162841797, 0.04195100784301758, 0.04193280029296875, 0.04269465637207031, 0.04222512054443359, 0.04252316665649414, 0.04202880096435547, 0.04279439926147461, 0.0419951057434082, 0.04220716857910156, 0.04246739196777344, 0.04220064163208008, 0.04223980712890625, 0.04296384048461914, 0.04259404754638672, 0.04207603073120117, 0.042411678314208986, 0.04245142364501953, 0.042299198150634765, 0.04219513702392578, 0.042246143341064454, 0.0420404167175293, 0.041984928131103515, 0.042028736114501954, 0.042221729278564456, 0.042053791046142576, 0.042119102478027345, 0.04257555389404297, 0.045690975189208984, 0.044897857666015624, 0.04198873519897461, 0.042649375915527345, 0.041605438232421875, 0.043014686584472654, 0.042325729370117186, 0.04187750244140625, 0.0413298225402832, 0.04159695816040039, 0.04188857650756836, 0.04153897476196289, 0.041810527801513675, 0.0413647346496582, 0.04138265609741211, 0.04143247985839844, 0.04145414352416992, 0.04227484893798828, 0.04209235382080078, 0.04239993667602539, 0.04154937744140625, 0.04187180709838867, 0.04561942291259766, 0.049512222290039064, 0.04216579055786133, 0.04204576110839844, 0.04276035308837891, 0.04219087982177734, 0.04220451354980469, 0.04197644805908203, 0.04211097717285156, 0.04208025741577148, 0.042090496063232424, 0.041985313415527345, 0.04196220779418945, 0.04204748916625976, 0.042178558349609374, 0.04192182540893555, 0.042003360748291016, 0.042071361541748044, 0.04213593673706055, 0.04214182281494141, 0.04290691375732422, 0.0420107536315918, 0.042048095703125, 0.04223347091674805, 0.04249385452270508, 0.042015201568603514, 0.041987422943115235, 0.042201503753662106, 0.04271974563598633, 0.04233599853515625, 0.04227686309814453, 0.04223299026489258, 0.04222652816772461, 0.04246323013305664, 0.04239769744873047, 0.042221569061279295, 0.042338081359863285, 0.04232828903198242, 0.04233011245727539, 0.04236220932006836, 0.042100704193115235, 0.04197036743164063, 0.04238934326171875, 0.042374462127685544, 0.04314595031738281, 0.04208652877807617, 0.042969024658203125, 0.04216451263427735, 0.04223376083374023, 0.042218910217285154, 0.042539295196533204, 0.042307998657226564, 0.041966655731201175, 0.041603073120117184, 0.04179040145874023, 0.04154777526855469, 0.041488384246826174, 0.04158464050292969, 0.042061630249023436, 0.04203673553466797, 0.04263974380493164, 0.04173459243774414, 0.04170051193237305, 0.0416712646484375, 0.04219097518920899, 0.04255123138427734, 0.042242111206054686, 0.04192665481567383, 0.042485889434814454, 0.04183347320556641, 0.04157734298706055, 0.041338878631591795, 0.04136671829223633, 0.04133135986328125, 0.04134092712402344, 0.04135523223876953, 0.041669822692871096, 0.04183552169799805, 0.04195942306518555, 0.042489696502685546, 0.04157455825805664, 0.04161027145385742, 0.041642753601074216, 0.041467838287353516, 0.04237551879882812, 0.04159686279296875, 0.041417984008789065, 0.041343742370605466, 0.04171366500854492, 0.04166041564941406, 0.04156800079345703, 0.04142630386352539, 0.042251136779785155, 0.042049537658691405, 0.04198390579223633, 0.041910369873046874, 0.04177638244628906, 0.04165036773681641, 0.041468448638916015, 0.04138995361328125, 0.041535648345947265, 0.0416071662902832, 0.04180537414550781, 0.041763263702392577, 0.04143718338012695, 0.04174028778076172, 0.042053089141845704, 0.0433230094909668, 0.043496353149414066, 0.04261996841430664, 0.041974720001220704, 0.04211916732788086, 0.04206387329101562, 0.04197990417480469, 0.04181305694580078, 0.04188051223754883, 0.0416255989074707, 0.04186531066894531, 0.04224524688720703, 0.04215270233154297, 0.04208233642578125, 0.04301193618774414, 0.04189353561401367, 0.042735103607177735, 0.04647417449951172, 0.04221139144897461, 0.04232396697998047, 0.0421212158203125, 0.04203958511352539, 0.0420533447265625, 0.04211097717285156, 0.042051200866699216, 0.042484096527099606, 0.04224204635620117, 0.04243382263183594, 0.04240867233276367, 0.04237107086181641, 0.04254719924926758, 0.042404991149902344, 0.04232278442382813, 0.042256351470947265, 0.04245017623901367, 0.042519359588623046, 0.042390846252441404, 0.04210335922241211, 0.041992321014404296, 0.042549278259277346, 0.042108863830566404, 0.04344623947143555, 0.04219295883178711, 0.04193017578125, 0.0424288330078125, 0.04170563125610351, 0.0421396484375, 0.04192563247680664, 0.04254617691040039, 0.04182015991210938, 0.04186444854736328, 0.04205849456787109, 0.042092544555664066, 0.042180606842041016, 0.04247347259521484, 0.04207206344604492, 0.04195452880859375, 0.041863937377929684, 0.042539039611816404, 0.04204489517211914, 0.04193539047241211, 0.04182246398925781, 0.04182195281982422, 0.04362575912475586, 0.042031711578369144, 0.04241356658935547, 0.041701889038085936, 0.04174848175048828, 0.04589363098144531, 0.04178348922729492, 0.04148515319824219, 0.04243952178955078, 0.04195094299316406, 0.04182611083984375, 0.042385631561279294, 0.042199359893798825, 0.04191007995605469, 0.04213942337036133, 0.04408163070678711, 0.044423168182373046, 0.04185651016235352, 0.04317440032958984, 0.04155187225341797, 0.0416437759399414, 0.042463489532470707, 0.041912254333496095, 0.04154374313354492, 0.041692222595214844, 0.042148799896240235, 0.04151500701904297, 0.04151295852661133, 0.041801727294921875, 0.04165631866455078, 0.04184617614746094, 0.04312124633789063, 0.04183990478515625, 0.041810657501220705, 0.0417628173828125, 0.04188911819458008, 0.04166108703613281, 0.04165836715698242, 0.04170137786865234, 0.041693088531494144, 0.04232611083984375, 0.042352638244628905, 0.042866687774658206, 0.04184796905517578, 0.04169606399536133, 0.04252675247192383, 0.04571340942382812, 0.042031105041503904, 0.041940990447998046, 0.043872257232666016, 0.0416409912109375, 0.04217667388916015, 0.04214566421508789, 0.04210073471069336, 0.04203411102294922, 0.042014495849609375, 0.04169136047363281, 0.04188131332397461, 0.041642143249511716, 0.041844352722167966, 0.04177766418457031, 0.04187516784667969, 0.04239798355102539, 0.04245471954345703, 0.04158700942993164]",tokens/s,23.744983508598548,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4604.7232,7227.768832,0.0,6849.298432,6444.4416,s,1,11.1192958984375,11.1192958984375,0.0,11.1192958984375,11.1192958984375,11.1192958984375,11.1192958984375,[11.1192958984375],,kWh,0.00012169366612085166,1.3416325208518818e-05,3.794947480398325e-05,0.00017305946613335374,,MB,2513.809408,7580.090368,0.0,7172.25984,6822.664192,s,10,3.846915252685547,0.38469152526855466,0.0043924681989724905,0.3848624725341797,0.38978692932128906,0.39062753448486326,0.39130001861572267,"[0.3787657470703125, 0.38294061279296876, 0.3786228332519531, 0.3816611633300781, 0.38704428100585936, 0.3867843322753906, 0.3812321472167969, 0.3896001281738281, 0.3887958679199219, 0.3914681396484375]",tokens/s,665.4682601112291,kWh,1.1370956313141551e-05,1.2540068247944957e-06,7.534941925385192e-06,2.015990506332124e-05,tokens/kWh,12698472.497559736,MB,2518.134784,7582.18752,0.0,7174.356992,6822.666752,s,10,29.13516845703124,2.913516845703125,0.014485912871789908,2.9170203857421875,2.9270207763671876,2.928691906738281,2.9300288110351564,"[2.914953125, 2.9266494140625, 2.926602294921875, 2.919087646484375, 2.924760009765625, 2.899684814453125, 2.901277099609375, 2.9091103515625, 2.930363037109375, 2.8826806640625]",tokens/s,21.62335189271785,kWh,8.464683113810947e-05,9.336741846009291e-06,5.5548302985614514e-05,0.00014953187596973327,tokens/kWh,421314.8507061586,,s,630,29.13214030838015,0.04624149255298433,0.0007635289994446567,0.04621358299255371,0.04679281387329101,0.04714868049621582,0.049186877937316895,"[0.047429630279541016, 0.046777664184570314, 0.046368606567382814, 0.04650419235229492, 0.046519073486328125, 0.04616377639770508, 0.04576847839355469, 0.045975616455078125, 0.04576694488525391, 0.04573974227905273, 0.04565980911254883, 0.04560070419311523, 0.049310272216796874, 0.04598112106323242, 0.0456668815612793, 0.04575775909423828, 0.045660865783691405, 0.04570521545410156, 0.04565929412841797, 0.04563395309448242, 0.04575388717651367, 0.04573072052001953, 0.04565737533569336, 0.045733760833740235, 0.04581046295166016, 0.045792606353759764, 0.04572643280029297, 0.04577859115600586, 0.04562160110473633, 0.04600400161743164, 0.045865184783935545, 0.046085887908935544, 0.04598780822753906, 0.04621855926513672, 0.04605027389526367, 0.046209022521972655, 0.04614908981323242, 0.04626486587524414, 0.04634588623046875, 0.046271041870117186, 0.046976001739501956, 0.04687747192382812, 0.046181632995605466, 0.046363391876220704, 0.0463699836730957, 0.046302017211914064, 0.04765491104125977, 0.046601249694824216, 0.0462628173828125, 0.046614974975585935, 0.046311424255371096, 0.04642969512939453, 0.04631808090209961, 0.04662444686889648, 0.04639081573486328, 0.04690835189819336, 0.046438465118408205, 0.04665116882324219, 0.04669219207763672, 0.04706841659545898, 0.046582687377929685, 0.04677737426757812, 0.04655203247070312, 0.04749337768554687, 0.047298561096191405, 0.04679065704345703, 0.04670019149780273, 0.04669475173950195, 0.04661804962158203, 0.04663312149047852, 0.049396385192871095, 0.0465401611328125, 0.046370784759521486, 0.04650844955444336, 0.046413822174072264, 0.04860927963256836, 0.04674899291992188, 0.04611894226074219, 0.04630524826049805, 0.04630550384521484, 0.0463296012878418, 0.046305694580078126, 0.04632585525512695, 0.04619900894165039, 0.04653833770751953, 0.04640739059448242, 0.046367424011230465, 0.04645996856689453, 0.04681625747680664, 0.046780033111572264, 0.04620719909667969, 0.04619232177734375, 0.04621958541870117, 0.04602889633178711, 0.046069759368896485, 0.04597516632080078, 0.04617254257202148, 0.046034942626953124, 0.0462479362487793, 0.046075233459472655, 0.04624860763549805, 0.04652236938476562, 0.0462490234375, 0.04656825637817383, 0.04630684661865234, 0.04628335952758789, 0.04614348983764648, 0.046059551239013674, 0.04611641693115234, 0.046001632690429686, 0.04605843353271484, 0.0458526725769043, 0.04612460708618164, 0.04614393615722656, 0.0461578254699707, 0.04609638214111328, 0.04629852676391601, 0.046842239379882813, 0.04693398284912109, 0.04638131332397461, 0.046360576629638675, 0.04637446212768555, 0.04621052932739258, 0.04601897430419922, 0.046172737121582035, 0.046514175415039063, 0.04702022552490234, 0.04620889663696289, 0.04620505523681641, 0.04602291107177734, 0.045954334259033204, 0.0459637451171875, 0.04588329696655274, 0.04594492721557617, 0.045889537811279295, 0.04592745590209961, 0.04614371109008789, 0.04618739318847656, 0.0464505615234375, 0.04627046585083008, 0.04604108810424805, 0.04626988983154297, 0.04640169525146484, 0.04628547286987305, 0.046226497650146484, 0.04632547378540039, 0.046416000366210936, 0.046424190521240236, 0.04614806365966797, 0.04641839981079102, 0.04643344116210937, 0.04639603042602539, 0.04624998474121094, 0.04673292922973633, 0.046313888549804685, 0.046311073303222657, 0.04626985549926758, 0.04629801559448242, 0.04614166259765625, 0.04640444946289062, 0.046861248016357424, 0.04632380676269531, 0.04620278549194336, 0.04652995300292969, 0.04653116989135742, 0.04624803161621094, 0.04615142440795898, 0.046217601776123045, 0.0463779182434082, 0.046430942535400394, 0.046450080871582033, 0.046593856811523435, 0.04658473587036133, 0.046812225341796875, 0.04656428909301758, 0.04682137680053711, 0.04667801666259765, 0.04667327880859375, 0.04664131164550781, 0.04683760070800781, 0.046645889282226564, 0.04738800048828125, 0.046827392578125, 0.04794812774658203, 0.04846412658691406, 0.04683180618286133, 0.04630915069580078, 0.04626467132568359, 0.046517822265625, 0.047301025390625, 0.04690665435791016, 0.046297855377197265, 0.04638883209228516, 0.046400096893310545, 0.04610044860839844, 0.04739599990844726, 0.04614374542236328, 0.046024673461914065, 0.04628937530517578, 0.04613750457763672, 0.046137344360351565, 0.0460120964050293, 0.046287166595458985, 0.04915631866455078, 0.04710089492797852, 0.04686726379394531, 0.04642803192138672, 0.0468645133972168, 0.047253440856933594, 0.04643401718139648, 0.0472737922668457, 0.046574142456054686, 0.046609569549560546, 0.04637984085083008, 0.04722809600830078, 0.04649363327026367, 0.04743824005126953, 0.04623183822631836, 0.04704012680053711, 0.04611916732788086, 0.046313793182373046, 0.0460403823852539, 0.04590252685546875, 0.04563711929321289, 0.04562201690673828, 0.045774593353271484, 0.04588710403442383, 0.04586841583251953, 0.045780193328857424, 0.045835617065429685, 0.045924415588378904, 0.04614944076538086, 0.04589625549316406, 0.04573388671875, 0.04640972900390625, 0.04580278396606445, 0.04584521484375, 0.04687200164794922, 0.04624236679077148, 0.045894847869873044, 0.04612179183959961, 0.04612505722045898, 0.046077953338623044, 0.04577689743041992, 0.045639678955078124, 0.04572774505615235, 0.04600012969970703, 0.045969406127929685, 0.046171775817871095, 0.045779327392578124, 0.04629913711547851, 0.04633020782470703, 0.04876806259155273, 0.04646393585205078, 0.04632166290283203, 0.05014479827880859, 0.04643065643310547, 0.046186016082763674, 0.04601641464233398, 0.046037601470947265, 0.04595507049560547, 0.05178121566772461, 0.046066078186035156, 0.04599603271484375, 0.04605145645141601, 0.04620889663696289, 0.04614982223510742, 0.04607145690917969, 0.04603670501708984, 0.046249664306640625, 0.04614009475708008, 0.04631868743896484, 0.046340991973876956, 0.0465880012512207, 0.04627865600585938, 0.04634790420532227, 0.04651251220703125, 0.04663065719604492, 0.04627872085571289, 0.046344383239746094, 0.04633190536499023, 0.046236961364746094, 0.04645692825317383, 0.046446273803710934, 0.046072769165039065, 0.04611481475830078, 0.04586905670166016, 0.045932544708251956, 0.045921825408935545, 0.04633190536499023, 0.0462845458984375, 0.04624867248535156, 0.046081184387207035, 0.0463834228515625, 0.04622201538085938, 0.04663075256347656, 0.04654415893554688, 0.04627734375, 0.046448448181152346, 0.04649593734741211, 0.04601446533203125, 0.04598134231567383, 0.04578134536743164, 0.045981056213378904, 0.045689697265625, 0.045905086517333986, 0.04610108947753906, 0.046276607513427735, 0.04619468688964844, 0.04685823822021484, 0.04614963150024414, 0.04619065475463867, 0.04655916976928711, 0.046219264984130856, 0.046516128540039066, 0.04642816162109375, 0.045428737640380856, 0.04506419372558594, 0.04511350250244141, 0.04505174255371094, 0.044943359375, 0.04531824111938477, 0.04537443161010742, 0.04543379211425781, 0.045330432891845705, 0.04518678283691406, 0.045164192199707034, 0.04495219039916992, 0.045123584747314455, 0.045543071746826175, 0.045898078918457035, 0.04568476867675781, 0.045517791748046876, 0.04555263900756836, 0.04604240036010742, 0.045245151519775394, 0.045107200622558595, 0.04487081527709961, 0.04488687896728515, 0.04539801788330078, 0.04500889587402344, 0.04491059112548828, 0.044948799133300785, 0.045185150146484374, 0.04606556701660156, 0.04549903869628906, 0.04587216186523437, 0.0458106575012207, 0.04657379150390625, 0.04640534210205078, 0.046266433715820315, 0.04621327972412109, 0.046616416931152344, 0.04591817474365235, 0.046155231475830075, 0.04635500717163086, 0.04628236770629883, 0.04639180755615235, 0.046387073516845706, 0.04628297424316406, 0.04628799819946289, 0.046213886260986325, 0.046383102416992186, 0.04647315216064453, 0.04676588821411133, 0.04621327972412109, 0.04629491043090821, 0.04629721450805664, 0.05578956985473633, 0.046063617706298826, 0.046212352752685544, 0.046155712127685544, 0.04669494247436524, 0.04644655990600586, 0.046566879272460934, 0.049199359893798825, 0.046520992279052736, 0.04602259063720703, 0.04690713500976563, 0.047386878967285155, 0.04627391815185547, 0.04624867248535156, 0.04603871917724609, 0.04633417510986328, 0.04627167892456055, 0.046494110107421875, 0.046351806640625, 0.04642649459838867, 0.04647091293334961, 0.046521183013916015, 0.04625612640380859, 0.04637900924682617, 0.04603903961181641, 0.04622857666015625, 0.04966191864013672, 0.04632057571411133, 0.046835712432861325, 0.04589132690429688, 0.0458603515625, 0.045714176177978516, 0.04538777542114258, 0.04511334228515625, 0.04484291076660156, 0.0450683822631836, 0.04547071838378906, 0.04561414337158203, 0.04564940643310547, 0.04569891357421875, 0.045789825439453126, 0.04577059173583985, 0.04567827224731445, 0.04557660675048828, 0.04559408187866211, 0.04560748672485351, 0.045391712188720706, 0.04534860610961914, 0.04506880187988281, 0.04508972930908203, 0.04502422332763672, 0.04551475143432617, 0.04577084732055664, 0.045830047607421875, 0.045946304321289065, 0.046034942626953124, 0.04584096145629883, 0.046063488006591796, 0.04578265762329101, 0.04586102294921875, 0.04581343841552735, 0.04819420623779297, 0.04613119888305664, 0.04617830276489258, 0.046159999847412106, 0.04623654556274414, 0.046048255920410154, 0.046153728485107424, 0.04656742477416992, 0.04622694396972656, 0.046088287353515625, 0.04630979156494141, 0.04651020812988281, 0.046593280792236326, 0.045986560821533205, 0.04842278289794922, 0.04583200073242188, 0.04579359817504883, 0.04574518585205078, 0.0458568000793457, 0.04604121780395508, 0.045705375671386717, 0.045404830932617185, 0.04545510482788086, 0.04591571044921875, 0.04758393478393555, 0.046276256561279296, 0.046091777801513675, 0.046150623321533205, 0.046524288177490235, 0.04637491226196289, 0.04617987060546875, 0.04630499267578125, 0.04633033752441406, 0.046026081085205076, 0.046018943786621094, 0.04617001724243164, 0.04623164749145508, 0.046260833740234375, 0.0459048957824707, 0.046160865783691406, 0.046170112609863284, 0.046249183654785156, 0.045720352172851565, 0.04566185760498047, 0.045698814392089844, 0.045832801818847656, 0.04603606414794922, 0.04626467132568359, 0.04616012954711914, 0.04642969512939453, 0.047033054351806644, 0.046383232116699216, 0.04613321685791016, 0.04621673583984375, 0.04610271835327148, 0.04630374526977539, 0.04612003326416016, 0.04628310394287109, 0.046424415588378905, 0.04645273590087891, 0.0462110710144043, 0.04635660934448242, 0.046229377746582034, 0.04606671905517578, 0.04599292755126953, 0.046048641204833984, 0.045719520568847656, 0.045780895233154296, 0.04696547317504883, 0.04590595245361328, 0.04589977645874024, 0.04622480010986328, 0.04620553588867188, 0.04605763244628906, 0.04610236740112305, 0.04763852691650391, 0.046637054443359374, 0.04628793716430664, 0.04629600143432617, 0.04635062408447266, 0.046419681549072264, 0.046359935760498044, 0.04661635208129883, 0.047059391021728514, 0.04664080047607422, 0.04664499282836914, 0.04649168014526367, 0.0465409927368164, 0.04659827041625977, 0.04730438232421875, 0.046564319610595706, 0.046507614135742184, 0.046491744995117185, 0.04673891067504883, 0.04669731140136719, 0.046473217010498044, 0.04634624099731445, 0.04611427307128906, 0.04608668899536133, 0.04606320190429687, 0.04647305679321289, 0.04650390243530273, 0.04634479904174805, 0.04661782455444336, 0.04633603286743164, 0.046297855377197265, 0.047099903106689454, 0.046266368865966793, 0.04654060745239258, 0.046300479888916016, 0.04623603057861328, 0.046236351013183595, 0.04645001602172852, 0.04610889434814453, 0.046422271728515624, 0.04649369430541992, 0.046354686737060544, 0.046562881469726564, 0.04631478500366211, 0.046359455108642575, 0.046465023040771485, 0.04603289413452148, 0.04630652618408203, 0.046088993072509764, 0.04615724945068359, 0.04618502426147461, 0.048639999389648435, 0.0462330551147461, 0.04647171020507813, 0.04634624099731445, 0.04638515090942383, 0.04638131332397461, 0.04658560180664063, 0.04639334487915039, 0.04689715194702149, 0.046450687408447267, 0.04696268844604492, 0.04675993728637695, 0.047143489837646484, 0.04653039932250977, 0.046260223388671876, 0.047126529693603515, 0.046095424652099606, 0.046137439727783204, 0.04602761459350586, 0.046179840087890625, 0.04581577682495117, 0.04602320098876953, 0.04544102478027344, 0.04552908706665039, 0.04534799957275391, 0.04582038497924805, 0.045787647247314454, 0.04566944122314453, 0.04551968002319336, 0.04559667205810547, 0.0483061752319336, 0.046098270416259766, 0.046069889068603515, 0.045658145904541016, 0.04560076904296875, 0.04566780853271484, 0.04550915145874023, 0.04569190216064453, 0.045448192596435545, 0.04532428741455078, 0.045287200927734375, 0.045448894500732424, 0.04542108917236328, 0.045534400939941405, 0.04554163360595703, 0.045777473449707035, 0.04561932754516602, 0.045566078186035155, 0.04569903945922851, 0.045622753143310546, 0.045857086181640624, 0.0455819206237793, 0.04558480072021484, 0.045563137054443356, 0.045349632263183594, 0.04560451126098633, 0.04522003173828125, 0.04564092636108399, 0.04611305618286133, 0.0460909423828125, 0.045330398559570315, 0.04532595062255859, 0.045078208923339844, 0.04498303985595703, 0.0449060173034668, 0.04715292739868164, 0.045066497802734376, 0.045377918243408207, 0.045034751892089844, 0.04559539031982422, 0.045864959716796876, 0.045716510772705075, 0.04557104110717773, 0.046080001831054686, 0.04577510452270508]",tokens/s,21.625599538210885,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,890.884096,638.517248,0.0,260.046848,253.520896,s,1,8.0936240234375,8.0936240234375,0.0,8.0936240234375,8.0936240234375,8.0936240234375,8.0936240234375,[8.0936240234375],,kWh,2.178515275418249e-05,2.395556827885202e-06,6.556116355999064e-06,3.073682593806676e-05,,MB,1186.992128,743.374848,0.0,335.54432,312.653824,s,11,0.17854700851440433,0.016231546228582208,0.000630683680541129,0.016063615798950194,0.016344608306884764,0.017258992195129393,0.017990499305725098,"[0.016101696014404296, 0.018173376083374024, 0.015791199684143067, 0.01610633659362793, 0.01587868785858154, 0.01605958366394043, 0.016063615798950194, 0.01605830383300781, 0.016344608306884764, 0.01611008071899414, 0.015859519958496094]",tokens/s,15771.75682432573,kWh,4.6838706919478506e-07,5.165474401098827e-08,3.0948883328730977e-07,8.29530646493083e-07,tokens/kWh,308608248.6310343,MB,1220.722688,768.540672,0.0,360.710144,313.442816,s,11,10.188017028808595,0.9261833662553268,0.012551883267073127,0.9244647827148438,0.94057861328125,0.9455420227050781,0.9495127502441406,"[0.9354699096679687, 0.9142109375, 0.9175198364257813, 0.9177374877929687, 0.910325439453125, 0.9118135375976563, 0.94057861328125, 0.9352012329101562, 0.9505054321289063, 0.9301898193359375, 0.9244647827148438]",tokens/s,68.02108771907311,kWh,2.6536873391409425e-05,2.9263875942598974e-06,1.000289315398753e-05,3.946615413965684e-05,tokens/kWh,1596304.5139150159,,s,693,10.182142732620244,0.014692846656017661,0.0003464994459911771,0.014606592178344726,0.015085388946533204,0.015232236480712891,0.015802830352783206,"[0.014614239692687988, 0.01470902442932129, 0.014744992256164552, 0.014729344367980957, 0.014690784454345704, 0.014905344009399414, 0.014589183807373047, 0.014713824272155762, 0.01480454444885254, 0.014604512214660645, 0.014548224449157715, 0.014502655982971192, 0.014434304237365723, 0.014461312294006348, 0.01456009578704834, 0.01461958408355713, 0.01464031982421875, 0.014836383819580077, 0.015117664337158204, 0.01471555233001709, 0.015302463531494141, 0.015609663963317871, 0.015290656089782716, 0.014775775909423828, 0.014844544410705566, 0.014679072380065918, 0.014588895797729493, 0.014684063911437988, 0.014686304092407227, 0.014724960327148437, 0.014532959938049317, 0.01443827247619629, 0.014497664451599121, 0.014600255966186523, 0.01458348846435547, 0.014523967742919922, 0.014648063659667969, 0.014520319938659668, 0.015224831581115723, 0.015095071792602539, 0.014969599723815919, 0.014935551643371582, 0.014897120475769044, 0.01520400047302246, 0.01493238353729248, 0.014997407913208008, 0.014850591659545899, 0.014904576301574707, 0.014811296463012696, 0.014827584266662598, 0.014745823860168457, 0.014808799743652344, 0.014698335647583008, 0.0147009916305542, 0.014768447875976563, 0.014739456176757813, 0.014978400230407715, 0.01481385612487793, 0.014935135841369629, 0.014817952156066894, 0.014854399681091309, 0.015026176452636719, 0.01817728042602539, 0.014867391586303711, 0.01584323215484619, 0.014706784248352051, 0.014812191963195801, 0.014917920112609864, 0.014889663696289062, 0.01480835247039795, 0.014758624076843262, 0.014587743759155274, 0.014481568336486816, 0.014499679565429687, 0.014436511993408204, 0.01445683193206787, 0.014389247894287109, 0.014480671882629394, 0.014519136428833007, 0.01451580810546875, 0.014390912055969238, 0.014434975624084473, 0.014417920112609863, 0.014432095527648925, 0.01487433624267578, 0.014385120391845703, 0.014516703605651856, 0.014431488037109376, 0.014664447784423829, 0.01471232032775879, 0.01463321590423584, 0.01460431957244873, 0.014417695999145507, 0.014420607566833495, 0.014417856216430665, 0.014481568336486816, 0.01434934425354004, 0.014382816314697266, 0.014328800201416015, 0.01439305591583252, 0.01442966365814209, 0.014330495834350586, 0.014367103576660156, 0.01438918399810791, 0.014426336288452149, 0.01446828842163086, 0.014430047988891602, 0.014381695747375488, 0.014360575675964356, 0.014430208206176758, 0.014317567825317384, 0.014369855880737305, 0.014385855674743653, 0.014327232360839843, 0.014361408233642578, 0.014321727752685547, 0.014345439910888672, 0.014293727874755859, 0.014419103622436523, 0.014627679824829102, 0.014380319595336914, 0.014381792068481445, 0.014395615577697755, 0.014350111961364747, 0.014395392417907715, 0.014369119644165039, 0.014117183685302735, 0.014398112297058105, 0.014665727615356445, 0.014528512001037597, 0.014538847923278808, 0.014790559768676758, 0.01469161605834961, 0.014564064025878906, 0.014527935981750489, 0.014588607788085937, 0.014555232048034668, 0.014437664031982421, 0.014676480293273926, 0.014425439834594727, 0.014437024116516112, 0.014375967979431152, 0.014429408073425293, 0.014372639656066895, 0.01439072036743164, 0.014428832054138184, 0.014454976081848144, 0.014473983764648438, 0.01443068790435791, 0.0144071683883667, 0.014535264015197754, 0.014464384078979492, 0.014821663856506347, 0.014568511962890624, 0.01455235195159912, 0.01445241641998291, 0.014437055587768555, 0.014471199989318848, 0.014391263961791993, 0.014435680389404296, 0.014373536109924317, 0.014475263595581055, 0.014367839813232423, 0.014383616447448731, 0.014451328277587891, 0.014391072273254394, 0.014424287796020507, 0.014352160453796387, 0.014342016220092773, 0.014331168174743652, 0.014411680221557617, 0.0145863676071167, 0.014642880439758301, 0.014778112411499024, 0.014664704322814942, 0.014690303802490234, 0.014650976181030274, 0.01466204833984375, 0.014618271827697754, 0.014717280387878418, 0.015019840240478515, 0.014759584426879883, 0.014907391548156738, 0.014824064254760742, 0.014813311576843262, 0.014983967781066894, 0.014774335861206054, 0.014844863891601563, 0.01488486385345459, 0.014479552268981934, 0.014705727577209473, 0.01476848030090332, 0.014633376121520996, 0.014566720008850098, 0.014535264015197754, 0.014512384414672851, 0.014499360084533692, 0.014529024124145508, 0.014560480117797852, 0.014809727668762207, 0.014993311882019043, 0.014733375549316407, 0.014892928123474121, 0.014780223846435547, 0.014715392112731934, 0.01469545555114746, 0.01475436782836914, 0.014534879684448242, 0.01453273582458496, 0.014486880302429199, 0.01444819164276123, 0.014502880096435546, 0.014491968154907227, 0.014456512451171875, 0.014628864288330079, 0.014548992156982422, 0.014505023956298829, 0.014558143615722656, 0.014520319938659668, 0.014714112281799316, 0.014648063659667969, 0.014716416358947755, 0.014848095893859863, 0.014899328231811523, 0.014731552124023438, 0.014530783653259277, 0.014485280036926269, 0.014374784469604492, 0.014373056411743164, 0.014350272178649903, 0.014413567543029785, 0.014388959884643554, 0.014424608230590821, 0.014438400268554688, 0.014392800331115722, 0.014416416168212891, 0.014585856437683106, 0.01444377613067627, 0.014432000160217286, 0.014425087928771972, 0.014475104331970214, 0.014553088188171387, 0.014522047996520996, 0.014737759590148926, 0.014430335998535156, 0.01452006435394287, 0.014571359634399414, 0.01448185634613037, 0.014391391754150391, 0.014407456398010254, 0.014351648330688476, 0.014405728340148926, 0.01460870361328125, 0.014989631652832031, 0.014610207557678223, 0.014567328453063964, 0.014427583694458007, 0.014369440078735352, 0.014406944274902344, 0.014377344131469726, 0.014529184341430664, 0.014462143898010253, 0.014496255874633789, 0.014506336212158202, 0.014372832298278808, 0.014397055625915527, 0.014415936470031738, 0.014475263595581055, 0.014882207870483399, 0.014498399734497071, 0.014442239761352539, 0.014440287590026855, 0.014360544204711913, 0.014375359535217284, 0.014340096473693848, 0.014364992141723633, 0.014370495796203614, 0.014323264122009278, 0.014354880332946778, 0.01438640022277832, 0.014323776245117187, 0.014392352104187012, 0.014356351852416993, 0.014393247604370118, 0.0144301118850708, 0.014428352355957031, 0.014393152236938477, 0.014565535545349121, 0.014665568351745605, 0.01444812774658203, 0.014350496292114257, 0.014416223526000976, 0.014366720199584961, 0.014358592033386231, 0.014393280029296875, 0.014359775543212891, 0.014422271728515625, 0.014377504348754884, 0.014375007629394532, 0.014453824043273927, 0.014449503898620606, 0.014454784393310547, 0.014462976455688477, 0.01437228775024414, 0.014409503936767577, 0.014327808380126953, 0.01437507152557373, 0.014299776077270508, 0.014430208206176758, 0.014557184219360352, 0.01450598430633545, 0.014397279739379883, 0.014460960388183593, 0.014378496170043945, 0.014529055595397949, 0.014119520187377929, 0.014393376350402833, 0.014448543548583985, 0.014396832466125489, 0.014375455856323242, 0.01447766399383545, 0.014409472465515136, 0.014401472091674805, 0.014417568206787109, 0.014398880004882812, 0.014375904083251952, 0.014432255744934081, 0.014567423820495605, 0.014413824081420898, 0.014524415969848633, 0.01459609603881836, 0.014657695770263672, 0.014692192077636718, 0.01463651180267334, 0.0146212158203125, 0.014690400123596192, 0.014626015663146972, 0.014977631568908691, 0.014694016456604004, 0.014780896186828613, 0.014547200202941894, 0.014524160385131836, 0.014519840240478516, 0.01439311981201172, 0.014412480354309082, 0.014397024154663086, 0.014307583808898926, 0.014373023986816407, 0.014307647705078126, 0.014395423889160156, 0.014458304405212402, 0.014535008430480957, 0.014538623809814454, 0.014443936347961426, 0.014395999908447265, 0.014384544372558594, 0.014768735885620117, 0.01437491226196289, 0.014316703796386719, 0.014332192420959473, 0.014332480430603027, 0.014397439956665039, 0.014442432403564453, 0.014370400428771973, 0.014315008163452148, 0.014290911674499512, 0.01438003158569336, 0.014419424057006835, 0.014352928161621094, 0.01450979232788086, 0.01436451244354248, 0.014465472221374513, 0.014389311790466309, 0.01438304042816162, 0.014411775588989258, 0.014348287582397461, 0.014509696006774903, 0.014784159660339356, 0.014319616317749024, 0.014532768249511718, 0.014601663589477539, 0.014612671852111816, 0.014545120239257813, 0.01455907154083252, 0.014571680068969726, 0.01458140754699707, 0.014508543968200683, 0.014475359916687011, 0.015687423706054686, 0.015167296409606934, 0.01461676788330078, 0.014574751853942872, 0.014708864212036132, 0.014679072380065918, 0.014587615966796875, 0.014855968475341796, 0.014998720169067383, 0.014881600379943847, 0.015017888069152833, 0.014896672248840332, 0.015026944160461426, 0.01489635181427002, 0.015038463592529297, 0.014926336288452148, 0.015067328453063964, 0.01544540786743164, 0.01521123218536377, 0.015208415985107422, 0.015228639602661133, 0.015042719841003417, 0.014979328155517578, 0.01500054359436035, 0.015006624221801757, 0.015005696296691894, 0.015031871795654296, 0.015147456169128418, 0.015120384216308593, 0.015183327674865722, 0.015212736129760741, 0.015103712081909179, 0.015173664093017578, 0.015030783653259277, 0.015546463966369628, 0.014925919532775878, 0.014909343719482422, 0.014907391548156738, 0.01490124797821045, 0.014796064376831055, 0.014870880126953125, 0.014865023612976074, 0.014750816345214844, 0.014678463935852051, 0.014590175628662109, 0.014512224197387695, 0.014552096366882324, 0.014738400459289551, 0.01580022430419922, 0.016285696029663087, 0.014831744194030761, 0.014757760047912597, 0.014772031784057617, 0.014481023788452148, 0.014837887763977051, 0.014665984153747558, 0.01463644790649414, 0.014535327911376954, 0.014540960311889648, 0.014528448104858399, 0.014579551696777344, 0.014470560073852539, 0.014491488456726074, 0.014545087814331054, 0.014586432456970215, 0.014566656112670898, 0.01471673583984375, 0.014826432228088379, 0.014934016227722167, 0.014888959884643555, 0.014899200439453125, 0.014970879554748535, 0.015467743873596191, 0.015020832061767578, 0.014929920196533204, 0.014934240341186523, 0.015047743797302246, 0.015020095825195312, 0.014965503692626954, 0.01500489616394043, 0.01503916835784912, 0.014921440124511718, 0.014979359626770019, 0.015069503784179688, 0.015308480262756347, 0.014982720375061036, 0.014976799964904785, 0.015966943740844727, 0.01477830410003662, 0.014673919677734374, 0.014792703628540039, 0.015450336456298829, 0.014694175720214843, 0.01466915225982666, 0.01478927993774414, 0.01482755184173584, 0.014856160163879395, 0.014761823654174805, 0.014881952285766602, 0.014806015968322754, 0.014869600296020509, 0.015147808074951172, 0.014712960243225097, 0.01466982364654541, 0.014606559753417969, 0.014619903564453125, 0.0146746244430542, 0.014758848190307617, 0.014719903945922852, 0.01472822380065918, 0.014679007530212403, 0.0147807035446167, 0.014726880073547364, 0.014744768142700196, 0.01479967975616455, 0.015048735618591309, 0.014773951530456544, 0.014869855880737306, 0.014902527809143067, 0.014991071701049805, 0.015122431755065918, 0.01496678352355957, 0.014996512413024903, 0.014933152198791504, 0.014984383583068847, 0.014866720199584961, 0.014836064338684082, 0.014826623916625977, 0.014895999908447265, 0.014890496253967284, 0.014895135879516602, 0.014989215850830078, 0.01519699192047119, 0.015832799911499024, 0.015226911544799805, 0.015153152465820312, 0.015089664459228515, 0.015126527786254883, 0.015046655654907226, 0.015157471656799317, 0.015032095909118652, 0.015071231842041016, 0.015036416053771973, 0.015095808029174805, 0.015102047920227051, 0.01518172836303711, 0.01519001579284668, 0.01508892822265625, 0.015329279899597169, 0.01512112045288086, 0.015469568252563477, 0.015430111885070801, 0.015186464309692383, 0.015420736312866211, 0.015291168212890624, 0.015312800407409668, 0.015278016090393066, 0.015239232063293457, 0.015313983917236328, 0.015227840423583985, 0.015194304466247558, 0.015224639892578124, 0.015208127975463868, 0.015093215942382813, 0.015339424133300781, 0.015237631797790528, 0.01574678421020508, 0.015131360054016113, 0.014887167930603027, 0.014800992012023926, 0.014777664184570313, 0.014836064338684082, 0.014712832450866698, 0.014816351890563965, 0.015000479698181152, 0.01466163158416748, 0.014682111740112304, 0.014724543571472168, 0.014832192420959473, 0.014684160232543946, 0.014946559906005859, 0.014970751762390136, 0.014952320098876953, 0.014841535568237304, 0.014899359703063966, 0.015005632400512696, 0.014923775672912597, 0.015030495643615723, 0.014972576141357422, 0.01497270393371582, 0.014933695793151855, 0.014852992057800293, 0.014999135971069336, 0.014901856422424316, 0.014874431610107422, 0.014974944114685059, 0.014817440032958984, 0.014624608039855958, 0.014554176330566406, 0.014545439720153808, 0.014555583953857422, 0.014718976020812988, 0.014593855857849121, 0.014660863876342774, 0.01458886432647705, 0.014499839782714843, 0.014533632278442383, 0.014533632278442383, 0.014589952468872071, 0.014518272399902344, 0.01459404754638672, 0.01458892822265625, 0.014652640342712402, 0.014548768043518066, 0.014442496299743653, 0.014552479743957519, 0.014517919540405273, 0.014930879592895508, 0.014667776107788086, 0.014800704002380372, 0.015275584220886231, 0.016126592636108397, 0.01489305591583252, 0.015364128112792969, 0.014784735679626465, 0.014690048217773437, 0.01478438377380371, 0.01480947208404541, 0.014774016380310059, 0.014694399833679199, 0.014593695640563965, 0.014720640182495116, 0.014914272308349609, 0.015392607688903808, 0.014518048286437988, 0.014575167655944824, 0.014459712028503418, 0.014483455657958985, 0.014360223770141602, 0.014389599800109863, 0.014359807968139649, 0.014379872322082519, 0.01410483169555664, 0.014473119735717773, 0.014403327941894532, 0.01431497573852539, 0.014499072074890137, 0.014427935600280762, 0.014423711776733399, 0.014729472160339355, 0.014969152450561523, 0.015613280296325683, 0.014866815567016602, 0.014525823593139649, 0.014502528190612794, 0.014456607818603516, 0.014381471633911133, 0.014397248268127442, 0.014507519721984862, 0.014579968452453613, 0.01440118408203125, 0.014354847908020019, 0.014507648468017578, 0.014489439964294434, 0.014462847709655762, 0.014944671630859375, 0.015681983947753907, 0.01439948844909668, 0.014361791610717774, 0.01451296043395996, 0.017330080032348632, 0.015478879928588866, 0.014782079696655273, 0.014641471862792969, 0.014608063697814942, 0.014493439674377441, 0.014694304466247558, 0.014639616012573242, 0.014704544067382813, 0.014704416275024414, 0.014606592178344726, 0.014633248329162598, 0.014667776107788086, 0.014657088279724121, 0.014944704055786132, 0.01485807991027832, 0.01493008041381836, 0.014620672225952149, 0.014595999717712402, 0.01454089641571045, 0.01454905605316162, 0.014630463600158692, 0.014600576400756836, 0.014684191703796387, 0.014694368362426758, 0.014485504150390625, 0.014575615882873535, 0.014544896125793457, 0.014522656440734864, 0.014476256370544434, 0.014951168060302734, 0.014427904129028321, 0.014528767585754394, 0.014419360160827637, 0.014467231750488281]",tokens/s,68.06033054121858,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1550.06976,1546.584064,0.0,1168.113664,1154.613248,s,1,8.5274755859375,8.5274755859375,0.0,8.5274755859375,8.5274755859375,8.5274755859375,8.5274755859375,[8.5274755859375],,kWh,4.070899408332024e-05,4.479947221899825e-06,1.3120843829997342e-05,5.8309785135217407e-05,,MB,1664.581632,1796.145152,0.0,1388.314624,1334.065152,s,10,0.8429779586791992,0.08429779586791993,0.0010402176392791894,0.08405886459350587,0.08557998428344726,0.08595285682678223,0.0862511548614502,"[0.08632572937011719, 0.08418966674804687, 0.08447596740722656, 0.08356633758544922, 0.08264189147949219, 0.08356393432617187, 0.08392806243896485, 0.08549712371826172, 0.08518681335449219, 0.08360243225097656]",tokens/s,3036.8528306612875,kWh,2.654149214393838e-06,2.9270417739078424e-07,1.7587312049636856e-06,4.705584596748308e-06,tokens/kWh,54403442.27939356,MB,1672.728576,1796.145152,0.0,1388.314624,1372.847616,s,10,17.796296264648436,1.7796296264648437,0.018806836818460466,1.7843126831054688,1.7998743530273436,1.8018918273925781,1.8035058068847656,"[1.754376953125, 1.8039093017578125, 1.797359130859375, 1.775194580078125, 1.799426025390625, 1.780257568359375, 1.790001953125, 1.7883677978515624, 1.7484576416015625, 1.7589453125]",tokens/s,35.40062441259013,kWh,5.1194525785188586e-05,5.646437273394922e-06,2.3576205729637518e-05,8.041716878822102e-05,tokens/kWh,783414.7974782697,,s,630,17.79308543968202,0.02824299276140001,0.0011451475570666266,0.028156975746154785,0.028709516906738283,0.02893239507675171,0.03039027721405029,"[0.028347967147827148, 0.028173887252807617, 0.02804630470275879, 0.027898815155029295, 0.027630207061767578, 0.027489696502685547, 0.02750262451171875, 0.02818307113647461, 0.028131744384765626, 0.027820287704467775, 0.027647743225097655, 0.02754150390625, 0.027928640365600586, 0.02751055908203125, 0.027539520263671874, 0.027938079833984376, 0.032262687683105466, 0.02765622329711914, 0.02770966339111328, 0.027470111846923828, 0.02747737693786621, 0.027402624130249024, 0.027320320129394532, 0.027471168518066406, 0.02734355163574219, 0.0273242244720459, 0.027297760009765627, 0.0273670711517334, 0.0274192008972168, 0.027426816940307616, 0.02728550338745117, 0.02736947250366211, 0.027922431945800782, 0.027507871627807618, 0.027493215560913085, 0.027508544921875, 0.02774822425842285, 0.02799238395690918, 0.02780364799499512, 0.027418464660644532, 0.027539615631103517, 0.027536415100097657, 0.027580671310424805, 0.02775321578979492, 0.02748752021789551, 0.027631776809692383, 0.027656192779541015, 0.027713247299194336, 0.02773801612854004, 0.02775721549987793, 0.02753971290588379, 0.02813747215270996, 0.02847123146057129, 0.02823583984375, 0.02827894401550293, 0.028329824447631834, 0.028383232116699218, 0.02840131187438965, 0.028323360443115234, 0.02848031997680664, 0.028266239166259765, 0.02809881591796875, 0.02840985679626465, 0.028321184158325196, 0.027980480194091797, 0.02810259246826172, 0.02776470375061035, 0.027598751068115233, 0.02824140739440918, 0.027593311309814454, 0.02793408012390137, 0.027883552551269532, 0.027746719360351564, 0.0280347843170166, 0.02834899139404297, 0.0281343994140625, 0.028369823455810548, 0.028210432052612304, 0.028240447998046876, 0.02841779136657715, 0.02820243263244629, 0.028291456222534178, 0.028301952362060546, 0.028539199829101563, 0.02896988868713379, 0.028889888763427733, 0.02857779121398926, 0.02877644729614258, 0.028463104248046874, 0.028487327575683594, 0.02857401657104492, 0.028485376358032225, 0.028525920867919923, 0.028576704025268556, 0.02860032081604004, 0.028614912033081055, 0.028430080413818358, 0.02853865623474121, 0.02829737663269043, 0.02852022361755371, 0.028561279296875, 0.028684703826904297, 0.028667552947998047, 0.029266271591186523, 0.028671871185302733, 0.028827775955200197, 0.02849577522277832, 0.02862886428833008, 0.02839369583129883, 0.028612607955932616, 0.028532224655151366, 0.02891811180114746, 0.028677919387817382, 0.028694623947143554, 0.02841539192199707, 0.028382080078125, 0.02847100830078125, 0.028913375854492187, 0.03040108871459961, 0.038317310333251954, 0.028854560852050782, 0.028921951293945314, 0.028577152252197265, 0.028563840866088867, 0.028265056610107423, 0.02826857566833496, 0.028184576034545897, 0.028231264114379883, 0.028139936447143556, 0.027901952743530273, 0.02787708854675293, 0.028129568099975587, 0.028105056762695313, 0.028063392639160155, 0.027867136001586915, 0.027824127197265625, 0.028065792083740236, 0.028045312881469726, 0.028097568511962892, 0.027932863235473632, 0.028128000259399415, 0.028043296813964842, 0.027934879302978517, 0.02798316764831543, 0.028115488052368163, 0.027918048858642578, 0.028054880142211913, 0.02830636787414551, 0.02913385581970215, 0.028821983337402345, 0.029135360717773437, 0.028641279220581056, 0.028837888717651368, 0.02853388786315918, 0.028982431411743163, 0.0286778564453125, 0.028540191650390626, 0.028574432373046875, 0.028938112258911134, 0.028708255767822266, 0.028738271713256835, 0.028656896591186524, 0.028739904403686522, 0.028539295196533202, 0.03204691314697266, 0.0287227840423584, 0.028695167541503905, 0.028473344802856446, 0.028641183853149413, 0.028676416397094725, 0.02856723213195801, 0.02895462417602539, 0.02892540740966797, 0.029159839630126954, 0.031244640350341798, 0.028682111740112304, 0.028450815200805665, 0.02853424072265625, 0.028308000564575195, 0.028440576553344726, 0.028318912506103515, 0.02851308822631836, 0.028380704879760743, 0.028062175750732422, 0.028473663330078124, 0.02842505645751953, 0.028358879089355468, 0.02838796806335449, 0.028399072647094726, 0.028412864685058593, 0.028499008178710938, 0.028568063735961914, 0.02851980781555176, 0.028526496887207032, 0.02850934410095215, 0.028517663955688475, 0.02850774383544922, 0.028530847549438475, 0.02863577651977539, 0.028563167572021483, 0.028403743743896485, 0.02866828727722168, 0.028665855407714845, 0.02854707145690918, 0.0284649600982666, 0.02846473693847656, 0.028299583435058593, 0.0282708797454834, 0.029572256088256837, 0.028367712020874025, 0.02833203125, 0.028167327880859374, 0.028121952056884766, 0.02811404800415039, 0.02822028732299805, 0.028296319961547852, 0.0281298885345459, 0.028228000640869142, 0.02823360061645508, 0.028096511840820314, 0.028269920349121094, 0.028242591857910157, 0.027910144805908203, 0.027777023315429687, 0.028016639709472657, 0.028052543640136717, 0.028124095916748047, 0.027918336868286132, 0.027760255813598634, 0.02764633560180664, 0.027873279571533204, 0.02807734489440918, 0.0278822078704834, 0.027623424530029295, 0.02753126335144043, 0.027678720474243163, 0.027975391387939454, 0.02793017578125, 0.02766659164428711, 0.027761215209960936, 0.02747542381286621, 0.027478559494018555, 0.027429920196533203, 0.027726816177368163, 0.027803199768066406, 0.028300031661987305, 0.0281146240234375, 0.028085695266723634, 0.028172672271728514, 0.0284136962890625, 0.02830175971984863, 0.028376768112182617, 0.028276895523071287, 0.028364383697509765, 0.028458879470825194, 0.028431327819824218, 0.03004729652404785, 0.05117843246459961, 0.02872118377685547, 0.02840982437133789, 0.028112895965576173, 0.028033023834228517, 0.027996160507202147, 0.027813888549804686, 0.027606624603271485, 0.02763408088684082, 0.028135744094848633, 0.02817513656616211, 0.028349344253540038, 0.028194816589355468, 0.02814300727844238, 0.02824457550048828, 0.02813862419128418, 0.028111743927001952, 0.02821686363220215, 0.02834480094909668, 0.02835446357727051, 0.02835465621948242, 0.028280832290649413, 0.028285152435302736, 0.028450592041015625, 0.028362207412719727, 0.028424032211303712, 0.028492128372192383, 0.02845280075073242, 0.028489248275756836, 0.028382080078125, 0.028366559982299804, 0.028551136016845703, 0.02860678482055664, 0.028708864212036132, 0.028308736801147462, 0.02853923225402832, 0.028408159255981447, 0.028465215682983398, 0.028446720123291015, 0.028237152099609374, 0.028225311279296873, 0.028033279418945314, 0.027943552017211912, 0.027768863677978515, 0.027643871307373048, 0.027741344451904296, 0.02765500831604004, 0.027676671981811524, 0.027699199676513672, 0.028555328369140626, 0.028042879104614258, 0.027953472137451172, 0.027792640686035156, 0.027548063278198243, 0.027640159606933595, 0.02766758346557617, 0.027653343200683595, 0.027841760635375978, 0.02763795280456543, 0.027631423950195313, 0.027664384841918944, 0.027641855239868163, 0.027744255065917968, 0.027840448379516602, 0.027870975494384765, 0.02790777587890625, 0.027751039505004883, 0.02790809631347656, 0.02817024040222168, 0.028057855606079103, 0.027937728881835936, 0.027847423553466796, 0.02787743949890137, 0.028012287139892577, 0.02787353515625, 0.028089696884155274, 0.02827552032470703, 0.028347808837890624, 0.028282720565795897, 0.02845756721496582, 0.028661855697631834, 0.02875775909423828, 0.028553375244140623, 0.02830326461791992, 0.02851878356933594, 0.028472511291503907, 0.02827248001098633, 0.028547136306762696, 0.02882009506225586, 0.02853446388244629, 0.028469568252563478, 0.028459007263183594, 0.028580863952636718, 0.028834815979003905, 0.029111648559570314, 0.028807840347290038, 0.028478975296020507, 0.028272607803344726, 0.028276960372924806, 0.028380928039550782, 0.02832441520690918, 0.028604415893554686, 0.02862598419189453, 0.02816819190979004, 0.027999103546142576, 0.02792246437072754, 0.02786511993408203, 0.027631616592407225, 0.027643903732299805, 0.02770652770996094, 0.027667295455932616, 0.027684864044189454, 0.028004352569580077, 0.02868409538269043, 0.028317888259887694, 0.028375040054321288, 0.028550752639770506, 0.029904287338256837, 0.028767232894897462, 0.02860851287841797, 0.028777503967285158, 0.02865190315246582, 0.028595487594604493, 0.028899904251098632, 0.029198495864868165, 0.0286167049407959, 0.028628992080688476, 0.028804096221923828, 0.028681215286254884, 0.028921344757080077, 0.028273151397705077, 0.02795315170288086, 0.028103679656982423, 0.028205440521240233, 0.02824425506591797, 0.028066144943237305, 0.027867136001586915, 0.02795302391052246, 0.028059776306152345, 0.02956492805480957, 0.02929360008239746, 0.028584831237792967, 0.028504159927368163, 0.02852681541442871, 0.028417503356933594, 0.028531007766723633, 0.029417472839355467, 0.03166435241699219, 0.02990582466125488, 0.028686496734619142, 0.028770111083984376, 0.02883260726928711, 0.028548959732055665, 0.028539264678955078, 0.028515968322753906, 0.02854729652404785, 0.02856243133544922, 0.028275808334350585, 0.029476768493652345, 0.02832399940490723, 0.02798896026611328, 0.028429183959960937, 0.027692895889282226, 0.027678911209106444, 0.02820627212524414, 0.027910400390625, 0.027757087707519532, 0.02759884834289551, 0.02773526382446289, 0.02769817543029785, 0.027866912841796876, 0.027611135482788086, 0.027463903427124025, 0.02823516845703125, 0.027617664337158204, 0.02793878364562988, 0.028162080764770506, 0.0280927677154541, 0.028001951217651366, 0.028057727813720703, 0.0281814079284668, 0.02803990364074707, 0.028211456298828125, 0.028301151275634765, 0.028246751785278322, 0.027991968154907225, 0.027998207092285156, 0.028185951232910157, 0.028715391159057618, 0.02825449562072754, 0.02813337516784668, 0.027995168685913088, 0.027949920654296877, 0.027926336288452147, 0.027947328567504884, 0.02791753578186035, 0.027996959686279296, 0.028116159439086914, 0.028574527740478514, 0.02867100715637207, 0.028666847229003905, 0.028634944915771485, 0.02863644790649414, 0.028533151626586914, 0.029053440093994142, 0.028706016540527343, 0.028749664306640624, 0.028744640350341796, 0.028702720642089844, 0.028688383102416993, 0.028667903900146483, 0.02878268814086914, 0.028759231567382814, 0.02867635154724121, 0.028561248779296874, 0.02840025520324707, 0.028188671112060547, 0.02792038345336914, 0.028002368927001954, 0.02809878349304199, 0.027899423599243165, 0.02772812843322754, 0.027600160598754885, 0.027589536666870116, 0.028364608764648438, 0.027756479263305663, 0.027684864044189454, 0.027747392654418945, 0.027784128189086914, 0.029300512313842772, 0.029427936553955078, 0.02851430320739746, 0.028366847991943358, 0.028307136535644532, 0.028549440383911134, 0.028876800537109375, 0.02838105583190918, 0.028481599807739257, 0.028417728424072267, 0.02840025520324707, 0.028271808624267578, 0.028291648864746093, 0.028431840896606445, 0.027945344924926757, 0.030363807678222655, 0.029683616638183592, 0.028149856567382812, 0.02797964859008789, 0.028027904510498046, 0.028443647384643556, 0.0277708797454834, 0.027559232711791993, 0.027859615325927733, 0.02813465690612793, 0.028023584365844727, 0.02804908752441406, 0.028262367248535158, 0.028155296325683594, 0.028170656204223633, 0.02801308822631836, 0.028020736694335937, 0.027997215270996093, 0.02822857666015625, 0.027950687408447264, 0.028074016571044923, 0.027637151718139647, 0.027653087615966798, 0.027926336288452147, 0.02780588722229004, 0.027891712188720705, 0.028100608825683594, 0.02775359916687012, 0.02756857681274414, 0.027524927139282226, 0.027510591506958008, 0.02752582359313965, 0.027549407958984376, 0.0274619197845459, 0.02741036796569824, 0.0277956485748291, 0.027865087509155274, 0.027749568939208984, 0.027547679901123046, 0.027495168685913087, 0.027539424896240235, 0.027529151916503906, 0.027419967651367186, 0.02747065544128418, 0.02749849510192871, 0.02758844757080078, 0.02748355293273926, 0.029563648223876953, 0.027705184936523436, 0.027649248123168945, 0.027410400390625, 0.027398752212524413, 0.027502975463867187, 0.027469823837280274, 0.02746918487548828, 0.027447328567504883, 0.027451999664306642, 0.027376991271972656, 0.027402912139892578, 0.027688159942626953, 0.027606143951416015, 0.027440832138061522, 0.02783228874206543, 0.02764419174194336, 0.02765180778503418, 0.027490495681762695, 0.027822080612182616, 0.027830272674560546, 0.02750771141052246, 0.027539968490600586, 0.02776655960083008, 0.02800924873352051, 0.02794879913330078, 0.027959327697753906, 0.02828441619873047, 0.028017311096191405, 0.027958879470825194, 0.027988288879394533, 0.02789596748352051, 0.027592639923095703, 0.027786975860595704, 0.027967775344848633, 0.02780076789855957, 0.027708223342895508, 0.02759884834289551, 0.02754764747619629, 0.02736262321472168, 0.027426816940307616, 0.027738559722900392, 0.027621631622314454, 0.02769273567199707, 0.02750889587402344, 0.02790220832824707, 0.02832988739013672, 0.028059167861938475, 0.029640287399291993, 0.02838969612121582, 0.028481439590454103, 0.02816067123413086, 0.028060735702514647, 0.02800326347351074, 0.02794419288635254, 0.02799827194213867, 0.02790179252624512, 0.027799968719482423, 0.028021120071411134, 0.027707456588745117, 0.02746339225769043, 0.027762271881103515, 0.027985984802246094, 0.02816067123413086, 0.027860960006713866, 0.027838464736938476, 0.02783776092529297, 0.027716287612915037, 0.027561056137084962, 0.027519071578979492, 0.02798409652709961, 0.027565919876098632, 0.027634431838989258, 0.027815168380737304, 0.028237855911254883, 0.028158655166625978, 0.02821651268005371, 0.02811382484436035, 0.02824799919128418, 0.028110336303710938, 0.028142080307006836, 0.028379135131835938]",tokens/s,35.40701258000924,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1849.87648,2718.892032,0.0,2340.421632,2284.9536,s,1,9.00453125,9.00453125,0.0,9.00453125,9.00453125,9.00453125,9.00453125,[9.00453125],,kWh,6.025863509175906e-05,6.639789652902034e-06,1.9503071158000207e-05,8.64014959026613e-05,,MB,1681.14176,3087.990784,0.0,2680.160256,2578.241536,s,10,0.8896338195800781,0.0889633819580078,0.0006379165192925978,0.08893408203125,0.08968330001831054,0.08981941032409667,0.08992829856872558,"[0.08995552062988281, 0.08916140747070313, 0.08871475219726563, 0.08959474945068359, 0.08895865631103515, 0.08818233489990235, 0.08890950775146485, 0.0896530532836914, 0.08775977325439453, 0.08874406433105468]",tokens/s,2877.58844555658,kWh,2.857336710314902e-06,3.15112401888477e-07,1.9002739042522327e-06,5.0727230164556125e-06,tokens/kWh,50465992.16427769,MB,1608.097792,3087.990784,0.0,2680.160256,2578.244096,s,10,20.728130615234377,2.0728130615234375,0.009648469638570125,2.071475952148438,2.089635302734375,2.0900453124999996,2.0903733203124997,"[2.07177734375, 2.089544189453125, 2.07070458984375, 2.06433447265625, 2.071174560546875, 2.057871826171875, 2.066839599609375, 2.072115234375, 2.090455322265625, 2.0733134765625]",tokens/s,30.39347887633313,kWh,6.107347214593199e-05,6.7359504432220745e-06,2.906802864814823e-05,9.68774512373023e-05,tokens/kWh,650306.1258876523,,s,630,20.725994392395044,0.03289840379745241,0.0006040979539939706,0.03274156761169433,0.03352360649108886,0.033829865646362305,0.03522202232360841,"[0.03391078567504883, 0.03322675323486328, 0.033157119750976564, 0.03295222473144531, 0.03280495834350586, 0.03286966323852539, 0.03305340957641602, 0.03284787368774414, 0.03264675140380859, 0.0331987190246582, 0.03305654525756836, 0.03350236892700195, 0.03312112045288086, 0.03334143829345703, 0.03304636764526367, 0.03334672164916992, 0.033120990753173825, 0.033399070739746094, 0.03311206436157227, 0.03293312072753906, 0.03267046356201172, 0.03283148956298828, 0.033006816864013674, 0.032993503570556644, 0.03315491104125977, 0.03318243026733399, 0.03274095916748047, 0.03259164810180664, 0.03274399948120117, 0.03290528106689453, 0.032780288696289066, 0.03299238586425781, 0.03324198532104492, 0.033023998260498046, 0.03297689437866211, 0.03268966293334961, 0.03272496032714844, 0.032928321838378904, 0.032597183227539066, 0.03246979141235352, 0.03255091094970703, 0.03275737762451172, 0.032858081817626957, 0.03244073486328125, 0.03245260620117187, 0.03250137710571289, 0.03249395370483398, 0.03257548904418945, 0.0328007698059082, 0.03266963195800781, 0.03275968170166016, 0.03257977676391602, 0.03296377563476562, 0.03281593704223633, 0.032571392059326174, 0.0328642578125, 0.032679935455322266, 0.03274342346191406, 0.03266121673583984, 0.03273961639404297, 0.0326409912109375, 0.03276297760009766, 0.032800800323486326, 0.03390486526489258, 0.0329268798828125, 0.03312646484375, 0.032860992431640625, 0.03285590362548828, 0.03315635299682617, 0.03359632110595703, 0.03264102554321289, 0.03289807891845703, 0.03265212631225586, 0.03247526550292969, 0.03264070510864258, 0.0326577262878418, 0.03250995254516602, 0.03252633666992188, 0.03288374328613281, 0.03273791885375977, 0.03268009567260742, 0.0329048957824707, 0.032898880004882815, 0.033427711486816405, 0.032776447296142576, 0.03263302230834961, 0.03269001770019531, 0.03268009567260742, 0.03259328079223633, 0.03299596786499023, 0.03271596908569336, 0.0327562255859375, 0.032786304473876954, 0.032774593353271486, 0.03277414321899414, 0.03315097427368164, 0.033592735290527344, 0.03351919937133789, 0.03363123321533203, 0.033587200164794925, 0.03380569458007812, 0.03322534561157227, 0.03356444931030273, 0.033503456115722655, 0.03348867034912109, 0.03352783966064453, 0.0335843505859375, 0.03368844985961914, 0.03360976028442383, 0.03361183929443359, 0.033673152923583985, 0.03383097457885742, 0.033912193298339846, 0.03394575881958008, 0.03359587097167969, 0.033478656768798826, 0.033501182556152344, 0.033288192749023435, 0.03321206283569336, 0.03320252990722656, 0.033345535278320314, 0.03335782241821289, 0.03324041748046875, 0.03325942230224609, 0.03320310211181641, 0.03302105712890625, 0.03383350372314453, 0.03401043319702148, 0.03332198333740234, 0.03310313415527344, 0.032860641479492185, 0.032694305419921875, 0.032615966796875, 0.03240137481689453, 0.03234812927246094, 0.032436992645263674, 0.0324826545715332, 0.0325495376586914, 0.03252019119262695, 0.03252169418334961, 0.03305635070800781, 0.0327806396484375, 0.03246960067749023, 0.03252604675292969, 0.0324918098449707, 0.0324505615234375, 0.032452190399169925, 0.03240796661376953, 0.032451744079589846, 0.0326910400390625, 0.03258777618408203, 0.032606208801269534, 0.03246031951904297, 0.03263126373291016, 0.032546241760253905, 0.03250028610229492, 0.03299903869628906, 0.032564990997314455, 0.032594558715820315, 0.032605438232421874, 0.032733951568603516, 0.03262464141845703, 0.035332096099853515, 0.03325478363037109, 0.03284646224975586, 0.03282080078125, 0.03300396728515625, 0.032825439453125, 0.03276995086669922, 0.032933887481689454, 0.03290726470947265, 0.033279647827148436, 0.03294857788085938, 0.03272451019287109, 0.03262694549560547, 0.03278985595703125, 0.032797119140625, 0.03296505737304688, 0.03313663864135742, 0.03319398498535156, 0.034326526641845705, 0.03326464080810547, 0.033030654907226564, 0.03317958450317383, 0.03311251068115234, 0.03296598434448242, 0.03322048187255859, 0.032775039672851565, 0.03254191970825195, 0.03382851028442383, 0.033089664459228514, 0.032976192474365236, 0.03282422256469727, 0.0330313606262207, 0.033027103424072266, 0.03274729537963867, 0.03267718505859375, 0.03288899230957031, 0.03335427093505859, 0.03332710266113281, 0.033255390167236325, 0.033135967254638674, 0.0326519660949707, 0.0326195182800293, 0.03261452865600586, 0.03269472122192383, 0.03283603286743164, 0.03280607986450195, 0.03278316879272461, 0.03279052734375, 0.03252364730834961, 0.032635520935058594, 0.03253184127807617, 0.032844417572021486, 0.03274342346191406, 0.03253247833251953, 0.03252633666992188, 0.03240959930419922, 0.032393280029296874, 0.032274337768554685, 0.03232771301269531, 0.032382942199707034, 0.03240963363647461, 0.03232972717285156, 0.032399360656738284, 0.032296031951904294, 0.032914302825927735, 0.03238095855712891, 0.03232352066040039, 0.03255222320556641, 0.03250460815429688, 0.03235017776489258, 0.03224291229248047, 0.03221731185913086, 0.03230985641479492, 0.03226236724853516, 0.03233097457885742, 0.03243679809570312, 0.03239916610717773, 0.03332425689697266, 0.035050464630126954, 0.03259187316894531, 0.03410940933227539, 0.03476793670654297, 0.033292831420898436, 0.033083839416503905, 0.0328960952758789, 0.03258591842651367, 0.03252707290649414, 0.03254249572753906, 0.032319713592529296, 0.03229875183105469, 0.03421798324584961, 0.03332067108154297, 0.033328670501708985, 0.03320908737182617, 0.033067230224609376, 0.035292095184326175, 0.03331520080566406, 0.033166881561279296, 0.03729913711547852, 0.03336601638793945, 0.03299123382568359, 0.03288063812255859, 0.033043617248535155, 0.033090400695800784, 0.03321241760253906, 0.03315670394897461, 0.03332134246826172, 0.032989215850830075, 0.03292335891723633, 0.03283123016357422, 0.03273168182373047, 0.032718849182128903, 0.03316121673583984, 0.03372963333129883, 0.03303081512451172, 0.03277657699584961, 0.03246681594848633, 0.03242598342895508, 0.032550048828125, 0.032639198303222657, 0.03249216079711914, 0.03231948852539063, 0.032382976531982424, 0.03240758514404297, 0.03259142303466797, 0.032815521240234374, 0.0325909423828125, 0.03252054214477539, 0.03274095916748047, 0.03244892883300781, 0.03239123153686523, 0.0324405746459961, 0.0324508171081543, 0.03258163070678711, 0.03249356842041016, 0.03249103927612305, 0.03250569534301758, 0.03247679901123047, 0.03245948791503906, 0.032436511993408204, 0.03265945434570312, 0.03270054244995117, 0.032545726776123045, 0.03265427017211914, 0.032511329650878905, 0.03255363082885742, 0.03241958236694336, 0.03242348861694336, 0.032464897155761716, 0.03238947296142578, 0.03245820617675781, 0.032489952087402345, 0.03240796661376953, 0.03380009460449219, 0.03256326293945312, 0.03246371078491211, 0.03247292709350586, 0.03225600051879883, 0.032855777740478515, 0.032374401092529294, 0.032416671752929685, 0.03243590545654297, 0.03235539245605469, 0.032449535369873043, 0.03245260620117187, 0.03253833770751953, 0.03277635192871094, 0.03241936111450195, 0.032604766845703126, 0.03249887847900391, 0.03254278564453125, 0.032436992645263674, 0.0324136962890625, 0.032454654693603514, 0.03242393493652344, 0.03237273788452148, 0.032509441375732424, 0.03233958435058594, 0.03238374328613281, 0.03237286376953125, 0.03240700912475586, 0.032330272674560546, 0.03263071823120117, 0.03241580963134766, 0.03240345764160156, 0.03244236755371094, 0.03247228622436524, 0.03237968063354492, 0.03227961730957031, 0.03226310348510742, 0.03239430236816406, 0.0323306884765625, 0.03252044677734375, 0.03261824035644531, 0.03264921569824219, 0.03265910339355469, 0.03304483032226563, 0.032505855560302735, 0.03281203079223633, 0.03293286514282227, 0.03257078552246094, 0.032444095611572264, 0.032401344299316404, 0.03283561706542969, 0.03294713592529297, 0.03283967971801758, 0.032892929077148435, 0.03311964797973633, 0.03313724899291992, 0.033148929595947264, 0.033377857208251954, 0.033392574310302736, 0.03339049530029297, 0.03338016128540039, 0.033879905700683596, 0.03338940811157227, 0.034476318359375, 0.03386342239379883, 0.03365340805053711, 0.03350137710571289, 0.033455745697021484, 0.03345222473144531, 0.033320960998535154, 0.03315302276611328, 0.03316934585571289, 0.03328006362915039, 0.03302377700805664, 0.03273513412475586, 0.03273734283447265, 0.03271295928955078, 0.03305574417114258, 0.03297766494750977, 0.0327723503112793, 0.03354592132568359, 0.032678112030029294, 0.0326036491394043, 0.03259852981567383, 0.032559200286865236, 0.032421886444091795, 0.03322415924072265, 0.03259555053710937, 0.03257644653320312, 0.03352313613891601, 0.03259174346923828, 0.03250246429443359, 0.03250783920288086, 0.03239686584472656, 0.032567806243896484, 0.03248271942138672, 0.032567935943603514, 0.03267977523803711, 0.03277747344970703, 0.03235638427734375, 0.03248169708251953, 0.032438720703125, 0.03365881729125977, 0.03260124969482422, 0.03267059326171875, 0.0323480339050293, 0.03249987030029297, 0.03240095901489258, 0.03238662338256836, 0.03260505676269531, 0.032366497039794925, 0.032245857238769535, 0.032464897155761716, 0.032389118194580076, 0.03249331283569336, 0.03266739273071289, 0.03250985717773437, 0.0327125129699707, 0.03291830444335937, 0.03276380920410156, 0.03341648101806641, 0.03260054397583008, 0.03258403015136719, 0.032405502319335935, 0.03249497604370117, 0.03240204620361328, 0.03382067108154297, 0.033089534759521484, 0.032968704223632815, 0.03341110229492188, 0.03280892944335938, 0.03270995330810547, 0.03275846481323242, 0.03241094589233398, 0.032932350158691406, 0.03248761749267578, 0.032824638366699216, 0.032412353515625, 0.034361278533935544, 0.034791358947753905, 0.03299955368041992, 0.03264921569824219, 0.032545982360839845, 0.032583873748779295, 0.032418113708496094, 0.03230752182006836, 0.03322841644287109, 0.033591808319091795, 0.03298060989379883, 0.03248940658569336, 0.03237532806396484, 0.03242793655395508, 0.03245011138916016, 0.03243244934082031, 0.03232153701782227, 0.03246694564819336, 0.03239116668701172, 0.03238019180297851, 0.03324796676635742, 0.03245225524902344, 0.03353225708007813, 0.03264921569824219, 0.03254441452026367, 0.0326429443359375, 0.032342720031738284, 0.032331775665283204, 0.032407329559326174, 0.03241779327392578, 0.03254003143310547, 0.032471519470214844, 0.03261455917358398, 0.03253452682495117, 0.032489471435546875, 0.032694271087646484, 0.033007614135742186, 0.03308163070678711, 0.0338135986328125, 0.033187614440917966, 0.03464278411865234, 0.03335168075561523, 0.03492659378051758, 0.03305654525756836, 0.03292550277709961, 0.032938400268554685, 0.03266355133056641, 0.032780288696289066, 0.03307929611206055, 0.03290454483032226, 0.03281167984008789, 0.03365513610839844, 0.03338240051269531, 0.03315670394897461, 0.033081470489501955, 0.03316559982299805, 0.03325747299194336, 0.03277414321899414, 0.03267583847045898, 0.03255820846557617, 0.03262704086303711, 0.03240399932861328, 0.03248096084594727, 0.03240342330932617, 0.03255331039428711, 0.032894977569580076, 0.03241116714477539, 0.0324510383605957, 0.03259324645996094, 0.03257001495361328, 0.0324692497253418, 0.03248041534423828, 0.03254726409912109, 0.032495777130126954, 0.03240345764160156, 0.03247462463378906, 0.03287196731567383, 0.032791519165039064, 0.03696995162963867, 0.03336684799194336, 0.032719806671142576, 0.03431884765625, 0.037418880462646485, 0.035735198974609375, 0.03276582336425781, 0.032581600189208984, 0.03284182357788086, 0.0325599365234375, 0.03237897491455078, 0.034969505310058595, 0.03252016067504883, 0.03602630233764648, 0.03345119857788086, 0.033127201080322265, 0.033746528625488284, 0.034873886108398436, 0.03324310302734375, 0.034187297821044925, 0.03331651306152344, 0.033142974853515625, 0.033153182983398435, 0.03280486297607422, 0.03282124710083008, 0.03273523330688476, 0.03293523025512695, 0.03280486297607422, 0.03270316696166992, 0.032833534240722655, 0.0328089599609375, 0.03274710464477539, 0.03287196731567383, 0.03301635360717774, 0.0329315185546875, 0.033161407470703126, 0.03335500717163086, 0.03274399948120117, 0.0326445426940918, 0.03266579055786133, 0.03236924743652344, 0.03240342330932617, 0.03254419326782226, 0.032459327697753906, 0.0323768310546875, 0.03243971252441406, 0.03225660705566406, 0.032738815307617186, 0.032761760711669925, 0.03236924743652344, 0.03237887954711914, 0.03230499267578125, 0.032415008544921874, 0.032502655029296876, 0.03237017440795899, 0.03265283203125, 0.03269311904907227, 0.03260563278198242, 0.032524959564208984, 0.03267583847045898, 0.03252598571777344, 0.033162593841552734, 0.03270892715454102, 0.0327248649597168, 0.0325681266784668, 0.03252326583862305, 0.03310601425170898, 0.0326253776550293, 0.03286220932006836, 0.033185375213623046, 0.03304713439941406, 0.032806110382080075, 0.032742176055908206, 0.0328559684753418, 0.03283158493041992, 0.03360268783569336, 0.03317644882202148, 0.034852863311767575, 0.03409305572509766, 0.033570816040039066, 0.03341686248779297, 0.03346857452392578, 0.033771713256835936, 0.033454078674316406, 0.03339619064331055, 0.033509151458740234, 0.033579071044921874, 0.03346707153320312, 0.03362815856933594, 0.033247230529785156, 0.03332505416870117, 0.03343539047241211, 0.03372057723999024, 0.032813152313232424, 0.03253647994995117, 0.032456703186035156, 0.032401214599609374, 0.03235615921020508, 0.0323131217956543]",tokens/s,30.396611524278214,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,872.189952,556.72832,0.0,178.25792,176.190464,s,1,7.68558203125,7.68558203125,0.0,7.68558203125,7.68558203125,7.68558203125,7.68558203125,[7.68558203125],,kWh,2.0834734899998844e-05,2.2905349177520978e-06,6.4622273920067475e-06,2.958749720975769e-05,,MB,1168.748544,661.58592,0.0,253.755392,220.750336,s,12,0.183656286239624,0.01530469051996867,0.00019884486284842023,0.015267151832580567,0.015476128292083741,0.015652350568771362,0.015812454061508177,"[0.015237055778503417, 0.015360063552856445, 0.015313632011413574, 0.01585247993469238, 0.015135199546813964, 0.015297247886657714, 0.015488608360290528, 0.015225055694580079, 0.015363807678222657, 0.015139295578002929, 0.01512771224975586, 0.015116127967834472]",tokens/s,16726.8981797434,kWh,4.6199605640822427e-07,5.0950060289420976e-08,3.079489981170813e-07,8.208951148147265e-07,tokens/kWh,311854700.2899127,MB,1202.958336,676.265984,0.0,268.435456,220.752896,s,12,10.052539978027344,0.837711664835612,0.008306164220725462,0.8395189514160156,0.8449926452636718,0.8473671356201171,0.8496844134521484,"[0.8449517211914063, 0.8366310424804687, 0.84122119140625, 0.8502637329101562, 0.8379326782226563, 0.841105224609375, 0.829469482421875, 0.8449971923828125, 0.8243952026367187, 0.8356583251953125, 0.821886474609375, 0.8440277099609375]",tokens/s,75.20487375851783,kWh,2.4577269508523106e-05,2.7104509462638155e-06,9.10293816021661e-06,3.6390658615003534e-05,tokens/kWh,1731213.514614041,,s,756,10.046389430999769,0.013288874908729837,0.0003385731659422532,0.013247408390045165,0.01360647964477539,0.01367526388168335,0.014316770839691167,"[0.012775808334350587, 0.01313587188720703, 0.013276576042175293, 0.013025823593139649, 0.013002816200256348, 0.012983519554138184, 0.012978976249694824, 0.013080351829528809, 0.013222111701965331, 0.013008895874023438, 0.01298588752746582, 0.012978655815124512, 0.012959744453430176, 0.012958815574645996, 0.013015968322753906, 0.013074432373046875, 0.013033727645874023, 0.013094464302062988, 0.013131967544555664, 0.013303808212280274, 0.013328384399414063, 0.013404159545898438, 0.013449024200439454, 0.013556991577148438, 0.01385478401184082, 0.013627424240112304, 0.013599424362182618, 0.013629311561584473, 0.013701343536376953, 0.013651424407958985, 0.013652576446533202, 0.013563936233520507, 0.01362723159790039, 0.013964703559875488, 0.013586688041687011, 0.013632160186767578, 0.013633343696594238, 0.013619199752807617, 0.013561856269836426, 0.013598719596862792, 0.013543231964111328, 0.013719167709350586, 0.0142741756439209, 0.013514880180358887, 0.013534015655517578, 0.01349555206298828, 0.013529855728149414, 0.01353932762145996, 0.013518848419189454, 0.013431072235107422, 0.013487199783325195, 0.013378175735473632, 0.013453311920166015, 0.013424639701843261, 0.01339187240600586, 0.013451264381408692, 0.013471903800964356, 0.01358240032196045, 0.013481760025024414, 0.013641728401184081, 0.013444735527038574, 0.013451647758483887, 0.013399231910705566, 0.013090944290161133, 0.013522687911987305, 0.013469696044921875, 0.013444671630859375, 0.013418944358825683, 0.013432479858398438, 0.014053728103637695, 0.014391039848327637, 0.013562111854553223, 0.013399200439453125, 0.013398112297058106, 0.013384063720703126, 0.013654175758361817, 0.013381535530090333, 0.013387999534606933, 0.013353055953979492, 0.013238271713256837, 0.01323209571838379, 0.013024479866027831, 0.013020064353942871, 0.01313980770111084, 0.013219903945922852, 0.01313587188720703, 0.013254015922546387, 0.01299078369140625, 0.013058367729187012, 0.01305190372467041, 0.013246015548706054, 0.013338784217834472, 0.014383392333984375, 0.013547616004943848, 0.0130600004196167, 0.013125727653503418, 0.012982175827026368, 0.013030688285827636, 0.01408073616027832, 0.01304371166229248, 0.013029024124145508, 0.013369695663452148, 0.013107168197631835, 0.013086784362792968, 0.013041631698608398, 0.013000639915466308, 0.013066304206848145, 0.013372703552246094, 0.013110239982604981, 0.013144864082336425, 0.012975071907043457, 0.01293875217437744, 0.0129684476852417, 0.012979455947875976, 0.012888992309570312, 0.012938816070556641, 0.013052191734313965, 0.013059616088867188, 0.013057696342468261, 0.01307875156402588, 0.013123744010925293, 0.013519295692443848, 0.013383071899414062, 0.013422623634338379, 0.013400927543640136, 0.013414143562316894, 0.013178879737854005, 0.013540351867675781, 0.013551520347595216, 0.013635680198669434, 0.013719008445739747, 0.01344313621520996, 0.013518655776977539, 0.01355020809173584, 0.013637120246887208, 0.0135316801071167, 0.013554816246032714, 0.013605759620666505, 0.013637439727783203, 0.01346726417541504, 0.013435680389404297, 0.01341004753112793, 0.013452447891235352, 0.013441408157348633, 0.013347264289855956, 0.013545536041259765, 0.013455360412597657, 0.013506239891052246, 0.013443391799926757, 0.013434816360473632, 0.013482048034667968, 0.013410304069519043, 0.013249919891357423, 0.013355392456054688, 0.013310208320617675, 0.013357376098632813, 0.013292863845825195, 0.013318528175354004, 0.01324614429473877, 0.013295743942260742, 0.01365830421447754, 0.013312000274658203, 0.013303808212280274, 0.013228032112121582, 0.013467647552490235, 0.013735263824462891, 0.013191840171813966, 0.013254783630371093, 0.013221504211425782, 0.013078656196594238, 0.013119615554809571, 0.01309017562866211, 0.01309887981414795, 0.013019904136657715, 0.013038880348205566, 0.012986592292785645, 0.013083135604858399, 0.013008607864379882, 0.01300211238861084, 0.013060959815979004, 0.013028639793395996, 0.013062687873840332, 0.013109567642211915, 0.013148287773132324, 0.013252351760864258, 0.013328351974487306, 0.013588576316833497, 0.013392928123474122, 0.013422911643981933, 0.013099616050720214, 0.01356060791015625, 0.013556639671325683, 0.013639552116394043, 0.013594752311706542, 0.013704575538635255, 0.013659839630126952, 0.013688960075378419, 0.013689727783203124, 0.013621536254882813, 0.013638943672180175, 0.01359648036956787, 0.013672351837158203, 0.013607680320739746, 0.013729727745056153, 0.013684000015258789, 0.013652000427246094, 0.013623680114746094, 0.013607199668884278, 0.013639039993286133, 0.013522848129272461, 0.013559871673583985, 0.01358505630493164, 0.013529088020324707, 0.013469696044921875, 0.01357430362701416, 0.013541215896606446, 0.013555328369140624, 0.01348646354675293, 0.013510656356811524, 0.013590527534484862, 0.013602144241333007, 0.013531807899475098, 0.013479552268981933, 0.013541343688964844, 0.013576383590698243, 0.013537504196166993, 0.01350169563293457, 0.013512479782104491, 0.013497311592102051, 0.0135863037109375, 0.013506015777587891, 0.013515423774719239, 0.013438752174377441, 0.013316351890563965, 0.014213088035583497, 0.01339187240600586, 0.013337984085083007, 0.013292415618896485, 0.013389887809753418, 0.01361411190032959, 0.013421567916870117, 0.013330080032348634, 0.013252608299255371, 0.013444512367248536, 0.013367136001586915, 0.013283295631408691, 0.0131878719329834, 0.0130600004196167, 0.01296985626220703, 0.01298044776916504, 0.012924032211303711, 0.01299135971069336, 0.013043775558471679, 0.01372332763671875, 0.013128543853759765, 0.013276896476745605, 0.013135744094848632, 0.013136320114135742, 0.013123519897460937, 0.013042079925537109, 0.013112607955932618, 0.013081151962280274, 0.014055328369140625, 0.014036543846130371, 0.01401039981842041, 0.01385923194885254, 0.013256640434265136, 0.013236288070678712, 0.013432607650756836, 0.013301759719848634, 0.013058143615722656, 0.013056127548217774, 0.013057760238647461, 0.01303171157836914, 0.012904255867004395, 0.013010463714599609, 0.013003423690795898, 0.013154399871826173, 0.013152159690856934, 0.013103296279907226, 0.01310700798034668, 0.013109248161315918, 0.013091103553771972, 0.013102815628051758, 0.013098272323608398, 0.012952287673950196, 0.013042943954467773, 0.012874496459960938, 0.012936223983764648, 0.012901344299316407, 0.012996512413024902, 0.013021280288696289, 0.013238143920898438, 0.013226112365722657, 0.013223039627075195, 0.013339455604553224, 0.014368831634521484, 0.01600307273864746, 0.013165984153747558, 0.014186431884765625, 0.013122048377990723, 0.013144224166870118, 0.013037599563598633, 0.01313100814819336, 0.013218624114990234, 0.013565855979919434, 0.013295904159545898, 0.013117152214050293, 0.01318057632446289, 0.013017375946044922, 0.013080351829528809, 0.01295798397064209, 0.013065279960632325, 0.013257823944091796, 0.014761183738708496, 0.012696864128112793, 0.013700063705444337, 0.01628927993774414, 0.014596416473388671, 0.013072447776794434, 0.013100864410400391, 0.013045887947082519, 0.01311574363708496, 0.013001952171325683, 0.012971776008605957, 0.012985055923461913, 0.012948703765869141, 0.012935263633728027, 0.012966303825378419, 0.01298185634613037, 0.013535008430480958, 0.013042559623718262, 0.012975584030151368, 0.013168224334716797, 0.01330681610107422, 0.013307904243469238, 0.013342720031738281, 0.013447168350219727, 0.013436351776123047, 0.013498751640319825, 0.013446784019470215, 0.013426495552062987, 0.013443424224853515, 0.013433279991149903, 0.013428640365600587, 0.013531200408935547, 0.013430879592895508, 0.013437855720520019, 0.013576640129089355, 0.013459936141967773, 0.013510751724243163, 0.013418496131896973, 0.013414719581604003, 0.013301440238952637, 0.013277503967285156, 0.01336473560333252, 0.013508864402770996, 0.013315327644348144, 0.013300415992736816, 0.013370656013488769, 0.013508576393127442, 0.013384448051452637, 0.013393695831298829, 0.0133820161819458, 0.013213312149047852, 0.013289695739746093, 0.013191167831420898, 0.01313372802734375, 0.013123680114746093, 0.013141311645507813, 0.013171392440795899, 0.013220159530639649, 0.013184351921081543, 0.013175168037414552, 0.013191136360168457, 0.01344921588897705, 0.013359359741210938, 0.013215680122375489, 0.0129518404006958, 0.013181280136108399, 0.013251839637756347, 0.013316512107849121, 0.013230463981628419, 0.013196576118469238, 0.013205408096313476, 0.013213536262512208, 0.0132424955368042, 0.01334768009185791, 0.013469696044921875, 0.013423999786376954, 0.013202207565307617, 0.013109087944030762, 0.013391263961791992, 0.01301302433013916, 0.013046336174011231, 0.013072640419006347, 0.013067839622497558, 0.013266847610473633, 0.013244064331054687, 0.013224224090576171, 0.013160896301269532, 0.013161760330200195, 0.013095552444458008, 0.013076031684875488, 0.01302950382232666, 0.013029696464538575, 0.012976127624511719, 0.013025152206420898, 0.013187199592590333, 0.013217791557312012, 0.013191167831420898, 0.01315782356262207, 0.01318057632446289, 0.013073311805725098, 0.013010944366455078, 0.012978400230407715, 0.01306595230102539, 0.013051039695739746, 0.013203680038452148, 0.013238975524902344, 0.013189120292663574, 0.013148127555847168, 0.0132892484664917, 0.013407839775085448, 0.01337388801574707, 0.013316320419311524, 0.013172736167907715, 0.013091936111450195, 0.013142815589904785, 0.013054047584533691, 0.013008288383483887, 0.01303382396697998, 0.013017375946044922, 0.013047807693481446, 0.0131146240234375, 0.013152064323425293, 0.013123871803283692, 0.013154815673828125, 0.013125791549682617, 0.013078720092773437, 0.013147616386413574, 0.01319974422454834, 0.01366972827911377, 0.013640064239501953, 0.013629728317260743, 0.01358233642578125, 0.013567551612854004, 0.013612480163574218, 0.013717696189880372, 0.013714048385620117, 0.013611200332641602, 0.013584159851074218, 0.01359488010406494, 0.013660223960876464, 0.013567904472351074, 0.013658111572265624, 0.013539232254028321, 0.0135578556060791, 0.01349955177307129, 0.013773344039916991, 0.013476384162902833, 0.013501728057861328, 0.013446751594543458, 0.013444095611572266, 0.01346127986907959, 0.013399359703063964, 0.01347049617767334, 0.013471776008605958, 0.013428640365600587, 0.013381695747375489, 0.013285759925842286, 0.01363491153717041, 0.01330735969543457, 0.01323209571838379, 0.013169792175292969, 0.013254176139831543, 0.013246432304382324, 0.013162752151489258, 0.013115391731262208, 0.013130784034729004, 0.013097951889038086, 0.013119615554809571, 0.013100768089294433, 0.013135711669921874, 0.013054047584533691, 0.013061440467834473, 0.013092896461486817, 0.013046976089477538, 0.012936800003051759, 0.013055904388427735, 0.01332038402557373, 0.013328384399414063, 0.013315296173095703, 0.0133721923828125, 0.013420063972473145, 0.0133853120803833, 0.013626239776611327, 0.013479935646057128, 0.013478976249694824, 0.013825087547302247, 0.013451295852661133, 0.013438816070556641, 0.013457375526428223, 0.013441151618957519, 0.013008864402770996, 0.013377568244934083, 0.01333801555633545, 0.013374048233032227, 0.013344896316528321, 0.013323583602905274, 0.013248384475708008, 0.013365951538085938, 0.013294624328613282, 0.013278240203857422, 0.013303359985351563, 0.013391231536865235, 0.013433088302612304, 0.013424960136413575, 0.013269375801086426, 0.013115391731262208, 0.013107263565063476, 0.013172032356262207, 0.013052096366882324, 0.013165056228637695, 0.013103103637695313, 0.01307852840423584, 0.013033760070800782, 0.01305519962310791, 0.013036031723022461, 0.013031328201293945, 0.012977663993835448, 0.012951807975769043, 0.012978528022766113, 0.012940735816955566, 0.012908384323120118, 0.01291324806213379, 0.012900639533996582, 0.012933247566223145, 0.013004511833190917, 0.012951423645019531, 0.013006560325622559, 0.012990880012512206, 0.012972031593322754, 0.012965696334838867, 0.012949503898620606, 0.012906304359436035, 0.01294979190826416, 0.012903743743896485, 0.012951807975769043, 0.012976672172546387, 0.012953599929809571, 0.012982272148132324, 0.01315225601196289, 0.012998656272888183, 0.013061951637268067, 0.013019328117370605, 0.013032671928405761, 0.013007648468017578, 0.013020544052124024, 0.013020992279052734, 0.01300153636932373, 0.012976127624511719, 0.01303337574005127, 0.012959839820861817, 0.012939040184020995, 0.012947456359863281, 0.013062687873840332, 0.012665087699890137, 0.013041152000427245, 0.012989695549011231, 0.013134847640991211, 0.013049023628234863, 0.013032256126403808, 0.012985952377319336, 0.012988096237182617, 0.013060383796691895, 0.01297862434387207, 0.012959839820861817, 0.012944767951965332, 0.012968640327453613, 0.012922719955444336, 0.012992511749267579, 0.01290060806274414, 0.01308249568939209, 0.012953472137451173, 0.012959039688110352, 0.013012672424316406, 0.012915712356567383, 0.013067584037780761, 0.013324992179870605, 0.013424768447875976, 0.01343609619140625, 0.013506367683410644, 0.013501343727111816, 0.01358028793334961, 0.013457056045532226, 0.013500384330749512, 0.01356326389312744, 0.013537823677062988, 0.013566240310668946, 0.01350496006011963, 0.013650912284851074, 0.013570624351501465, 0.013552127838134765, 0.013460607528686523, 0.013437503814697266, 0.013432127952575684, 0.013512543678283691, 0.013541631698608399, 0.013657983779907226, 0.013447392463684081, 0.013396479606628419, 0.013478079795837402, 0.013446975708007812, 0.013500415802001953, 0.013493824005126952, 0.013725664138793945, 0.013451807975769043, 0.0133602237701416, 0.01326576042175293, 0.013274368286132812, 0.013194144248962402, 0.013201248168945313, 0.013153440475463867, 0.013146976470947265, 0.013091872215270996, 0.013138912200927734, 0.013082431793212891, 0.012992704391479492, 0.013029376029968261, 0.012685312271118163, 0.012984319686889649, 0.012913760185241699, 0.012929951667785645, 0.01289241600036621, 0.012965215682983398, 0.012930784225463867, 0.012958047866821289, 0.012879743576049804, 0.01287548828125, 0.01296678352355957, 0.012991968154907227, 0.012927136421203612, 0.012953824043273925, 0.013031359672546386, 0.012949567794799804, 0.012954784393310546, 0.01296895980834961, 0.013039487838745117, 0.012909695625305176, 0.012917183876037597, 0.01309331226348877, 0.012857343673706055, 0.01287168025970459, 0.012928031921386718, 0.013063136100769044, 0.012881024360656738, 0.012860447883605956, 0.012859552383422852, 0.012958880424499511, 0.012847647666931153, 0.012918784141540527, 0.012890144348144532, 0.01291808032989502, 0.012965727806091308, 0.01294159984588623, 0.012964127540588379, 0.013041919708251953, 0.012985631942749024, 0.012945728302001953, 0.012939680099487304, 0.012907903671264648, 0.01315062427520752, 0.016723712921142577, 0.013594304084777832, 0.013100000381469726, 0.013029151916503907, 0.013027711868286133, 0.012992159843444824, 0.012984064102172852, 0.013037376403808594, 0.012985856056213378, 0.012948415756225585, 0.013020511627197265, 0.012959648132324218, 0.01297596836090088, 0.012946335792541504, 0.012969792366027832, 0.012975647926330566, 0.013073408126831054, 0.01328707218170166, 0.013286815643310548, 0.01304035186767578, 0.01278771209716797, 0.013139967918395995, 0.013100607872009278, 0.013107711791992188, 0.01305299186706543, 0.013032320022583008, 0.013086112022399902, 0.013121824264526367, 0.013062463760375977, 0.013013312339782715, 0.012998335838317871, 0.013000191688537598, 0.01297049617767334, 0.0129617919921875, 0.013021183967590331, 0.01306601619720459, 0.013153504371643067, 0.013424960136413575, 0.0134519681930542, 0.013462719917297363, 0.013384511947631836, 0.01349836826324463, 0.013414400100708008, 0.013413439750671387, 0.013446047782897949, 0.013586496353149414, 0.013549535751342774, 0.01358243179321289, 0.013538623809814453, 0.013546303749084472, 0.013778719902038574, 0.013981696128845214, 0.013533184051513672, 0.013594688415527344, 0.013522879600524902, 0.013528512001037598, 0.013468416213989257, 0.013442879676818847, 0.013491264343261719, 0.013448127746582031, 0.0134782075881958, 0.013442272186279297, 0.013449695587158203, 0.013514431953430175, 0.013476160049438477, 0.013610527992248536, 0.013522720336914063, 0.013611488342285157, 0.013588704109191895, 0.013389823913574218, 0.01338368034362793, 0.013334527969360351, 0.013316096305847168, 0.01351593589782715, 0.013978495597839355, 0.013553631782531739, 0.013499391555786134, 0.013511679649353027, 0.013540512084960937, 0.013478752136230468, 0.013484031677246093, 0.013532608032226563, 0.013480031967163086]",tokens/s,75.25091528576823,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 1013, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 839, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 467, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 397, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 199, in _flash_attention_forward attn_output_unpad = flash_attn_varlen_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 1124, in flash_attn_varlen_func return FlashAttnVarlenFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 620, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_varlen_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 90, in _flash_attn_varlen_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.varlen_fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3798, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ self.model = DeciCoderModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ self.self_attn = DeciCoderAttention(config=config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ self._init_rope() File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1709, in __getattr__ raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4357.210112,4562.28864,0.0,4183.81824,4182.069248,s,1,10.031142578125,10.031142578125,0.0,10.031142578125,10.031142578125,10.031142578125,10.031142578125,[10.031142578125],,kWh,8.981634780416621e-05,9.900313537311436e-06,3.0006690671999584e-05,0.00012972335201347722,,MB,1635.110912,4715.380736,0.0,4307.550208,4281.174016,s,10,3.7298596801757813,0.3729859680175781,0.00239330107419596,0.3737974243164063,0.3747413604736328,0.37496203460693356,0.3751385739135742,"[0.3665159912109375, 0.373134521484375, 0.37361480712890627, 0.37194921875, 0.3744526062011719, 0.37398004150390624, 0.3743899230957031, 0.37194754028320315, 0.37518270874023435, 0.37469232177734374]",tokens/s,686.3528978332375,kWh,1.0785647206249867e-05,1.1891202052040404e-06,7.142733888785716e-06,1.9117501300239624e-05,tokens/kWh,13390871.32672465,MB,1641.34912,4730.0608,0.0,4322.230272,4281.176576,s,10,19.28433227539062,1.928433227539062,0.011992368182706773,1.9251874999999998,1.9393903198242188,1.9469975280761718,1.9530832946777343,"[1.9149078369140624, 1.92698876953125, 1.92338623046875, 1.9225906982421874, 1.954604736328125, 1.9357330322265625, 1.936275634765625, 1.917160888671875, 1.914984619140625, 1.9376998291015626]",tokens/s,32.66900772104845,kWh,5.615197611333324e-05,6.192956502939655e-06,3.712638485981421e-05,9.947131747608708e-05,tokens/kWh,633348.4023185398,,s,630,19.2810475769043,0.030604837423657612,0.00047886753596131467,0.030571392059326172,0.03109004878997803,0.03141498889923096,0.032441209793090835,"[0.030980224609375, 0.03203251266479492, 0.030491968154907227, 0.03043529510498047, 0.030155744552612305, 0.029895967483520507, 0.029991647720336915, 0.02984511947631836, 0.02998828887939453, 0.02998281669616699, 0.030333791732788086, 0.03010918426513672, 0.029904767990112303, 0.029932159423828125, 0.02993561553955078, 0.029888511657714844, 0.029808895111083984, 0.029858720779418944, 0.029913471221923827, 0.029925823211669922, 0.02991619110107422, 0.03002060890197754, 0.02979248046875, 0.029812223434448244, 0.03008336067199707, 0.029960031509399413, 0.030494688034057617, 0.030851264953613282, 0.030752767562866212, 0.029865983963012696, 0.029716320037841797, 0.029941375732421876, 0.029991455078125, 0.030666751861572264, 0.03087123107910156, 0.030623296737670898, 0.030740480422973632, 0.030692159652709963, 0.030547903060913085, 0.03059916877746582, 0.030748544692993166, 0.03073766326904297, 0.030724992752075197, 0.031192672729492187, 0.03087919998168945, 0.03069228744506836, 0.030691328048706053, 0.03086774444580078, 0.030506399154663084, 0.03044528007507324, 0.031002687454223632, 0.03039481544494629, 0.030324832916259765, 0.030405920028686525, 0.03050150489807129, 0.03039836883544922, 0.03145948791503906, 0.0306297607421875, 0.030623903274536134, 0.030642175674438478, 0.030765056610107422, 0.03038198471069336, 0.030250879287719728, 0.031472896575927736, 0.03066080093383789, 0.03020038414001465, 0.030064640045166017, 0.029894176483154296, 0.02993814468383789, 0.02995609664916992, 0.030144512176513674, 0.030275423049926756, 0.030224544525146484, 0.03330227279663086, 0.031366559982299806, 0.030351936340332033, 0.030357791900634767, 0.030521343231201172, 0.03013382339477539, 0.030124479293823243, 0.030926847457885744, 0.03182796859741211, 0.03020796775817871, 0.03027961540222168, 0.03036169624328613, 0.03032268714904785, 0.030871456146240234, 0.03058492851257324, 0.030527488708496094, 0.03046944046020508, 0.03063043212890625, 0.030556320190429687, 0.030488576889038086, 0.03047644805908203, 0.030477312088012694, 0.030415775299072266, 0.030492607116699218, 0.030623743057250977, 0.030308351516723633, 0.030211551666259766, 0.03023721694946289, 0.030224384307861327, 0.030126079559326172, 0.030107295989990235, 0.03030575942993164, 0.0307077751159668, 0.030720832824707032, 0.030958688735961915, 0.030741151809692384, 0.030822656631469728, 0.031106271743774415, 0.030659360885620116, 0.030672895431518556, 0.03079270362854004, 0.030775583267211915, 0.030697248458862306, 0.030645408630371095, 0.030664127349853517, 0.0307159366607666, 0.030533344268798827, 0.030816640853881836, 0.030671072006225587, 0.030465120315551757, 0.030569055557250976, 0.030961984634399413, 0.03091823959350586, 0.03139583969116211, 0.030961023330688477, 0.030832767486572266, 0.030923456192016602, 0.030781248092651366, 0.030819616317749023, 0.0307412166595459, 0.03116646385192871, 0.031282623291015624, 0.031146591186523437, 0.030942495346069337, 0.030919328689575195, 0.031174688339233397, 0.03079756736755371, 0.030621952056884765, 0.030519264221191406, 0.03012201690673828, 0.030345216751098632, 0.030650367736816408, 0.030514432907104493, 0.03059174346923828, 0.03064793586730957, 0.030500608444213866, 0.030572511672973632, 0.030760896682739257, 0.030728128433227538, 0.030620704650878905, 0.030571456909179687, 0.03068377685546875, 0.030576831817626954, 0.0303636474609375, 0.030338815689086914, 0.03027484893798828, 0.030423871994018553, 0.0304682559967041, 0.03020595169067383, 0.030212287902832032, 0.030537567138671874, 0.030271455764770507, 0.030252479553222657, 0.030217023849487306, 0.03018832015991211, 0.03001852798461914, 0.030306304931640625, 0.03065251159667969, 0.030684064865112305, 0.030301183700561524, 0.030056447982788087, 0.031082496643066407, 0.030319904327392577, 0.030570463180541994, 0.030544639587402344, 0.03018137550354004, 0.03020595169067383, 0.03023072052001953, 0.030229568481445312, 0.03028156852722168, 0.03020879936218262, 0.03007663917541504, 0.03012998390197754, 0.03012668800354004, 0.030116960525512694, 0.0300982723236084, 0.030994592666625978, 0.030203744888305663, 0.03014240074157715, 0.03010361671447754, 0.030300159454345704, 0.030287872314453124, 0.030330751419067385, 0.030406784057617188, 0.03057868766784668, 0.03057459259033203, 0.03054591941833496, 0.030612543106079103, 0.030519615173339842, 0.030345344543457033, 0.030310176849365236, 0.030374624252319335, 0.03025916862487793, 0.030263328552246095, 0.030088287353515625, 0.02989967918395996, 0.030024927139282228, 0.029929983139038087, 0.0300437126159668, 0.030179840087890625, 0.03015817642211914, 0.030156671524047853, 0.030063488006591796, 0.030627967834472657, 0.03082444763183594, 0.030676000595092772, 0.03101795196533203, 0.031362720489501957, 0.03086128044128418, 0.030634368896484375, 0.030638080596923828, 0.03074608039855957, 0.030751167297363283, 0.03071996879577637, 0.030844863891601564, 0.030663904190063478, 0.03050204849243164, 0.030801088333129882, 0.030702144622802734, 0.03069958305358887, 0.030482431411743165, 0.030631616592407228, 0.030527904510498048, 0.031088544845581056, 0.030871551513671876, 0.030928127288818358, 0.030746784210205078, 0.03092336082458496, 0.031916032791137694, 0.030916608810424805, 0.03051728057861328, 0.03045577621459961, 0.030139551162719727, 0.030161760330200196, 0.03016089630126953, 0.030466047286987305, 0.030416864395141602, 0.03008515167236328, 0.030105215072631836, 0.031170495986938475, 0.03076710319519043, 0.03077529525756836, 0.03081827163696289, 0.030838367462158203, 0.03057913589477539, 0.03080303955078125, 0.030837055206298827, 0.03072800064086914, 0.030980064392089845, 0.031043840408325196, 0.032731712341308596, 0.03123404884338379, 0.031024288177490235, 0.030978591918945312, 0.030910783767700196, 0.030942623138427734, 0.03110767936706543, 0.03132371139526367, 0.03171286392211914, 0.03108540725708008, 0.03105996894836426, 0.03127673530578613, 0.031053312301635744, 0.03102131271362305, 0.031013439178466797, 0.030910463333129884, 0.030801088333129882, 0.030795936584472657, 0.03078620719909668, 0.030770208358764647, 0.03063088035583496, 0.030627744674682617, 0.030799615859985353, 0.030781791687011718, 0.030994400024414063, 0.03106768035888672, 0.030800384521484377, 0.030631263732910155, 0.03054560089111328, 0.030546911239624025, 0.030600383758544923, 0.03056108856201172, 0.03061555290222168, 0.030843040466308595, 0.031022943496704102, 0.03102019119262695, 0.031019456863403322, 0.030871551513671876, 0.0308874568939209, 0.03128358459472656, 0.031135808944702147, 0.03143519973754883, 0.032239295959472655, 0.031116992950439452, 0.031194816589355467, 0.031220544815063478, 0.031295616149902346, 0.03125148773193359, 0.031181631088256837, 0.031105184555053712, 0.03197660827636719, 0.031029312133789063, 0.030978784561157227, 0.030494464874267577, 0.030472415924072266, 0.030656288146972657, 0.03041689682006836, 0.03039232063293457, 0.030427135467529298, 0.03063590431213379, 0.030719167709350587, 0.03091961669921875, 0.030453760147094725, 0.03058483123779297, 0.030402528762817384, 0.03029327964782715, 0.030042272567749023, 0.030124639511108397, 0.032833534240722655, 0.03083263969421387, 0.030471391677856445, 0.030605663299560548, 0.03054537582397461, 0.03154646492004395, 0.03168243217468262, 0.030641567230224608, 0.030558656692504883, 0.03061065673828125, 0.030635072708129884, 0.030545215606689453, 0.030517375946044922, 0.030693023681640626, 0.0307857608795166, 0.030691455841064454, 0.030443391799926757, 0.03058131217956543, 0.03043328094482422, 0.031942655563354495, 0.03143065643310547, 0.030620864868164063, 0.030548864364624024, 0.030631263732910155, 0.031103584289550783, 0.030691328048706053, 0.03073843193054199, 0.03077238464355469, 0.0337017936706543, 0.030593984603881835, 0.030398080825805664, 0.03027142333984375, 0.03059119987487793, 0.030714080810546874, 0.03059721565246582, 0.030358623504638672, 0.030136415481567383, 0.030253791809082033, 0.030490495681762694, 0.030519424438476564, 0.03056025505065918, 0.03057244873046875, 0.03064227294921875, 0.030718048095703124, 0.030790815353393553, 0.03065727996826172, 0.030684864044189453, 0.03122364807128906, 0.031167104721069337, 0.03077337646484375, 0.03077529525756836, 0.030670047760009766, 0.030870304107666016, 0.030658559799194338, 0.030770975112915037, 0.0306845760345459, 0.03146198463439941, 0.03080022430419922, 0.03063360023498535, 0.030839040756225587, 0.030678112030029295, 0.03072038459777832, 0.030702112197875976, 0.030697023391723633, 0.03058086395263672, 0.030512672424316406, 0.03046188735961914, 0.030376800537109373, 0.03060531234741211, 0.030498815536499024, 0.030453760147094725, 0.03054182434082031, 0.03054182434082031, 0.03062076759338379, 0.03071788787841797, 0.030667743682861327, 0.030646272659301758, 0.03150569534301758, 0.03084707260131836, 0.03153772735595703, 0.03095961570739746, 0.03073849678039551, 0.03075856018066406, 0.030723487854003906, 0.031075263977050783, 0.03075071907043457, 0.030801919937133788, 0.03083673667907715, 0.030914560317993164, 0.030727935791015626, 0.030748607635498047, 0.030746335983276366, 0.030972320556640624, 0.03082575988769531, 0.0307281608581543, 0.030546239852905274, 0.030521984100341796, 0.030633983612060548, 0.03076848030090332, 0.030519424438476564, 0.030351232528686524, 0.030360223770141602, 0.030220287322998047, 0.030271488189697264, 0.03039459228515625, 0.030418815612792968, 0.03034511947631836, 0.030275487899780275, 0.03159622383117676, 0.03112739181518555, 0.030824480056762697, 0.030556127548217772, 0.03055615997314453, 0.030608415603637695, 0.03067580795288086, 0.030517375946044922, 0.030348608016967774, 0.030470848083496094, 0.03037948799133301, 0.03028438377380371, 0.0300882568359375, 0.030098304748535157, 0.02998886489868164, 0.030192703247070313, 0.030415807723999023, 0.03039823913574219, 0.03045583915710449, 0.030552255630493165, 0.030603296279907228, 0.030604768753051757, 0.030877504348754883, 0.030679136276245116, 0.03176508712768555, 0.030993696212768554, 0.031185216903686523, 0.030517824172973634, 0.03043280029296875, 0.030218559265136717, 0.030212095260620117, 0.030477312088012694, 0.03021670341491699, 0.03003670310974121, 0.030039840698242188, 0.03007600021362305, 0.030019519805908203, 0.030216928482055663, 0.030044416427612304, 0.030072383880615235, 0.029931968688964843, 0.03051456069946289, 0.030558656692504883, 0.03048076820373535, 0.030469600677490234, 0.030828575134277343, 0.03043155288696289, 0.030508800506591795, 0.030555871963500975, 0.030550559997558593, 0.030552064895629883, 0.030537727355957032, 0.030390272140502928, 0.03032268714904785, 0.03057663917541504, 0.030561344146728515, 0.03049158477783203, 0.030462175369262694, 0.030314271926879882, 0.0301711368560791, 0.030209823608398436, 0.030123296737670897, 0.030192575454711913, 0.03015475273132324, 0.03027519989013672, 0.03087545585632324, 0.03027084732055664, 0.029981504440307616, 0.02986934471130371, 0.029797088623046874, 0.02985100746154785, 0.029880960464477538, 0.02996019172668457, 0.02980361557006836, 0.030181663513183594, 0.03006118392944336, 0.03022652816772461, 0.02992083168029785, 0.03081660842895508, 0.03001468849182129, 0.02992617607116699, 0.031170560836791993, 0.030115840911865234, 0.030031871795654298, 0.03057868766784668, 0.03061075210571289, 0.030519744873046876, 0.030380319595336915, 0.030486495971679687, 0.03159449577331543, 0.030564031600952148, 0.030669120788574217, 0.030537727355957032, 0.03047340774536133, 0.030571327209472657, 0.030204032897949217, 0.030365568161010742, 0.03048374366760254, 0.030567136764526368, 0.03041279983520508, 0.03032428741455078, 0.03153359985351562, 0.032523681640625, 0.03040086364746094, 0.030177536010742186, 0.02995574378967285, 0.030112064361572266, 0.030631744384765625, 0.03067046356201172, 0.03059939193725586, 0.030646560668945312, 0.030484479904174806, 0.0303636474609375, 0.03014575958251953, 0.030103519439697267, 0.030015775680541992, 0.03004070472717285, 0.029927488327026366, 0.02993062400817871, 0.03008905601501465, 0.030112640380859375, 0.030130176544189452, 0.03095759963989258, 0.030275360107421875, 0.030231039047241212, 0.030652095794677734, 0.031197183609008788, 0.03065376091003418, 0.03155795288085937, 0.031054304122924804, 0.030932159423828126, 0.03196182441711426, 0.031049728393554688, 0.030865535736083985, 0.030818336486816405, 0.03087753677368164, 0.030697471618652345, 0.03055820846557617, 0.03042323112487793, 0.03035308837890625, 0.030295551300048826, 0.03046022415161133, 0.030277280807495116, 0.030130847930908203, 0.030157983779907225, 0.030444095611572266, 0.030056032180786132, 0.030369983673095704, 0.030379743576049806, 0.030300960540771485, 0.030168767929077148, 0.03053984069824219, 0.030723840713500976, 0.03147120094299317, 0.03092777633666992, 0.03092889595031738, 0.030661855697631836, 0.030636415481567383, 0.03064259147644043, 0.030626848220825197, 0.03075584030151367, 0.030917695999145508, 0.030899103164672852, 0.03040787124633789, 0.03058070373535156, 0.03041267204284668, 0.03041302490234375, 0.030421152114868164, 0.030288480758666993, 0.030117055892944337, 0.030200639724731446, 0.033230846405029296, 0.03119923210144043, 0.03018956756591797, 0.030416223526000978, 0.030712480545043944, 0.03193459129333496, 0.030858207702636718, 0.03102943992614746, 0.030624479293823243, 0.030634143829345702, 0.030635871887207032, 0.03060326385498047, 0.03055820846557617, 0.03062579154968262, 0.030509023666381835, 0.030640159606933594, 0.0306889591217041, 0.03042950439453125, 0.03285606384277344, 0.03216790390014648]",tokens/s,32.67457317799694,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1141, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 944, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 677, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 500, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 1013, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 839, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 467, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 397, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 199, in _flash_attention_forward attn_output_unpad = flash_attn_varlen_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 1124, in flash_attn_varlen_func return FlashAttnVarlenFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 620, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_varlen_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 90, in _flash_attn_varlen_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.varlen_fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3798, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ self.model = DeciCoderModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ self.self_attn = DeciCoderAttention(config=config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ self._init_rope() File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1709, in __getattr__ raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4356.325376,5355.012096,0.0,4959.76448,4769.731072,s,1,11.4017490234375,11.4017490234375,0.0,11.4017490234375,11.4017490234375,11.4017490234375,11.4017490234375,[11.4017490234375],,kWh,0.00013048237477916112,1.438593316229139e-05,5.61742116059992e-05,0.0002010425195474517,,MB,1614.225408,5373.886464,0.0,4966.055936,4251.027456,s,10,30.64483813476562,3.0644838134765626,0.0036675148485847946,3.063621826171875,3.069316064453125,3.0702912109375,3.071071328125,"[3.0587421875, 3.061697021484375, 3.060749755859375, 3.063556396484375, 3.063182861328125, 3.063687255859375, 3.065486572265625, 3.067370361328125, 3.069099365234375, 3.071266357421875]",tokens/s,83.53772301690701,kWh,8.943002240083321e-05,9.863613228052827e-06,5.943954755160009e-05,0.00015873318318048612,tokens/kWh,1612769.2702345515,MB,1623.69536,5388.566528,0.0,4980.736,4251.030016,s,10,17.700633666992186,1.7700633666992185,0.011951188198258642,1.7681280517578126,1.7857709716796875,1.78604716796875,1.7862681249999999,"[1.7573978271484374, 1.7863233642578125, 1.7594388427734375, 1.769110595703125, 1.7671455078125, 1.7812459716796876, 1.78075, 1.7857095947265624, 1.7613912353515624, 1.7521207275390625]",tokens/s,35.591946133251284,kWh,5.177968356166691e-05,5.712164755921341e-06,3.399308274999995e-05,9.148493106758823e-05,tokens/kWh,688638.0004315266,,s,630,17.697427032470696,0.028091154019794774,0.00044194637218125394,0.028012063980102538,0.028631215286254882,0.028905468654632567,0.029727254791259778,"[0.02891379165649414, 0.02809449577331543, 0.02785273551940918, 0.027840736389160157, 0.027873056411743164, 0.028383264541625975, 0.02812460708618164, 0.02760918426513672, 0.027482559204101562, 0.02770534324645996, 0.027692256927490236, 0.0276582088470459, 0.027659072875976562, 0.02769430351257324, 0.027609567642211914, 0.027683135986328124, 0.0275230712890625, 0.027706880569458008, 0.02744576072692871, 0.02772332763671875, 0.027998336791992186, 0.028228960037231444, 0.02785804748535156, 0.027723487854003907, 0.027793119430541992, 0.027801279067993165, 0.02767283248901367, 0.028082656860351562, 0.027867136001586915, 0.02798371124267578, 0.027897151947021484, 0.027850719451904298, 0.027609983444213868, 0.027845792770385742, 0.0277922248840332, 0.028086271286010742, 0.02776425552368164, 0.028457439422607422, 0.028084224700927734, 0.02800230407714844, 0.028049407958984376, 0.028853376388549803, 0.0283284797668457, 0.02793916893005371, 0.02778892707824707, 0.027827808380126953, 0.027752256393432616, 0.027748992919921875, 0.027691551208496094, 0.027758464813232422, 0.028183807373046876, 0.027753087997436525, 0.02780985641479492, 0.027869056701660157, 0.02777510452270508, 0.02793267250061035, 0.027922367095947264, 0.02795667266845703, 0.02782476806640625, 0.027813695907592775, 0.02814188766479492, 0.027807680130004883, 0.027857952117919922, 0.02893212890625, 0.028485151290893556, 0.028633663177490234, 0.027979776382446288, 0.02846339225769043, 0.0281395206451416, 0.02815180778503418, 0.027888832092285157, 0.028161024093627928, 0.028318912506103515, 0.028555295944213868, 0.02916409683227539, 0.029050336837768555, 0.028887615203857422, 0.02832793617248535, 0.028341888427734375, 0.028437055587768555, 0.03174515151977539, 0.02854572868347168, 0.028524511337280272, 0.028352672576904298, 0.02841791915893555, 0.02835456085205078, 0.028272640228271483, 0.028082176208496092, 0.028717056274414062, 0.028194015502929687, 0.028404767990112306, 0.028171360015869142, 0.027986591339111327, 0.02813065528869629, 0.027932607650756835, 0.02813759994506836, 0.027994720458984376, 0.02803887939453125, 0.028079456329345703, 0.028302175521850586, 0.028263679504394533, 0.02794313621520996, 0.02788800048828125, 0.027924736022949218, 0.029859071731567384, 0.027860767364501954, 0.028046144485473632, 0.027957408905029298, 0.028078079223632812, 0.02777292823791504, 0.029067264556884766, 0.027924543380737306, 0.02824799919128418, 0.02815795135498047, 0.028106752395629882, 0.028688032150268553, 0.027980127334594727, 0.02788937568664551, 0.028080448150634766, 0.028110815048217774, 0.028233728408813476, 0.028241119384765624, 0.028221952438354493, 0.028178911209106445, 0.028578624725341797, 0.028363679885864256, 0.02889529609680176, 0.027979776382446288, 0.027797504425048827, 0.02793168067932129, 0.02800761604309082, 0.02787708854675293, 0.027920448303222656, 0.027716928482055665, 0.02784326362609863, 0.027768863677978515, 0.027829376220703125, 0.027705312728881836, 0.02788047981262207, 0.027766624450683595, 0.02780364799499512, 0.028837888717651368, 0.02819891166687012, 0.027962911605834962, 0.028819232940673827, 0.027800256729125977, 0.027590240478515625, 0.027742624282836914, 0.02776678466796875, 0.027869184494018553, 0.028011743545532226, 0.028205888748168945, 0.027889631271362306, 0.02776268768310547, 0.027844608306884764, 0.028138656616210938, 0.0276713924407959, 0.028010496139526365, 0.027757728576660156, 0.028791519165039064, 0.02778329658508301, 0.02773334312438965, 0.02774399948120117, 0.028031488418579102, 0.02765987205505371, 0.028775232315063477, 0.027822080612182616, 0.027731359481811522, 0.027783775329589845, 0.027726079940795897, 0.027663776397705078, 0.027732095718383788, 0.027752960205078125, 0.027639007568359374, 0.027714271545410157, 0.027803232192993164, 0.0277193603515625, 0.02822400093078613, 0.028035072326660155, 0.027743520736694335, 0.028711616516113283, 0.027727903366088866, 0.02776268768310547, 0.02779955291748047, 0.027830272674560546, 0.027696672439575194, 0.027822559356689452, 0.02776268768310547, 0.027813343048095702, 0.028882944107055664, 0.02792780876159668, 0.027771615982055665, 0.027654176712036134, 0.028198080062866213, 0.027670623779296875, 0.027709951400756837, 0.027369760513305665, 0.02842367935180664, 0.027683263778686525, 0.028046432495117186, 0.027644031524658202, 0.027810592651367188, 0.027657535552978514, 0.027867040634155273, 0.02771206474304199, 0.027621599197387697, 0.027637760162353517, 0.027621471405029296, 0.027719776153564454, 0.027795263290405273, 0.027897855758666993, 0.028192768096923827, 0.028078304290771485, 0.028222688674926756, 0.029454912185668945, 0.028165311813354493, 0.028102912902832032, 0.02813395118713379, 0.03039414405822754, 0.028463327407836914, 0.028282880783081055, 0.028079391479492188, 0.02828300857543945, 0.028143936157226563, 0.028454719543457033, 0.028395999908447267, 0.02828019142150879, 0.028281471252441407, 0.028419776916503905, 0.028193088531494142, 0.028223487854003908, 0.028331712722778322, 0.028254528045654297, 0.028102655410766602, 0.028251903533935547, 0.027816255569458007, 0.027825824737548827, 0.028032352447509765, 0.028013504028320313, 0.027824127197265625, 0.02815590476989746, 0.027962495803833008, 0.028038015365600587, 0.02787062454223633, 0.027761247634887694, 0.027799903869628908, 0.028177600860595703, 0.027949663162231447, 0.02779532814025879, 0.028861536026000976, 0.027757471084594726, 0.027672063827514647, 0.028893184661865235, 0.02804096031188965, 0.027818239212036133, 0.027549503326416015, 0.02763590431213379, 0.027712543487548827, 0.027618303298950195, 0.027487199783325197, 0.027613664627075197, 0.027631200790405274, 0.027694015502929686, 0.02769932746887207, 0.02761100769042969, 0.027546911239624022, 0.027595775604248047, 0.027571935653686524, 0.027672256469726562, 0.02752067184448242, 0.027574079513549805, 0.02766524887084961, 0.02769865608215332, 0.02757891273498535, 0.02769264030456543, 0.027590623855590822, 0.027783103942871094, 0.02808451271057129, 0.027777280807495117, 0.027693023681640627, 0.027731935501098634, 0.02806719970703125, 0.02794767951965332, 0.028611808776855468, 0.027798303604125975, 0.027725887298583985, 0.027930559158325194, 0.028256256103515624, 0.027977088928222656, 0.028012384414672853, 0.028053407669067384, 0.027884384155273438, 0.028081407546997072, 0.028401439666748046, 0.028492799758911135, 0.028358655929565428, 0.02847871971130371, 0.02901683235168457, 0.02899488067626953, 0.028805728912353515, 0.02839366340637207, 0.028420000076293944, 0.028188671112060547, 0.028719104766845704, 0.02873958396911621, 0.028622655868530272, 0.029308223724365236, 0.0283384952545166, 0.028434335708618166, 0.028303327560424803, 0.02832771110534668, 0.028350400924682616, 0.02822243118286133, 0.02804524803161621, 0.027739328384399416, 0.028654848098754883, 0.02865023994445801, 0.029125728607177735, 0.02809561538696289, 0.02805491256713867, 0.028078144073486327, 0.02784707260131836, 0.027811552047729494, 0.027813568115234374, 0.027914016723632813, 0.028040000915527344, 0.028327871322631835, 0.028016031265258787, 0.02797590446472168, 0.028124927520751953, 0.029278656005859376, 0.028863807678222657, 0.028034143447875977, 0.02821299171447754, 0.02889472007751465, 0.029096479415893554, 0.02796678352355957, 0.028449024200439453, 0.028123584747314453, 0.028194816589355468, 0.02824563217163086, 0.02827712059020996, 0.028674047470092775, 0.028508159637451173, 0.028220895767211915, 0.02783695983886719, 0.02800764846801758, 0.027996448516845702, 0.028170751571655273, 0.02792448043823242, 0.027895904541015624, 0.028034143447875977, 0.02949305534362793, 0.028134624481201173, 0.028165376663208008, 0.028636991500854494, 0.030010080337524413, 0.028289215087890625, 0.0283441276550293, 0.028223167419433592, 0.028038848876953126, 0.02791881561279297, 0.027716896057128907, 0.03000819206237793, 0.028020736694335937, 0.028078079223632812, 0.027809791564941407, 0.027641855239868163, 0.027797632217407227, 0.027848575592041017, 0.027898048400878905, 0.027797504425048827, 0.027954463958740235, 0.027801759719848634, 0.029243776321411133, 0.027975679397583008, 0.028692480087280273, 0.027920127868652344, 0.02894236755371094, 0.028241888046264648, 0.028228832244873048, 0.028179391860961914, 0.02814959907531738, 0.028023807525634766, 0.02810316848754883, 0.02838313674926758, 0.02950614356994629, 0.028439807891845702, 0.028461824417114256, 0.02812656021118164, 0.02822153663635254, 0.028150304794311524, 0.028332063674926758, 0.028278783798217775, 0.028182079315185547, 0.028453311920166015, 0.028145471572875978, 0.02837494468688965, 0.02873948860168457, 0.028363008499145508, 0.02808582305908203, 0.028239839553833006, 0.02825276756286621, 0.02800217628479004, 0.027832447052001955, 0.028016639709472657, 0.028630943298339845, 0.028277984619140627, 0.02803606414794922, 0.02798896026611328, 0.027826175689697266, 0.02797260856628418, 0.027801536560058595, 0.02824185562133789, 0.02801465606689453, 0.02879302406311035, 0.02797916793823242, 0.028219808578491212, 0.028251935958862304, 0.028311071395874025, 0.02832044792175293, 0.02893414306640625, 0.02856342315673828, 0.028114431381225585, 0.02801513671875, 0.027940927505493166, 0.02784454345703125, 0.028116640090942384, 0.0279553279876709, 0.028065343856811524, 0.02841667175292969, 0.028339744567871094, 0.028281312942504883, 0.02815590476989746, 0.02870681571960449, 0.02856550407409668, 0.028286527633666993, 0.0282891845703125, 0.028306751251220702, 0.028072864532470702, 0.028325183868408203, 0.02885273551940918, 0.028256256103515624, 0.02843846321105957, 0.02826576042175293, 0.028363487243652345, 0.028344575881958007, 0.028318880081176757, 0.028208831787109374, 0.028289119720458986, 0.028437568664550782, 0.02824678421020508, 0.02832067108154297, 0.028248064041137694, 0.028434528350830077, 0.02906105613708496, 0.02861609649658203, 0.02878121566772461, 0.028423423767089843, 0.028299871444702147, 0.028389312744140624, 0.028280704498291016, 0.028043615341186524, 0.028250080108642578, 0.02807360076904297, 0.02818035125732422, 0.027961280822753905, 0.029817567825317384, 0.028209184646606444, 0.028290912628173827, 0.028184288024902342, 0.028250463485717775, 0.029091680526733398, 0.02824355125427246, 0.028440351486206054, 0.028168735504150392, 0.02803936004638672, 0.028182783126831056, 0.029053760528564454, 0.028545568466186524, 0.02835433578491211, 0.028125823974609374, 0.028067840576171874, 0.028043264389038085, 0.027972896575927733, 0.02799603271484375, 0.02810761642456055, 0.027914239883422853, 0.027835424423217774, 0.027948127746582032, 0.02888096046447754, 0.027917631149291994, 0.02905753517150879, 0.03016214370727539, 0.028048160552978516, 0.02813132858276367, 0.02800390434265137, 0.027938400268554688, 0.028130144119262696, 0.028333568572998048, 0.028324287414550783, 0.0281146240234375, 0.02813171195983887, 0.027932575225830078, 0.028979583740234374, 0.028151552200317384, 0.028030975341796875, 0.027975679397583008, 0.028110847473144532, 0.02778726387023926, 0.02762112045288086, 0.02750694465637207, 0.027594335556030275, 0.027789215087890624, 0.027894271850585937, 0.0277708797454834, 0.027836063385009765, 0.02775059127807617, 0.028353824615478515, 0.028119871139526367, 0.028079999923706054, 0.028123327255249023, 0.02812518310546875, 0.028954784393310548, 0.028184288024902342, 0.02806387138366699, 0.02814771270751953, 0.02777635192871094, 0.028213375091552733, 0.02784908866882324, 0.027944864273071288, 0.02789811134338379, 0.027973024368286133, 0.027779647827148438, 0.027711135864257812, 0.02789414405822754, 0.027864992141723634, 0.027743711471557617, 0.02779404830932617, 0.028248064041137694, 0.027873279571533204, 0.02773606491088867, 0.028008575439453124, 0.02814761543273926, 0.028110815048217774, 0.028051456451416015, 0.027992223739624022, 0.027863967895507814, 0.027939775466918945, 0.02779292869567871, 0.02776316833496094, 0.027680192947387695, 0.027564607620239257, 0.027855968475341795, 0.027627487182617188, 0.027800512313842774, 0.027803232192993164, 0.027931039810180663, 0.027853919982910157, 0.028773279190063478, 0.02780499267578125, 0.028055295944213868, 0.027765695571899413, 0.02790399932861328, 0.028033023834228517, 0.027856767654418944, 0.027827327728271484, 0.02893619155883789, 0.028039167404174805, 0.028060863494873047, 0.027712320327758787, 0.027669504165649415, 0.027613920211791994, 0.027615520477294923, 0.02752012825012207, 0.027776927947998048, 0.02780620765686035, 0.028094655990600587, 0.027680288314819335, 0.02775321578979492, 0.027572223663330078, 0.02772991943359375, 0.0276378231048584, 0.027629247665405275, 0.028608640670776366, 0.0277544002532959, 0.02771171188354492, 0.027745311737060546, 0.027654176712036134, 0.027795839309692382, 0.027605567932128906, 0.027891712188720705, 0.027557472229003906, 0.027641855239868163, 0.027713951110839845, 0.027770975112915038, 0.02774131202697754, 0.0277225284576416, 0.027547040939331056, 0.027664480209350587, 0.02821990394592285, 0.027889663696289063, 0.02773324775695801, 0.027615936279296874, 0.02763987159729004, 0.029056608200073244, 0.0278573112487793, 0.027719808578491212, 0.02780556869506836, 0.0284117431640625, 0.02786729621887207, 0.02778726387023926, 0.02810425567626953, 0.027940704345703126, 0.027720544815063478, 0.027725215911865234, 0.02790575981140137, 0.02770172882080078, 0.027637216567993166, 0.027591360092163085, 0.027842559814453126, 0.027639808654785155, 0.027661983489990234, 0.027859296798706055, 0.027723392486572265, 0.02768499183654785, 0.027684415817260742, 0.027714111328125, 0.027533439636230467, 0.027594207763671875]",tokens/s,35.59839511382615,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1141, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 944, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 677, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 500, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4169.474048,5925.43744,0.0,5530.189824,5138.859008,s,1,11.6259453125,11.6259453125,0.0,11.6259453125,11.6259453125,11.6259453125,11.6259453125,[11.6259453125],,kWh,0.000136070056616677,1.5002031505387146e-05,5.769143504200103e-05,0.00020876352316406518,,MB,1376.415744,5944.311808,0.0,5534.384128,4844.878336,s,10,1.9729530181884767,0.19729530181884766,0.00043113796195293495,0.19714612579345703,0.1979267852783203,0.19802846374511718,0.19810980651855467,"[0.19710159301757812, 0.1971556091308594, 0.19666685485839844, 0.19703488159179688, 0.1971697235107422, 0.1971366424560547, 0.19698876953125, 0.19766461181640624, 0.19813014221191405, 0.19790419006347656]",tokens/s,1297.547370058785,kWh,5.785537226960829e-06,6.376602793393575e-07,3.82815121066692e-06,1.0251348716967105e-05,tokens/kWh,24972323.84420715,MB,1382.596608,5944.311808,0.0,5534.384128,5015.800832,s,10,15.389005493164063,1.5389005493164063,0.014977846303433613,1.5344813842773437,1.5553174926757813,1.5602898498535156,1.564267735595703,"[1.5148800048828126, 1.5278021240234374, 1.5240762939453125, 1.5535860595703126, 1.5323665771484376, 1.5480943603515624, 1.5542125244140625, 1.532129150390625, 1.56526220703125, 1.53659619140625]",tokens/s,40.93831796212249,kWh,4.5213650301372886e-05,4.987208540213296e-06,3.003357032313415e-05,8.023442916472033e-05,tokens/kWh,785199.0804429075,,s,630,15.385902879714974,0.024422068063039627,0.0005485261684029732,0.02434662437438965,0.024897196960449217,0.025076761531829833,0.026705567264556904,"[0.025111167907714844, 0.024440832138061523, 0.024295167922973634, 0.024030975341796875, 0.02390297508239746, 0.02444585609436035, 0.023923456192016603, 0.023861087799072266, 0.02385446357727051, 0.023780223846435546, 0.023910400390625, 0.023661983489990233, 0.02385161590576172, 0.023874879837036133, 0.023857856750488283, 0.023767040252685546, 0.023934656143188477, 0.023779647827148438, 0.023981983184814454, 0.023758304595947265, 0.023777919769287108, 0.024086528778076172, 0.023758111953735353, 0.023870080947875977, 0.023722272872924804, 0.02379961585998535, 0.024037376403808593, 0.023783424377441405, 0.023846912384033202, 0.02392064094543457, 0.023909439086914064, 0.02384172821044922, 0.023774591445922852, 0.023753343582153322, 0.02388172721862793, 0.023807584762573244, 0.02391823959350586, 0.023821056365966795, 0.02411520004272461, 0.024133216857910155, 0.023875648498535157, 0.023951711654663085, 0.023814144134521483, 0.02370172882080078, 0.023791391372680663, 0.023942304611206056, 0.023835487365722656, 0.023785472869873047, 0.02390425682067871, 0.027461631774902344, 0.024371200561523438, 0.02449807929992676, 0.02407334327697754, 0.0239400634765625, 0.02408448028564453, 0.02421116828918457, 0.024219263076782228, 0.024171167373657227, 0.024379392623901368, 0.024303327560424803, 0.024272224426269532, 0.02417132759094238, 0.02423616027832031, 0.025683359146118166, 0.024443487167358398, 0.024008544921875, 0.024058015823364257, 0.023973888397216796, 0.02384614372253418, 0.02376572799682617, 0.023784992218017578, 0.02598963165283203, 0.02432204818725586, 0.02408857536315918, 0.024202720642089844, 0.024287776947021486, 0.024003807067871093, 0.02390505599975586, 0.02390220832824707, 0.02393907165527344, 0.023813343048095702, 0.02380793571472168, 0.02385408020019531, 0.02385603141784668, 0.023897024154663087, 0.024554975509643556, 0.024046112060546875, 0.023834623336791993, 0.023848064422607423, 0.023889087677001954, 0.023824064254760743, 0.023711584091186524, 0.023675039291381837, 0.023734272003173826, 0.023758848190307616, 0.023764991760253908, 0.023789567947387694, 0.024696832656860353, 0.023789152145385743, 0.024035039901733397, 0.024072895050048827, 0.024796640396118164, 0.02447551918029785, 0.02452137565612793, 0.024633344650268556, 0.02632294464111328, 0.024600576400756836, 0.024415359497070313, 0.02455401611328125, 0.02428553581237793, 0.025557119369506835, 0.024440095901489257, 0.024383424758911133, 0.024187551498413087, 0.02412928009033203, 0.025329280853271484, 0.024369792938232424, 0.02423334312438965, 0.02397987174987793, 0.02393168067932129, 0.024206687927246093, 0.023955615997314453, 0.0239498233795166, 0.024008703231811524, 0.02507366371154785, 0.0247193603515625, 0.0248558406829834, 0.02416499137878418, 0.02388991928100586, 0.023840768814086914, 0.02386944007873535, 0.023920480728149413, 0.023832223892211915, 0.023924383163452148, 0.024167264938354492, 0.024306880950927735, 0.024126272201538086, 0.02396918487548828, 0.024164735794067384, 0.023867231369018554, 0.023853439331054688, 0.023792896270751953, 0.02388636779785156, 0.023637983322143556, 0.023778783798217774, 0.023712032318115233, 0.023959232330322267, 0.02391142463684082, 0.023955263137817383, 0.02385206413269043, 0.024101856231689454, 0.024067712783813477, 0.024436927795410155, 0.024412351608276366, 0.024426496505737305, 0.024331392288208006, 0.024423295974731446, 0.024352256774902343, 0.02426838493347168, 0.024408992767333985, 0.02429267120361328, 0.024105663299560546, 0.023964704513549803, 0.023952352523803712, 0.023963008880615234, 0.02397657585144043, 0.02411929512023926, 0.023969503402709962, 0.023945184707641603, 0.023869760513305666, 0.023867071151733397, 0.02401055908203125, 0.024053855895996092, 0.02501059150695801, 0.024049663543701173, 0.023883775711059572, 0.023967744827270508, 0.023734272003173826, 0.024625152587890626, 0.025053184509277345, 0.024426496505737305, 0.024180736541748047, 0.024258560180664062, 0.02422096061706543, 0.02445158386230469, 0.02430384063720703, 0.024247615814208985, 0.024326847076416015, 0.02845929527282715, 0.02529097557067871, 0.024987775802612303, 0.025051488876342773, 0.02453913688659668, 0.024392799377441408, 0.024314783096313478, 0.0243507194519043, 0.024401920318603516, 0.024565759658813476, 0.024416128158569337, 0.024593759536743164, 0.024584127426147462, 0.024513120651245116, 0.024671775817871094, 0.02459926414489746, 0.02452236747741699, 0.02453913688659668, 0.024458816528320312, 0.024646528244018556, 0.02515545654296875, 0.024550975799560545, 0.024611167907714844, 0.024716800689697265, 0.024658592224121093, 0.02454732894897461, 0.02450758361816406, 0.024740671157836912, 0.024633344650268556, 0.024454944610595702, 0.024330495834350586, 0.02450649642944336, 0.02459017562866211, 0.024885120391845702, 0.024487327575683594, 0.024705120086669922, 0.02473958396911621, 0.024854528427124024, 0.024853376388549803, 0.024845600128173828, 0.024826591491699218, 0.024737535476684572, 0.024729759216308593, 0.02480342483520508, 0.024962400436401366, 0.02496518325805664, 0.024984159469604493, 0.024809471130371095, 0.024965343475341798, 0.024864543914794923, 0.024898752212524414, 0.024875680923461915, 0.024770719528198242, 0.02469647979736328, 0.024807775497436523, 0.024647680282592774, 0.024558975219726564, 0.02449884796142578, 0.02442032051086426, 0.024448959350585938, 0.024260671615600585, 0.02421766471862793, 0.02427039909362793, 0.02444121551513672, 0.025168575286865235, 0.02469478416442871, 0.024625152587890626, 0.02455891227722168, 0.02750035285949707, 0.02478483200073242, 0.024761568069458006, 0.024567520141601563, 0.024350656509399413, 0.024254528045654297, 0.024145824432373047, 0.02460476875305176, 0.024682048797607423, 0.024950624465942383, 0.024232608795166016, 0.024393152236938477, 0.024172096252441405, 0.024120256423950194, 0.02432204818725586, 0.02408038330078125, 0.024174591064453126, 0.024162303924560546, 0.024070144653320313, 0.024043712615966797, 0.024001792907714845, 0.024107583999633787, 0.024070144653320313, 0.023961599349975587, 0.024149343490600585, 0.023972223281860352, 0.02402332878112793, 0.02402675247192383, 0.02421798324584961, 0.024058975219726563, 0.024130464553833008, 0.023977983474731446, 0.024180736541748047, 0.023914239883422853, 0.024143743515014648, 0.02409219169616699, 0.024101728439331054, 0.02398134422302246, 0.024304351806640624, 0.024337791442871095, 0.02414963150024414, 0.024022016525268555, 0.02407219123840332, 0.024102912902832032, 0.024106464385986327, 0.024102624893188478, 0.02406483268737793, 0.024214656829833984, 0.02427369689941406, 0.024341728210449217, 0.024275840759277342, 0.024186208724975587, 0.024244895935058595, 0.02429542350769043, 0.02434867286682129, 0.024376640319824217, 0.024564416885375976, 0.024481792449951172, 0.024623104095458984, 0.025324575424194334, 0.024835039138793945, 0.0246824951171875, 0.02472550392150879, 0.024639488220214844, 0.02454528045654297, 0.024565759658813476, 0.024690624237060546, 0.02447337532043457, 0.024314144134521484, 0.024393087387084962, 0.024412799835205078, 0.024349695205688478, 0.025010208129882812, 0.02425267219543457, 0.024326879501342772, 0.024483903884887696, 0.02447990417480469, 0.0245184326171875, 0.024625152587890626, 0.024571903228759767, 0.024840351104736327, 0.024897024154663085, 0.02489081573486328, 0.02508188819885254, 0.024562559127807616, 0.02462886428833008, 0.024550912857055664, 0.024556064605712892, 0.02463279914855957, 0.024961503982543945, 0.024508031845092773, 0.02459519958496094, 0.024500255584716798, 0.02446313667297363, 0.024355039596557618, 0.024290847778320312, 0.02446588706970215, 0.024459264755249024, 0.0243158073425293, 0.024305759429931642, 0.024327232360839845, 0.02442540740966797, 0.024278976440429687, 0.024270751953125, 0.024182111740112304, 0.02423263931274414, 0.02410918426513672, 0.024213504791259766, 0.024233375549316406, 0.024928287506103517, 0.024709695816040038, 0.02461676788330078, 0.024649919509887694, 0.024639488220214844, 0.024588287353515623, 0.024666112899780275, 0.024625152587890626, 0.024657920837402345, 0.02480512046813965, 0.024631616592407226, 0.025079296112060546, 0.02477004814147949, 0.025399616241455078, 0.024852479934692383, 0.02483404731750488, 0.024688640594482423, 0.024477344512939453, 0.024643936157226563, 0.02439740753173828, 0.02428316879272461, 0.024324480056762694, 0.024422399520874022, 0.024292640686035157, 0.024453184127807618, 0.02470992088317871, 0.024512384414672853, 0.025017824172973633, 0.024373855590820313, 0.024330175399780274, 0.024358816146850586, 0.02414396858215332, 0.024238079071044923, 0.024377023696899414, 0.024461088180541993, 0.024558111190795897, 0.02484454345703125, 0.024383232116699217, 0.0242872314453125, 0.024337408065795898, 0.024382463455200197, 0.02470297622680664, 0.024564735412597655, 0.024914464950561523, 0.024822240829467775, 0.024925504684448242, 0.025652223587036133, 0.025369983673095703, 0.024824127197265625, 0.024951007843017577, 0.025038623809814455, 0.025018367767333984, 0.025011327743530272, 0.02473651123046875, 0.02473353576660156, 0.02480156707763672, 0.024759679794311523, 0.024937088012695313, 0.025005952835083008, 0.024803071975708007, 0.024656255722045897, 0.02468467140197754, 0.024481664657592772, 0.024344575881958007, 0.02434272003173828, 0.024280000686645507, 0.02471331214904785, 0.024382240295410158, 0.024426368713378905, 0.024469215393066405, 0.0244105281829834, 0.02458624076843262, 0.02464102363586426, 0.02439014434814453, 0.024770559310913084, 0.02636390495300293, 0.024819711685180663, 0.024491392135620117, 0.02450432014465332, 0.024332767486572267, 0.024262815475463866, 0.02433827209472656, 0.024184255599975585, 0.02431001663208008, 0.024215103149414063, 0.02425129508972168, 0.02431337547302246, 0.024191455841064455, 0.02433433532714844, 0.024238079071044923, 0.024248064041137694, 0.02443084716796875, 0.024156160354614258, 0.024203264236450195, 0.02415590476989746, 0.024067359924316405, 0.02402403259277344, 0.02429078483581543, 0.024013343811035155, 0.024470687866210938, 0.024066240310668945, 0.025408063888549805, 0.024436832427978516, 0.024168447494506837, 0.023993440628051758, 0.02398611259460449, 0.024182783126831055, 0.023972831726074218, 0.024043392181396485, 0.02396758460998535, 0.02512268829345703, 0.025455007553100584, 0.024328191757202147, 0.024259904861450195, 0.024203968048095704, 0.024303615570068358, 0.024190208435058595, 0.0243143367767334, 0.024903968811035158, 0.02469478416442871, 0.024631296157836914, 0.024766719818115235, 0.024366111755371095, 0.024195936203002928, 0.024098432540893555, 0.02420966339111328, 0.024045312881469726, 0.024238336563110353, 0.024131584167480468, 0.024354591369628906, 0.024180959701538086, 0.024219648361206055, 0.02427289581298828, 0.024164352416992187, 0.024360960006713867, 0.024193023681640623, 0.0242523193359375, 0.02409401512145996, 0.024208160400390626, 0.02534534454345703, 0.02462175941467285, 0.024604671478271483, 0.0245449275970459, 0.02477916717529297, 0.025284095764160155, 0.024773056030273438, 0.024832000732421877, 0.02470911979675293, 0.024791040420532227, 0.02478489685058594, 0.02462067222595215, 0.024926591873168945, 0.024772607803344726, 0.024567808151245117, 0.02457088088989258, 0.02451968002319336, 0.024386560440063477, 0.02433945655822754, 0.024368928909301757, 0.024407615661621095, 0.024705631256103516, 0.02444211196899414, 0.024560447692871093, 0.024557567596435546, 0.024528480529785155, 0.024377407073974608, 0.02868396759033203, 0.029337919235229493, 0.025039199829101563, 0.024823808670043947, 0.02452992057800293, 0.024509439468383788, 0.02492620849609375, 0.024713216781616212, 0.025691743850708007, 0.024959360122680664, 0.024917024612426758, 0.024609792709350587, 0.02450841522216797, 0.02485862350463867, 0.024823040008544923, 0.0245296630859375, 0.024657695770263673, 0.025628896713256837, 0.02684511947631836, 0.02492742347717285, 0.024699743270874024, 0.024518688201904296, 0.024393888473510743, 0.024291168212890624, 0.0245166072845459, 0.024410015106201173, 0.024933567047119142, 0.024715999603271484, 0.025858240127563478, 0.02427494430541992, 0.024227840423583984, 0.024151935577392578, 0.02411065673828125, 0.024169023513793946, 0.02414364814758301, 0.02430793571472168, 0.025218399047851562, 0.024681055068969726, 0.02480975914001465, 0.024666112899780275, 0.02465977668762207, 0.02478086471557617, 0.02450444793701172, 0.02454047966003418, 0.02435158348083496, 0.02421743965148926, 0.024457216262817383, 0.024395776748657227, 0.024459264755249024, 0.024371200561523438, 0.024694976806640626, 0.024371007919311523, 0.024538496017456054, 0.02440665626525879, 0.024470880508422853, 0.02428995132446289, 0.02431385612487793, 0.024338655471801758, 0.024233919143676758, 0.024209184646606444, 0.02439331245422363, 0.02457638359069824, 0.024430496215820312, 0.02436934471130371, 0.024258272171020508, 0.024273183822631834, 0.024163328170776367, 0.024208383560180666, 0.024178688049316405, 0.02426825523376465, 0.02439017677307129, 0.024287391662597656, 0.02418671989440918, 0.024218719482421876, 0.02410179138183594, 0.02409881591796875, 0.024170495986938476, 0.02413907241821289, 0.024053695678710938, 0.024342432022094726, 0.024265567779541017, 0.024102432250976562, 0.02437513542175293, 0.02416703987121582, 0.024030879974365236, 0.02417033576965332, 0.024147903442382813, 0.024119871139526367, 0.02408399963378906, 0.024234464645385742, 0.02413273620605469, 0.024427391052246093, 0.024104383468627928, 0.024437311172485352, 0.027150335311889647, 0.024636800765991212, 0.024267295837402343, 0.024229984283447265, 0.024143871307373048]",tokens/s,40.946573296689834,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1339.94496,1148.125184,0.0,752.877568,710.554112,s,1,8.7731044921875,8.7731044921875,0.0,8.7731044921875,8.7731044921875,8.7731044921875,8.7731044921875,[8.7731044921875],,kWh,4.9749024966695284e-05,5.477843875071808e-06,1.6886957954015402e-05,7.211382679578249e-05,,MB,1337.827328,1429.143552,0.0,1019.215872,949.09696,s,10,0.26381101036071775,0.02638110103607178,0.0003819674675211477,0.026364831924438475,0.02667032299041748,0.026933353710174558,0.027143778285980222,"[0.026392799377441406, 0.026202783584594727, 0.02627484893798828, 0.02629033660888672, 0.02647478485107422, 0.02719638442993164, 0.02633686447143555, 0.02661187171936035, 0.026468511581420898, 0.025561824798583984]",tokens/s,9703.916438133589,kWh,7.683246981357689e-07,8.469786453777631e-08,5.001531779000312e-07,1.3531757405735765e-06,tokens/kWh,189184591.71569848,MB,1358.237696,1437.53216,0.0,1027.60448,949.09952,s,10,13.170380249023436,1.317038024902344,0.008806143524571609,1.3187987060546875,1.325497912597656,1.3257829528808593,1.3260109851074218,"[1.3121505126953126, 1.3254345703125, 1.3126083984375, 1.3182025146484375, 1.3260679931640624, 1.322536376953125, 1.3193948974609375, 1.323876953125, 1.315271240234375, 1.2948367919921875]",tokens/s,47.83460979015495,kWh,3.8256545591859534e-05,4.219283233828629e-06,1.577088761669926e-05,5.824671644238742e-05,tokens/kWh,1081606.0346047853,,s,630,13.167925653457653,0.020901469291202606,0.0003524974265339799,0.020911567687988283,0.021156941413879395,0.021330596446990965,0.022246307201385505,"[0.020983808517456053, 0.02114518356323242, 0.021186784744262697, 0.020934047698974608, 0.02098361587524414, 0.021183456420898437, 0.02122889518737793, 0.02095961570739746, 0.0210351676940918, 0.021003936767578123, 0.021058015823364258, 0.021014528274536134, 0.021112831115722656, 0.021120704650878907, 0.02130361557006836, 0.021121023178100586, 0.021405696868896484, 0.02085478401184082, 0.020712959289550782, 0.02057676887512207, 0.02070688056945801, 0.02066067123413086, 0.020688383102416993, 0.020515327453613282, 0.02048371124267578, 0.02095961570739746, 0.02078460884094238, 0.021479232788085938, 0.020513151168823244, 0.02039401626586914, 0.02045574378967285, 0.020321311950683593, 0.02055471992492676, 0.02044435119628906, 0.020327392578125, 0.02049007987976074, 0.02049843215942383, 0.020526559829711914, 0.02040425682067871, 0.020342784881591795, 0.020323999404907228, 0.020302688598632813, 0.020479999542236327, 0.020528799057006837, 0.02067433547973633, 0.02061574363708496, 0.02102681541442871, 0.02085478401184082, 0.02078220748901367, 0.02089459228515625, 0.02079539108276367, 0.020746240615844725, 0.020612224578857422, 0.02069183921813965, 0.020983808517456053, 0.020938304901123046, 0.020931039810180664, 0.020813791275024415, 0.020987680435180664, 0.02077939224243164, 0.022951776504516602, 0.020856800079345702, 0.020869152069091797, 0.020920320510864256, 0.02097113609313965, 0.020863359451293945, 0.020993247985839843, 0.02083510398864746, 0.020774911880493165, 0.02083430480957031, 0.020918272018432618, 0.02097171211242676, 0.020843711853027344, 0.021009023666381837, 0.02106572723388672, 0.020965087890625, 0.020864320755004884, 0.020851200103759765, 0.020797920227050782, 0.02092755126953125, 0.020761568069458006, 0.020750303268432618, 0.02099404716491699, 0.02092748832702637, 0.02082099151611328, 0.02085273551940918, 0.02074959945678711, 0.020941535949707032, 0.02143020820617676, 0.02170172882080078, 0.021799903869628906, 0.021288415908813477, 0.0210478401184082, 0.020815872192382814, 0.020882720947265624, 0.02095996856689453, 0.022290016174316408, 0.02213929557800293, 0.020899839401245117, 0.021001247406005858, 0.021110944747924805, 0.020933055877685548, 0.02075663948059082, 0.0210229434967041, 0.02076643180847168, 0.020859167098999022, 0.02070528030395508, 0.020834079742431642, 0.020816064834594725, 0.02309328079223633, 0.02149171257019043, 0.02098761558532715, 0.020857120513916017, 0.020809728622436522, 0.020872575759887695, 0.021139551162719726, 0.020830751419067383, 0.02071334457397461, 0.020781280517578125, 0.02120841598510742, 0.021853824615478516, 0.021191743850708007, 0.02083724784851074, 0.0207258243560791, 0.020802495956420898, 0.020707263946533203, 0.020719615936279297, 0.020711423873901368, 0.02072777557373047, 0.02055990409851074, 0.020584447860717774, 0.02087116813659668, 0.0208855037689209, 0.020811456680297852, 0.020750816345214845, 0.020701023101806642, 0.02087068748474121, 0.020759008407592774, 0.02071900749206543, 0.020683359146118165, 0.020527103424072265, 0.020692991256713866, 0.02099404716491699, 0.02065203285217285, 0.02048409652709961, 0.020625408172607423, 0.020707328796386718, 0.020985855102539062, 0.02086502456665039, 0.020748287200927733, 0.02065203285217285, 0.020768224716186525, 0.02059516716003418, 0.020594240188598633, 0.020602752685546875, 0.02069772720336914, 0.020999488830566407, 0.020807903289794923, 0.020718048095703125, 0.020800928115844726, 0.020946815490722658, 0.020775295257568358, 0.020807359695434572, 0.020774560928344725, 0.020810752868652343, 0.02090166473388672, 0.020907936096191407, 0.02096928024291992, 0.02111516761779785, 0.020985279083251953, 0.020929311752319334, 0.020960607528686524, 0.020997983932495117, 0.020834592819213866, 0.020965919494628907, 0.021387264251708983, 0.021127168655395507, 0.020930559158325195, 0.020993919372558595, 0.020916351318359373, 0.020948192596435548, 0.021006591796875, 0.021100511550903322, 0.02100918388366699, 0.02083203125, 0.020920160293579102, 0.020740095138549804, 0.020879520416259765, 0.02101795196533203, 0.02084966468811035, 0.020927776336669923, 0.020625791549682616, 0.020765024185180662, 0.020702816009521483, 0.02068931198120117, 0.02061311912536621, 0.020706367492675782, 0.02087376022338867, 0.02196291160583496, 0.020739456176757813, 0.020933504104614257, 0.02087731170654297, 0.020891136169433593, 0.020984319686889647, 0.02090729522705078, 0.02093129539489746, 0.02091007995605469, 0.020760000228881834, 0.02101715278625488, 0.02100160026550293, 0.020896255493164064, 0.021176448822021486, 0.022616064071655274, 0.021098367691040038, 0.02111296081542969, 0.02081996726989746, 0.020778112411499024, 0.02086591911315918, 0.020872928619384765, 0.02097107124328613, 0.021269216537475585, 0.021118783950805665, 0.02096771240234375, 0.020837343215942383, 0.02079372787475586, 0.020953664779663084, 0.020884992599487305, 0.02077337646484375, 0.020840448379516603, 0.020676607131958007, 0.020772863388061523, 0.020918272018432618, 0.02089779281616211, 0.020741952896118163, 0.02084230422973633, 0.020756864547729494, 0.020930559158325195, 0.020758527755737305, 0.020875263214111327, 0.020793344497680662, 0.021741567611694337, 0.02109008026123047, 0.021114784240722655, 0.020726079940795897, 0.020641408920288085, 0.020783487319946288, 0.020825279235839843, 0.02080441665649414, 0.02066441535949707, 0.020864927291870117, 0.020674560546875, 0.02074006462097168, 0.021083168029785156, 0.021128416061401367, 0.021044511795043946, 0.021080095291137694, 0.02108051109313965, 0.02103843116760254, 0.02103107261657715, 0.021107200622558595, 0.021018047332763672, 0.02100896072387695, 0.020965375900268556, 0.02101055908203125, 0.02097760009765625, 0.02082809638977051, 0.021090303421020508, 0.021086208343505858, 0.020940256118774415, 0.02109040069580078, 0.020935104370117186, 0.020976640701293944, 0.02093881607055664, 0.021033920288085938, 0.021340160369873046, 0.021028863906860353, 0.021325824737548828, 0.02095913505554199, 0.02127052879333496, 0.021026784896850587, 0.02124608039855957, 0.02113052749633789, 0.021054176330566405, 0.021178367614746094, 0.021104448318481444, 0.021176511764526368, 0.021045248031616212, 0.021198848724365234, 0.02121126365661621, 0.021402816772460937, 0.021037759780883788, 0.020961280822753905, 0.020996095657348633, 0.02102272033691406, 0.020856767654418944, 0.0209552001953125, 0.020919807434082033, 0.02091430473327637, 0.0209716796875, 0.021069183349609374, 0.020886367797851562, 0.021070016860961913, 0.02101638412475586, 0.02103500747680664, 0.02092380714416504, 0.02097158432006836, 0.021055967330932616, 0.021421728134155275, 0.02113577651977539, 0.02099404716491699, 0.020911136627197267, 0.020982976913452148, 0.020791072845458985, 0.02084659194946289, 0.02086092758178711, 0.02105180740356445, 0.021178207397460937, 0.021104639053344726, 0.021045248031616212, 0.021044288635253906, 0.021183391571044922, 0.02101046371459961, 0.020894975662231446, 0.021076255798339844, 0.021058015823364258, 0.021159936904907226, 0.021067743301391603, 0.021063711166381834, 0.020964351654052735, 0.021097471237182617, 0.0209420166015625, 0.02094163131713867, 0.020891616821289063, 0.020908063888549804, 0.021327871322631836, 0.021014528274536134, 0.020930080413818358, 0.020924448013305664, 0.02098419189453125, 0.02102672004699707, 0.02081398391723633, 0.02086092758178711, 0.02086092758178711, 0.021143552780151367, 0.02097724723815918, 0.020956607818603517, 0.020849632263183593, 0.021104639053344726, 0.021331647872924804, 0.021174591064453126, 0.02106572723388672, 0.021325855255126952, 0.021073888778686524, 0.020981760025024415, 0.021083711624145508, 0.02099407958984375, 0.022002080917358398, 0.02333695983886719, 0.021235488891601564, 0.020903615951538085, 0.02076214408874512, 0.02085785675048828, 0.020785152435302736, 0.020746240615844725, 0.02073353576660156, 0.020699167251586915, 0.020578624725341797, 0.02057632064819336, 0.020621440887451173, 0.02048192024230957, 0.020676607131958007, 0.020719615936279297, 0.020665664672851563, 0.021561151504516603, 0.020537567138671876, 0.02042947196960449, 0.020335647583007814, 0.020564960479736327, 0.020920320510864256, 0.020961408615112306, 0.021110784530639647, 0.020938623428344728, 0.020965343475341798, 0.02077654457092285, 0.020811840057373045, 0.020791807174682618, 0.020944896697998046, 0.02079871940612793, 0.02077568054199219, 0.02081996726989746, 0.02079539108276367, 0.020911231994628906, 0.02084668731689453, 0.02115660858154297, 0.020905088424682618, 0.020927391052246093, 0.020803455352783204, 0.020915584564208986, 0.020875616073608397, 0.020957599639892577, 0.021373088836669923, 0.020966527938842773, 0.020628032684326173, 0.020542816162109376, 0.0208023681640625, 0.020883455276489257, 0.020936704635620116, 0.020867071151733398, 0.020856351852416993, 0.02081622314453125, 0.020759967803955077, 0.020871904373168944, 0.0208154239654541, 0.02071340751647949, 0.02076416015625, 0.020789344787597655, 0.020917312622070312, 0.0207642879486084, 0.020652095794677736, 0.020812992095947266, 0.020882400512695312, 0.020918272018432618, 0.02081996726989746, 0.02088960075378418, 0.020955135345458984, 0.021192384719848634, 0.020871488571166993, 0.02099945640563965, 0.02117491149902344, 0.021014623641967774, 0.020971519470214844, 0.020979103088378907, 0.020830432891845704, 0.021531007766723634, 0.022556671142578123, 0.021022048950195313, 0.02111350440979004, 0.021007680892944337, 0.021153888702392577, 0.02103766441345215, 0.02097350311279297, 0.02103500747680664, 0.020987295150756837, 0.021027423858642577, 0.020905824661254884, 0.020930976867675782, 0.020975360870361327, 0.02087321662902832, 0.020866336822509764, 0.020957696914672853, 0.020951263427734373, 0.020783103942871094, 0.02076198387145996, 0.020793983459472656, 0.020911903381347657, 0.020887775421142576, 0.02091007995605469, 0.020950368881225586, 0.021004928588867187, 0.02102070426940918, 0.020973567962646485, 0.021006336212158205, 0.020891679763793945, 0.02090185546875, 0.020933759689331054, 0.020797439575195312, 0.02088025665283203, 0.020883455276489257, 0.020960575103759767, 0.021018720626831053, 0.02107209587097168, 0.021031295776367188, 0.020972639083862304, 0.0210251522064209, 0.021017120361328124, 0.021041183471679686, 0.02089967918395996, 0.02095871925354004, 0.021019264221191405, 0.02086195182800293, 0.020871456146240235, 0.0209968318939209, 0.02098099136352539, 0.020992767333984374, 0.02112291145324707, 0.02095939254760742, 0.020912128448486327, 0.020908031463623047, 0.02105881690979004, 0.02105625534057617, 0.021255327224731446, 0.02100227165222168, 0.020984640121459962, 0.022052608489990234, 0.021329311370849608, 0.02184012794494629, 0.02163539123535156, 0.021, 0.0210817928314209, 0.021015327453613283, 0.02113929557800293, 0.0209849910736084, 0.02084524726867676, 0.02090220832824707, 0.021114879608154297, 0.021102592468261717, 0.021159999847412108, 0.021017663955688475, 0.021363584518432618, 0.021111967086791993, 0.021070688247680665, 0.020999711990356447, 0.021272960662841796, 0.021136608123779297, 0.020941247940063478, 0.021034591674804686, 0.02111369514465332, 0.020926464080810548, 0.020961280822753905, 0.02107961654663086, 0.02107027244567871, 0.021307392120361326, 0.021057504653930664, 0.02102684783935547, 0.020940031051635742, 0.020861696243286133, 0.020733951568603515, 0.02063929557800293, 0.021856704711914064, 0.02120275115966797, 0.02068511962890625, 0.020909952163696288, 0.021112831115722656, 0.0209017276763916, 0.020934816360473632, 0.021102399826049806, 0.02099628829956055, 0.02104729652404785, 0.020880544662475586, 0.021015392303466798, 0.02093417549133301, 0.02095155143737793, 0.020891616821289063, 0.020951040267944337, 0.020953088760375976, 0.020817760467529298, 0.020910240173339843, 0.02089916801452637, 0.02093942451477051, 0.020874719619750976, 0.020734495162963867, 0.02067203140258789, 0.020564191818237303, 0.02057241630554199, 0.020539264678955078, 0.02063488006591797, 0.020979648590087892, 0.02110089683532715, 0.02053353691101074, 0.020582304000854493, 0.020351200103759765, 0.02025494384765625, 0.020174816131591798, 0.020162559509277343, 0.020121599197387697, 0.020090303421020507, 0.02018502426147461, 0.02026134490966797, 0.02069286346435547, 0.020222335815429687, 0.020354175567626955, 0.020198272705078124, 0.020281375885009764, 0.020148191452026367, 0.02012265586853027, 0.020236608505249023, 0.022420127868652343, 0.02115283203125, 0.020245439529418947, 0.020325952529907227, 0.02034432029724121, 0.02017990493774414, 0.02037478446960449, 0.020312480926513672, 0.02023664093017578, 0.020143455505371093, 0.020148895263671876, 0.02026108741760254, 0.02023811149597168, 0.020176895141601564, 0.020271104812622072, 0.020173856735229492, 0.020150400161743163, 0.020204416275024416, 0.020336095809936523, 0.020204191207885743, 0.020241376876831054, 0.020298847198486326, 0.02010044860839844, 0.020226112365722658, 0.020296064376831055, 0.020183263778686525, 0.0201889591217041, 0.02030182456970215, 0.02030723190307617, 0.02017353630065918, 0.020578304290771485, 0.020823328018188477, 0.02078998374938965, 0.020774335861206056, 0.020916799545288085, 0.021016576766967773, 0.02084864044189453, 0.02101862335205078, 0.020934560775756835, 0.020959327697753907, 0.02087321662902832, 0.020996095657348633, 0.02096143913269043, 0.020848703384399415, 0.020905759811401366, 0.020903488159179688, 0.0208218879699707, 0.020886079788208008, 0.020932607650756836, 0.020748287200927733, 0.02115100860595703, 0.02062758445739746, 0.020537952423095703, 0.022007551193237305]",tokens/s,47.84352650370366,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTJForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Deci/DeciCoder-1b contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Deci/DeciCoder-1b. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2166.366208,2964.258816,0.0,2569.0112,2295.745536,s,1,9.6473115234375,9.6473115234375,0.0,9.6473115234375,9.6473115234375,9.6473115234375,9.6473115234375,[9.6473115234375],,kWh,7.833504790416252e-05,8.63349490721445e-06,3.1614747514036035e-05,0.000118583290325413,,MB,2210.922496,2981.036032,0.0,2571.108352,2282.381824,s,10,0.8473688659667968,0.08473688659667969,0.00044418413667075065,0.08474364852905274,0.08524638366699219,0.08527711105346679,0.08530169296264647,"[0.08514192199707031, 0.08523955535888672, 0.08446864318847656, 0.08440092468261719, 0.08451583862304687, 0.08378530883789062, 0.0853078384399414, 0.08481427001953125, 0.08502153778076171, 0.08467302703857423]",tokens/s,3021.1164261731446,kWh,2.5314591589078797e-06,2.791743459482126e-07,1.6692015269310386e-06,4.47983503178713e-06,tokens/kWh,57144961.40673164,MB,2214.264832,2981.036032,0.0,2571.108352,2391.673344,s,10,14.580236450195313,1.4580236450195314,0.009167518722685945,1.455982666015625,1.47125048828125,1.47271826171875,1.4738924804687499,"[1.466255859375, 1.4489053955078126, 1.4572235107421876, 1.4547418212890626, 1.47418603515625, 1.4539447021484375, 1.449278076171875, 1.4592376708984376, 1.47092431640625, 1.4455390625]",tokens/s,43.20917580123062,kWh,4.241340228567793e-05,4.677850070407303e-06,2.331506846046541e-05,7.040632081655064e-05,tokens/kWh,894806.0240805878,,s,630,14.57759053802491,0.023139032600039525,0.00042756544902658537,0.023002351760864258,0.023539679908752442,0.023715660858154297,0.024710056915283204,"[0.023114336013793944, 0.02290483283996582, 0.02556844711303711, 0.023786304473876953, 0.02325503921508789, 0.023262624740600587, 0.023386720657348634, 0.023386112213134767, 0.023293664932250976, 0.023371135711669922, 0.02322934341430664, 0.02330009651184082, 0.023211967468261718, 0.02341484832763672, 0.02328780746459961, 0.023185407638549805, 0.02294169616699219, 0.022982656478881838, 0.02278201675415039, 0.022904352188110353, 0.02293494415283203, 0.023133184432983397, 0.023000799179077148, 0.0229215030670166, 0.022944927215576172, 0.022940607070922853, 0.022919071197509765, 0.022996992111206056, 0.02286591911315918, 0.022931455612182617, 0.02308095932006836, 0.02309334373474121, 0.022935455322265624, 0.023345151901245118, 0.023439231872558593, 0.023256671905517577, 0.023480863571166993, 0.0231461124420166, 0.02427097511291504, 0.022927616119384767, 0.02285977554321289, 0.022859424591064454, 0.022917375564575196, 0.023052383422851562, 0.022861824035644532, 0.022882272720336914, 0.022825151443481444, 0.023070016860961915, 0.022923103332519533, 0.02299769592285156, 0.023136255264282226, 0.023613439559936524, 0.023781375885009767, 0.02371174430847168, 0.023596319198608398, 0.02359779167175293, 0.024195072174072265, 0.023811712265014648, 0.02398041534423828, 0.02362771224975586, 0.02358278465270996, 0.023563360214233397, 0.023378847122192382, 0.023551616668701172, 0.02343584060668945, 0.023367488861083984, 0.023177024841308593, 0.022986879348754884, 0.023045631408691408, 0.023251520156860352, 0.02319068717956543, 0.022917024612426756, 0.023071680068969726, 0.02279952049255371, 0.02299171257019043, 0.0230951042175293, 0.023062623977661133, 0.023203008651733397, 0.023395231246948242, 0.023323999404907227, 0.023212064743041994, 0.022873952865600587, 0.02292780876159668, 0.02279772758483887, 0.02281977653503418, 0.022726015090942384, 0.02292390441894531, 0.02287615966796875, 0.02286319923400879, 0.02273961639404297, 0.022802112579345703, 0.022814495086669922, 0.023266944885253906, 0.022852512359619142, 0.022921247482299803, 0.022796255111694336, 0.023119871139526366, 0.02274492835998535, 0.022853792190551756, 0.022789791107177736, 0.022964351654052733, 0.022740224838256835, 0.02275222396850586, 0.02272480010986328, 0.022841215133666992, 0.022874048233032226, 0.022794240951538085, 0.022749183654785156, 0.022816160202026366, 0.022820959091186522, 0.02280012893676758, 0.023089920043945313, 0.023856672286987304, 0.024348735809326172, 0.02326095962524414, 0.02292390441894531, 0.022908384323120118, 0.022835039138793947, 0.022926015853881834, 0.022841215133666992, 0.022922847747802736, 0.022841312408447265, 0.02295452880859375, 0.02276902389526367, 0.02285430335998535, 0.02285510444641113, 0.022922880172729494, 0.024281471252441407, 0.02438118362426758, 0.02303206443786621, 0.02290675163269043, 0.022838815689086914, 0.022979040145874024, 0.02292473602294922, 0.022819520950317383, 0.022861824035644532, 0.02287820816040039, 0.02308064079284668, 0.022878528594970703, 0.022855680465698244, 0.02288342475891113, 0.022721439361572265, 0.02277987289428711, 0.022784095764160156, 0.022783935546875, 0.022808576583862306, 0.022807584762573243, 0.022840288162231444, 0.022900384902954103, 0.023068735122680664, 0.023037824630737305, 0.023196063995361327, 0.023363584518432616, 0.023734272003173826, 0.023351295471191406, 0.02339571189880371, 0.02322412872314453, 0.023339839935302736, 0.02337123107910156, 0.023289472579956054, 0.023661088943481446, 0.026549888610839845, 0.023444223403930663, 0.023347103118896484, 0.023065887451171874, 0.023212352752685548, 0.0232043514251709, 0.023000640869140623, 0.022921855926513673, 0.022909759521484375, 0.022928384780883788, 0.02298396873474121, 0.022725343704223633, 0.022935552597045897, 0.02298044776916504, 0.022868127822875978, 0.02285772705078125, 0.023157983779907226, 0.02327628707885742, 0.023375648498535156, 0.02304025650024414, 0.022939647674560547, 0.022960128784179686, 0.02303340721130371, 0.022794687271118164, 0.022830432891845703, 0.022807199478149413, 0.022914144515991212, 0.02291315269470215, 0.022943744659423827, 0.022921215057373046, 0.022829055786132812, 0.02285539245605469, 0.023139904022216797, 0.022893280029296876, 0.02275107192993164, 0.02280790328979492, 0.02384160041809082, 0.022985824584960936, 0.02282534408569336, 0.022895135879516602, 0.022982656478881838, 0.022898687362670898, 0.022767616271972657, 0.022959583282470702, 0.022954208374023437, 0.023011648178100585, 0.022738208770751955, 0.02285968017578125, 0.022846271514892578, 0.02294166374206543, 0.02277174377441406, 0.02287513542175293, 0.022823936462402345, 0.022940704345703125, 0.022805471420288086, 0.022861087799072265, 0.022901472091674806, 0.02291097640991211, 0.02292278480529785, 0.02375641632080078, 0.02299737548828125, 0.02300156784057617, 0.023219520568847657, 0.02325299263000488, 0.023299968719482422, 0.023330720901489257, 0.02365532875061035, 0.02352742385864258, 0.023459487915039063, 0.02350668716430664, 0.023562816619873046, 0.023750688552856444, 0.02334422492980957, 0.023429088592529297, 0.023538623809814453, 0.023451391220092772, 0.022969888687133788, 0.023030399322509765, 0.023037696838378908, 0.023189184188842773, 0.023011199951171873, 0.02311587142944336, 0.022903327941894532, 0.023023775100708008, 0.023048255920410158, 0.023018976211547852, 0.023220703125, 0.023147647857666015, 0.02316716766357422, 0.02308780860900879, 0.022945024490356444, 0.02395136070251465, 0.023590911865234376, 0.02326937675476074, 0.023394304275512694, 0.023011327743530274, 0.023522592544555663, 0.02359164810180664, 0.023631872177124022, 0.023556095123291015, 0.023572383880615236, 0.02349679946899414, 0.023428543090820313, 0.023392223358154298, 0.02349484825134277, 0.02395084762573242, 0.023610240936279298, 0.023371807098388674, 0.02348236846923828, 0.023713407516479493, 0.023433792114257813, 0.02340553665161133, 0.02311827278137207, 0.023306655883789062, 0.023549184799194336, 0.02307753562927246, 0.023412223815917968, 0.023333663940429687, 0.023568191528320313, 0.023199743270874023, 0.023180383682250977, 0.02327235221862793, 0.02309119987487793, 0.022966272354125978, 0.02308064079284668, 0.02293552017211914, 0.02294963264465332, 0.022984928131103515, 0.022996768951416016, 0.022949600219726564, 0.02332966423034668, 0.023406591415405274, 0.023396352767944335, 0.023236608505249022, 0.02409267234802246, 0.023592863082885742, 0.024053855895996092, 0.023586111068725588, 0.023669439315795897, 0.02347542381286621, 0.023654176712036134, 0.024738815307617186, 0.02373017692565918, 0.02353561592102051, 0.02342857551574707, 0.02310812759399414, 0.0230830078125, 0.023044095993041993, 0.023138303756713868, 0.02327552032470703, 0.023258943557739258, 0.023124223709106446, 0.023070655822753906, 0.023052288055419923, 0.023145248413085937, 0.02327756881713867, 0.022943744659423827, 0.02329395294189453, 0.023113727569580078, 0.02327743911743164, 0.02326950454711914, 0.02319580841064453, 0.02293948745727539, 0.022889951705932617, 0.022841888427734373, 0.022929407119750975, 0.02285772705078125, 0.02287615966796875, 0.022828704833984376, 0.023011264801025392, 0.022923679351806642, 0.0230067195892334, 0.022915584564208984, 0.02290640068054199, 0.02284796714782715, 0.022978656768798827, 0.02304604721069336, 0.022996959686279298, 0.022947423934936522, 0.022995391845703126, 0.023172319412231444, 0.02296063995361328, 0.02289459228515625, 0.022921056747436525, 0.02290108871459961, 0.02298646354675293, 0.022801919937133788, 0.023032863616943358, 0.02299785614013672, 0.022967296600341795, 0.022971744537353515, 0.023016096115112305, 0.022914047241210937, 0.02299295997619629, 0.023124671936035155, 0.022892127990722655, 0.022934175491333007, 0.023165151596069335, 0.02287593650817871, 0.023717504501342773, 0.02285318374633789, 0.023328704833984373, 0.022962400436401367, 0.0229769287109375, 0.02303923225402832, 0.02313113594055176, 0.022968320846557616, 0.023023008346557617, 0.022955968856811525, 0.023059104919433592, 0.023023008346557617, 0.022968511581420898, 0.022931072235107423, 0.023370527267456056, 0.023044095993041993, 0.022983903884887694, 0.026568639755249025, 0.022988319396972656, 0.02299958419799805, 0.023079999923706053, 0.023392288208007813, 0.022922079086303712, 0.02289254379272461, 0.022945728302001953, 0.023099327087402345, 0.022841472625732422, 0.02300636863708496, 0.022811296463012696, 0.023220415115356444, 0.022771711349487304, 0.022976512908935546, 0.022888128280639648, 0.02291539192199707, 0.022798336029052735, 0.02285919952392578, 0.022814752578735352, 0.02283478355407715, 0.022952896118164062, 0.022965375900268554, 0.022840192794799805, 0.023134208679199218, 0.023158784866333007, 0.02293289566040039, 0.022798816680908204, 0.022912864685058595, 0.023049728393554687, 0.0228503360748291, 0.022788095474243163, 0.022906879425048828, 0.02281881523132324, 0.02291097640991211, 0.026089471817016603, 0.023821439743041992, 0.023293888092041016, 0.023110591888427734, 0.022896543502807617, 0.022888063430786133, 0.02281875228881836, 0.022856224060058595, 0.022829055786132812, 0.022926687240600586, 0.022815391540527342, 0.02289039993286133, 0.022892032623291016, 0.022901344299316406, 0.022824575424194336, 0.022859392166137697, 0.022823295593261718, 0.022897024154663086, 0.022826751708984374, 0.023224319458007812, 0.022916704177856444, 0.023007904052734375, 0.022943359375, 0.022936992645263672, 0.02294063949584961, 0.02300668716430664, 0.022859392166137697, 0.022909919738769532, 0.022935487747192382, 0.02346598434448242, 0.023219263076782227, 0.022892768859863282, 0.023433952331542968, 0.02298784065246582, 0.022905792236328125, 0.02280793571472168, 0.022815359115600585, 0.0228351993560791, 0.02288844871520996, 0.022814176559448243, 0.02286169624328613, 0.022851648330688475, 0.022878976821899415, 0.022876096725463868, 0.0229039363861084, 0.02291494369506836, 0.022864255905151367, 0.02293609619140625, 0.023054176330566407, 0.023242912292480468, 0.023400447845458985, 0.02332467269897461, 0.023658496856689453, 0.023674047470092774, 0.023537696838378905, 0.023523647308349608, 0.023535423278808594, 0.023412992477416992, 0.023525535583496095, 0.023463743209838867, 0.0234003849029541, 0.023058944702148438, 0.023007232666015624, 0.023011327743530274, 0.023418880462646483, 0.023377920150756838, 0.023396095275878905, 0.023201055526733398, 0.023237632751464843, 0.023060447692871095, 0.02307481575012207, 0.022931455612182617, 0.023418624877929686, 0.023279584884643555, 0.02318569564819336, 0.023214080810546874, 0.023185407638549805, 0.023125856399536134, 0.02307609558105469, 0.02327030372619629, 0.023551136016845702, 0.023488895416259765, 0.023220703125, 0.022932960510253907, 0.02292176055908203, 0.023045631408691408, 0.023044160842895508, 0.023189952850341797, 0.022932640075683595, 0.022862688064575195, 0.02333830451965332, 0.02300547218322754, 0.023264768600463868, 0.023192384719848632, 0.02319468879699707, 0.023486848831176757, 0.023400480270385743, 0.023588640213012695, 0.023461984634399413, 0.023441247940063477, 0.023339775085449217, 0.02342255973815918, 0.023362016677856444, 0.02346188735961914, 0.023459711074829102, 0.02347430419921875, 0.024040672302246095, 0.0246396484375, 0.023597696304321288, 0.023119871139526366, 0.022999040603637694, 0.02310745620727539, 0.023452896118164063, 0.023239583969116212, 0.023109216690063477, 0.023009567260742186, 0.022970495223999025, 0.023078527450561524, 0.023681407928466798, 0.022992895126342772, 0.022939647674560547, 0.022998783111572267, 0.02365670394897461, 0.0230699520111084, 0.022907232284545897, 0.024906143188476563, 0.024585887908935546, 0.023769439697265624, 0.023363487243652344, 0.023427167892456056, 0.023173120498657225, 0.02300851249694824, 0.02295680046081543, 0.023078432083129884, 0.023366111755371094, 0.023473247528076172, 0.026989376068115235, 0.023277280807495117, 0.0229967041015625, 0.022915744781494142, 0.022898687362670898, 0.023001087188720702, 0.02288643264770508, 0.02298240089416504, 0.02295625686645508, 0.023109632492065428, 0.022994016647338866, 0.02305731201171875, 0.022986751556396484, 0.023049983978271484, 0.02296575927734375, 0.023165695190429686, 0.02291916847229004, 0.02315807914733887, 0.023081567764282225, 0.023060480117797853, 0.023009279251098632, 0.023119871139526366, 0.02305843162536621, 0.022824960708618162, 0.023003135681152344, 0.02288332748413086, 0.02308310317993164, 0.022897567749023438, 0.02291097640991211, 0.022835071563720704, 0.022980735778808593, 0.02295225524902344, 0.022984384536743164, 0.02280243110656738, 0.022951936721801756, 0.023001087188720702, 0.02290483283996582, 0.022771615982055664, 0.02298019218444824, 0.022829599380493164, 0.02293552017211914, 0.022749183654785156, 0.02285158348083496, 0.022794048309326173, 0.02288435173034668, 0.02282624053955078, 0.022867935180664063, 0.022764511108398437, 0.022865888595581054, 0.02287151908874512, 0.022939647674560547, 0.02285420799255371, 0.022896352767944335, 0.02277507209777832, 0.022846464157104493, 0.02327452850341797, 0.023354335784912108, 0.023005184173583985, 0.022975872039794922, 0.022884992599487303, 0.02292300796508789, 0.022775104522705078, 0.022848447799682616, 0.023326719284057617, 0.022890495300292968, 0.022761472702026365, 0.023013376235961915, 0.02287615966796875, 0.022994464874267578, 0.02289302444458008, 0.023127071380615233, 0.02287926483154297, 0.023035839080810548, 0.02288787269592285, 0.02297657585144043, 0.02297088050842285, 0.022974431991577147, 0.022887935638427736, 0.02290060806274414, 0.02293827247619629, 0.02301139259338379, 0.023322559356689452]",tokens/s,43.217018502246795,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,7334.526976,9673.048064,0.0,9277.800448,8679.633408,s,1,14.6659150390625,14.6659150390625,0.0,14.6659150390625,14.6659150390625,14.6659150390625,14.6659150390625,[14.6659150390625],,kWh,0.00022638449367916183,2.496456093734959e-05,0.00010217091507000453,0.0003535199696865159,,MB,1633.624064,9687.728128,0.0,9277.800448,8206.444544,s,10,3.639275939941406,0.3639275939941406,0.0009292388946834563,0.3641373748779297,0.36494803466796877,0.36504625244140626,0.3651248266601563,"[0.36514447021484375, 0.3631889343261719, 0.36265203857421874, 0.363911865234375, 0.36459429931640625, 0.362269287109375, 0.36436288452148435, 0.36359759521484375, 0.36492620849609375, 0.36462835693359374]",tokens/s,703.4366292216954,kWh,1.0648591585416512e-05,1.1743468569112036e-06,7.045541350713927e-06,1.886847979304164e-05,tokens/kWh,13567600.718655048,MB,1637.482496,9687.728128,0.0,9277.800448,8483.104256,s,10,24.1464248046875,2.41464248046875,0.0032581539207504878,2.4154171142578127,2.4186192626953122,2.418891174316406,2.419108703613281,"[2.409552490234375, 2.41047900390625, 2.411179931640625, 2.415583984375, 2.4128466796875, 2.415250244140625, 2.418558837890625, 2.416442138671875, 2.4191630859375, 2.417368408203125]",tokens/s,26.09081903825776,kWh,7.050408184916646e-05,7.776659843467544e-06,4.701366856328689e-05,0.00012529441025592087,tokens/kWh,502815.7271447223,,s,630,24.142780403137223,0.038321873655773345,0.0003929093673359112,0.038308799743652346,0.03883981666564942,0.038901848220825194,0.03900824077606201,"[0.03857583999633789, 0.037775646209716796, 0.037539840698242184, 0.037614879608154295, 0.03764025497436523, 0.03761129760742187, 0.037563007354736326, 0.0376814079284668, 0.03783785629272461, 0.037826976776123046, 0.03775955200195313, 0.037741825103759764, 0.03778412628173828, 0.037865665435791014, 0.03782656097412109, 0.03779919815063477, 0.03781500625610352, 0.037804031372070314, 0.03781017684936523, 0.03796582412719727, 0.03805712127685547, 0.038060897827148436, 0.03808844757080078, 0.03813750457763672, 0.03818761444091797, 0.03812895965576172, 0.038124031066894534, 0.0381049919128418, 0.038050079345703126, 0.03806585693359375, 0.038029632568359374, 0.03810841751098633, 0.03817753601074219, 0.03830771255493164, 0.03822195053100586, 0.03827014541625977, 0.03833119964599609, 0.03832831954956055, 0.03827228927612305, 0.03833516693115235, 0.038515968322753905, 0.03848886489868164, 0.038408191680908206, 0.03844095993041992, 0.03840732955932617, 0.03849097442626953, 0.03859628677368164, 0.03869289779663086, 0.03867219161987305, 0.03879894256591797, 0.03871324920654297, 0.038925121307373044, 0.03890192031860352, 0.03879731369018555, 0.0386761589050293, 0.03863724899291992, 0.03876927947998047, 0.038811168670654296, 0.03883465576171875, 0.03886000061035156, 0.0388267822265625, 0.03886016082763672, 0.03880409622192383, 0.038560031890869144, 0.03792035293579102, 0.03766518402099609, 0.037703231811523436, 0.03769708633422852, 0.037668800354003905, 0.037635009765625, 0.0376995849609375, 0.037720062255859374, 0.03783200073242188, 0.03784569549560547, 0.037787647247314454, 0.037754878997802735, 0.037852607727050784, 0.037782081604003905, 0.037795360565185544, 0.03777788925170898, 0.03786751937866211, 0.03798777770996094, 0.03805606460571289, 0.037980609893798825, 0.03801657485961914, 0.03796217727661133, 0.037973281860351565, 0.038141918182373044, 0.03822665786743164, 0.03819113540649414, 0.03807392120361328, 0.038121921539306644, 0.03818905639648437, 0.0381376953125, 0.03819084930419922, 0.0383553581237793, 0.038458366394042966, 0.03842876815795898, 0.03840911865234375, 0.03836921691894531, 0.038541057586669925, 0.038426944732666016, 0.03845939254760742, 0.038413761138916015, 0.0383776969909668, 0.03852886581420899, 0.03859260940551758, 0.03858620834350586, 0.03861670303344727, 0.03856889724731445, 0.0385445442199707, 0.038742881774902344, 0.038599681854248044, 0.03875449752807617, 0.03874803161621094, 0.038712257385253905, 0.03868163299560547, 0.038642559051513675, 0.03861872100830078, 0.038615550994873044, 0.03878092956542969, 0.038745121002197264, 0.03875939178466797, 0.0387209587097168, 0.03877536010742187, 0.03879024124145508, 0.0385816650390625, 0.03784960174560547, 0.03766396713256836, 0.03768809509277344, 0.03763824081420898, 0.03768524932861328, 0.037779457092285154, 0.037894142150878905, 0.03788390350341797, 0.03782860946655273, 0.03786483383178711, 0.03784767913818359, 0.03784425735473633, 0.037884449005126955, 0.03784096145629883, 0.037844768524169924, 0.03785311889648438, 0.03789148712158203, 0.037956607818603515, 0.03802316665649414, 0.03811238479614258, 0.03809686279296875, 0.038056865692138675, 0.03800883102416992, 0.03812761688232422, 0.038400001525878906, 0.03826278305053711, 0.03814140701293945, 0.03809689712524414, 0.0380994873046875, 0.03811123275756836, 0.03815423965454102, 0.03824012756347656, 0.03822195053100586, 0.03821977615356445, 0.038299648284912106, 0.038305728912353516, 0.03838547134399414, 0.0384697265625, 0.03834812927246094, 0.03830252838134766, 0.0383631362915039, 0.03833651351928711, 0.03841228866577148, 0.038507553100585935, 0.03856067276000977, 0.03850451278686524, 0.038556926727294924, 0.03853388977050781, 0.03863273620605469, 0.03856662368774414, 0.038555648803710936, 0.038645759582519534, 0.038844417572021485, 0.038811168670654296, 0.03885513687133789, 0.03876454544067383, 0.03890176010131836, 0.03891795349121094, 0.0389183349609375, 0.03883612823486328, 0.03889775848388672, 0.03902000045776367, 0.03836476898193359, 0.03778601455688477, 0.03763814544677734, 0.03763600158691406, 0.03767305755615234, 0.03779542541503906, 0.037720481872558595, 0.03773235321044922, 0.03770307159423828, 0.0378201904296875, 0.037788478851318356, 0.03791462326049805, 0.03794124984741211, 0.038031295776367186, 0.03792083358764648, 0.03790643310546875, 0.03790848159790039, 0.037904384613037106, 0.037887649536132814, 0.038109535217285155, 0.038117374420166016, 0.038158302307128904, 0.0379692497253418, 0.03799488067626953, 0.03808492660522461, 0.03811894226074219, 0.03819308853149414, 0.038140449523925785, 0.038365184783935545, 0.03837747192382813, 0.038340576171875, 0.03835292816162109, 0.038373374938964845, 0.03845119857788086, 0.03844255828857422, 0.03856838226318359, 0.038506240844726564, 0.03853952026367188, 0.03864371109008789, 0.03854524612426758, 0.03859676742553711, 0.038723201751708985, 0.038613121032714845, 0.03854975891113281, 0.03865375900268555, 0.038691009521484375, 0.03868672180175781, 0.03864985656738281, 0.038741214752197266, 0.03875100708007812, 0.03874816131591797, 0.0387677116394043, 0.038679454803466795, 0.03884163284301758, 0.038849056243896486, 0.038797504425048826, 0.03875136184692383, 0.038851039886474606, 0.03899334335327148, 0.03891299057006836, 0.03889926528930664, 0.0392053108215332, 0.038793216705322264, 0.038599807739257815, 0.03794742584228516, 0.037808990478515624, 0.037717727661132815, 0.03768348693847656, 0.037782913208007814, 0.037835391998291015, 0.03790643310546875, 0.037889022827148434, 0.03793407821655274, 0.03789004898071289, 0.037953536987304685, 0.037904384613037106, 0.03793920135498047, 0.03791027069091797, 0.037932510375976565, 0.03788880157470703, 0.03806345748901367, 0.03818972778320313, 0.03833446502685547, 0.038188255310058594, 0.03811724853515625, 0.03809782409667969, 0.03800883102416992, 0.038141822814941405, 0.038258209228515624, 0.038255199432373044, 0.03830579376220703, 0.03829759979248047, 0.03824585723876953, 0.03820393753051758, 0.038153343200683594, 0.03818585586547851, 0.03819497680664063, 0.038291679382324216, 0.03830537414550781, 0.03832867050170898, 0.03832633590698242, 0.03831398391723633, 0.038414337158203124, 0.03837094497680664, 0.038420032501220704, 0.03846227264404297, 0.03860883331298828, 0.03855980682373047, 0.038569534301757816, 0.03856995010375976, 0.038523361206054686, 0.03858227157592774, 0.03865190505981445, 0.0385986557006836, 0.03870851135253906, 0.03870105743408203, 0.03878780746459961, 0.038739456176757815, 0.03871139144897461, 0.03865977478027344, 0.03868745422363281, 0.038688606262207034, 0.03881795120239258, 0.03883433532714844, 0.038784862518310544, 0.03876249694824219, 0.0386866569519043, 0.03793423843383789, 0.03774556732177734, 0.03773980712890625, 0.03768937683105469, 0.037708480834960936, 0.03769343948364258, 0.03774185562133789, 0.03777199935913086, 0.037766143798828124, 0.03773478317260742, 0.03776752090454102, 0.03787190246582031, 0.03793222427368164, 0.03795795059204102, 0.03791414260864258, 0.03792585754394531, 0.03788390350341797, 0.0379576301574707, 0.0379576301574707, 0.03812063980102539, 0.03811564636230469, 0.038091262817382815, 0.03802649688720703, 0.038123329162597655, 0.03820640182495117, 0.03821763229370117, 0.038227935791015626, 0.03820556640625, 0.0382597770690918, 0.03828627014160156, 0.038295040130615236, 0.03822611236572265, 0.0382938232421875, 0.03823206329345703, 0.03833651351928711, 0.038424575805664066, 0.038504447937011715, 0.038483585357666016, 0.038496639251708986, 0.03849622344970703, 0.03851779174804688, 0.03848908615112305, 0.03850239944458008, 0.038537025451660156, 0.03875235366821289, 0.03876464080810547, 0.03860684967041016, 0.038688385009765625, 0.03978812789916992, 0.0387380485534668, 0.038707489013671874, 0.03882841491699219, 0.03878297424316406, 0.03895199966430664, 0.03881228637695312, 0.03893065643310547, 0.03897967910766602, 0.03890380859375, 0.03888742446899414, 0.03885039901733398, 0.03895657730102539, 0.03888355255126953, 0.0388485107421875, 0.038019073486328124, 0.03775849533081055, 0.03778607940673828, 0.03777452850341797, 0.03779024124145508, 0.03786323165893555, 0.037937503814697265, 0.03796329498291016, 0.03796819305419922, 0.03789648056030273, 0.03792819213867187, 0.0379788818359375, 0.037986305236816405, 0.03788800048828125, 0.03787273788452149, 0.03795036697387695, 0.037988353729248046, 0.03794124984741211, 0.03805152130126953, 0.03819142532348633, 0.03813776016235351, 0.0381952018737793, 0.03828745651245117, 0.03833366394042969, 0.03841417694091797, 0.03828627014160156, 0.03823820877075195, 0.03815119934082031, 0.038257633209228516, 0.03829542541503906, 0.038319839477539065, 0.038340641021728517, 0.03844124984741211, 0.0384697265625, 0.03842009735107422, 0.03842700958251953, 0.038558719635009765, 0.03851161575317383, 0.038629344940185543, 0.038543392181396484, 0.03850239944458008, 0.038704769134521484, 0.038683006286621094, 0.038768638610839845, 0.038795265197753906, 0.03874816131591797, 0.038690174102783206, 0.03865663909912109, 0.03880550384521484, 0.038838241577148436, 0.03886633682250976, 0.03892428970336914, 0.03886959838867188, 0.03875433731079102, 0.038836223602294925, 0.038795265197753906, 0.038905567169189456, 0.03894822311401367, 0.038818401336669923, 0.03891436767578125, 0.03883961486816406, 0.03888172912597656, 0.03857555389404297, 0.03783523178100586, 0.03770582580566406, 0.03772825622558594, 0.03785279846191406, 0.037900543212890624, 0.0377918701171875, 0.03775897598266602, 0.03776921463012695, 0.03794124984741211, 0.037857086181640624, 0.03787094497680664, 0.03787452697753906, 0.03787776184082031, 0.03786073684692383, 0.037910465240478516, 0.037880512237548826, 0.037959136962890626, 0.03795817565917969, 0.03826454544067383, 0.03820982360839844, 0.03823616027832031, 0.03818905639648437, 0.03822182464599609, 0.0382355842590332, 0.0382490234375, 0.03820748901367187, 0.03820748901367187, 0.03829350280761719, 0.03837542343139649, 0.038332416534423826, 0.038294784545898436, 0.03825945663452148, 0.03829759979248047, 0.03841558456420899, 0.03854528045654297, 0.03848604965209961, 0.03850739288330078, 0.038540481567382816, 0.038555904388427736, 0.03858812713623047, 0.03860771179199219, 0.038637569427490234, 0.03857126235961914, 0.03859532928466797, 0.03871670532226563, 0.038693599700927735, 0.038637569427490234, 0.038741310119628905, 0.0387303352355957, 0.038670337677001954, 0.03871084976196289, 0.038724128723144534, 0.03889670562744141, 0.038917057037353514, 0.03887923049926758, 0.038782878875732424, 0.038803424835205075, 0.03886297607421875, 0.03889120101928711, 0.03887472152709961, 0.038849246978759765, 0.03888496017456055, 0.03861094284057617, 0.037934432983398436, 0.03767907333374024, 0.037690048217773435, 0.03764217758178711, 0.037719486236572265, 0.03771660614013672, 0.03784499359130859, 0.03787980651855469, 0.037925983428955076, 0.03794627380371094, 0.03792473602294922, 0.037934303283691406, 0.037990814208984376, 0.03799091339111328, 0.03794281768798828, 0.03799702453613281, 0.037966976165771486, 0.03808345413208008, 0.038223648071289064, 0.03821369552612305, 0.03816259384155273, 0.0381952018737793, 0.03815190505981445, 0.038317569732666014, 0.038294303894042966, 0.03829116821289062, 0.03835728073120117, 0.03827711868286133, 0.03829283142089844, 0.03829779052734375, 0.03833699035644531, 0.03842015838623047, 0.03851443099975586, 0.03854188919067383, 0.038785022735595705, 0.03859251022338867, 0.03862694549560547, 0.03858060836791992, 0.03860275268554687, 0.038550559997558596, 0.03857670211791992, 0.038544830322265626, 0.03858844757080078, 0.038646720886230466, 0.03863929748535156, 0.038723648071289064, 0.03874550247192383, 0.038797504425048826, 0.0388590087890625, 0.038801311492919925, 0.038832065582275394, 0.038763072967529295, 0.03876454544067383, 0.03877478408813476, 0.038903457641601566, 0.03889187240600586, 0.038965248107910154, 0.039003265380859374, 0.03894771194458008, 0.03888032150268555, 0.039010272979736325, 0.03905020904541016, 0.038674720764160155, 0.03798697662353516, 0.03767705535888672, 0.03768323135375977, 0.03772134399414063, 0.0377303695678711, 0.03779033660888672, 0.03785862350463867, 0.03780022430419922, 0.03791097640991211, 0.037881118774414066, 0.037868255615234374, 0.038002689361572264, 0.037975391387939456, 0.037949119567871094, 0.037999584197998044, 0.03797401428222656, 0.037961505889892576, 0.03807664108276367, 0.038133758544921875, 0.03818086242675781, 0.03814368057250977, 0.038072032928466795, 0.03815033721923828, 0.038236576080322264, 0.03830988693237305, 0.03832012939453125, 0.03822796630859375, 0.03823820877075195, 0.03829340744018555, 0.03823011016845703, 0.038266880035400394, 0.03830748748779297, 0.03845094299316406, 0.0384436149597168, 0.038438911437988284, 0.03844300842285156, 0.038419807434082034, 0.038369953155517576, 0.03854742431640625, 0.03843411254882813, 0.03844579315185547, 0.038532958984375, 0.03861315155029297, 0.03862694549560547, 0.038736255645751956, 0.03876995086669922, 0.03872227096557617, 0.03864166259765625, 0.03870300674438477, 0.038750049591064456, 0.03889120101928711, 0.038923934936523436, 0.038886302947998046, 0.03882169723510742, 0.038895809173583984, 0.03897679901123047, 0.038994625091552736, 0.03893660736083984, 0.03890108871459961, 0.03902531051635742, 0.039041023254394534, 0.03895827102661133]",tokens/s,26.09475750018152,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,8205.9264,12277.710848,0.0,11882.463232,11315.947008,s,1,16.886255859375,16.886255859375,0.0,16.886255859375,16.886255859375,16.886255859375,16.886255859375,[16.886255859375],,kWh,0.00029382946845003057,3.239582406865539e-05,0.00012235482010597432,0.00044858011262466027,,MB,4044.82048,12294.488064,0.0,11884.560384,11070.308352,s,10,3.9773246765136716,0.3977324676513672,0.0019224772243556913,0.3980414886474609,0.39965287475585937,0.40012190856933594,0.4004971356201172,"[0.3930919189453125, 0.39734170532226565, 0.39863742065429686, 0.39619915771484376, 0.39800732421875, 0.3980756530761719, 0.3974848022460937, 0.39954864501953125, 0.39834710693359376, 0.4005909423828125]",tokens/s,643.6487358240943,kWh,1.1554067951282704e-05,1.2737012906724567e-06,7.650198427846875e-06,2.0477967669802032e-05,tokens/kWh,12501240.559018562,MB,4049.084416,12296.585216,0.0,11886.657536,11070.310912,s,10,25.94716357421875,2.594716357421875,0.0037278301135179426,2.5959627685546875,2.598407739257812,2.599555920410156,2.6004744653320313,"[2.589595458984375, 2.590308837890625, 2.589087158203125, 2.5964873046875, 2.597010498046875, 2.595438232421875, 2.6007041015625, 2.593673095703125, 2.596706298828125, 2.598152587890625]",tokens/s,24.28011054842124,kWh,7.587325129079908e-05,8.36829190384015e-06,5.051604254955066e-05,0.00013475758574418988,tokens/kWh,467506.1492982874,,s,630,25.943864662170412,0.04118073755900065,0.00041361422757583017,0.041152158737182615,0.04148256874084473,0.04162520122528076,0.04329115642547607,"[0.04304620742797852, 0.041142974853515625, 0.04068518447875977, 0.04066345596313477, 0.04065884780883789, 0.04080031967163086, 0.04055449676513672, 0.041585983276367186, 0.04063711929321289, 0.04067324829101562, 0.04066275024414062, 0.040726497650146486, 0.04103728103637695, 0.04088918304443359, 0.04075475311279297, 0.04083145523071289, 0.04086761474609375, 0.04097660827636719, 0.04093308639526367, 0.04084793472290039, 0.04087548828125, 0.04087228775024414, 0.041002880096435546, 0.04102352142333984, 0.04114636611938476, 0.04117830276489258, 0.04121401596069336, 0.041073406219482425, 0.04101939010620117, 0.04103577423095703, 0.040928417205810544, 0.04102969741821289, 0.041032352447509766, 0.041025665283203124, 0.040908798217773434, 0.040925472259521485, 0.04104544067382813, 0.04109135818481445, 0.04109635162353516, 0.04102025604248047, 0.0409536018371582, 0.04107468795776367, 0.0412407341003418, 0.041210113525390626, 0.04124787139892578, 0.04132937622070312, 0.04231507110595703, 0.041175743103027344, 0.04117913436889648, 0.041283424377441404, 0.04131564712524414, 0.041243328094482425, 0.04143468856811523, 0.04141638565063477, 0.041294273376464845, 0.0413967056274414, 0.04133267211914062, 0.041187614440917966, 0.041223617553710935, 0.041191585540771486, 0.04113945770263672, 0.0413111686706543, 0.04125696182250976, 0.043718593597412106, 0.04122236633300781, 0.04070380783081055, 0.04063030242919922, 0.04066918563842774, 0.040574977874755856, 0.04062412643432617, 0.04071177673339844, 0.04063068771362305, 0.040586593627929685, 0.04067190551757813, 0.040793663024902345, 0.0407677116394043, 0.04069807815551758, 0.04061167907714844, 0.040736446380615236, 0.04100451278686523, 0.041032703399658206, 0.04080844879150391, 0.04090876770019531, 0.0409536018371582, 0.04086403274536133, 0.041390079498291016, 0.041390079498291016, 0.041471038818359375, 0.04128659057617187, 0.041203071594238284, 0.04100531387329102, 0.04100723266601562, 0.0408653450012207, 0.04086240005493164, 0.040804351806640625, 0.04080025482177734, 0.04099673461914063, 0.040951934814453125, 0.04104806518554688, 0.04114432144165039, 0.041233695983886716, 0.04114883041381836, 0.04108319854736328, 0.04126924896240235, 0.04119148635864258, 0.041170654296875, 0.04131967926025391, 0.04120038223266602, 0.04140054321289063, 0.041282943725585934, 0.041482559204101564, 0.04266815948486328, 0.04129606246948242, 0.04120899200439453, 0.041169471740722656, 0.04131568145751953, 0.04120207977294922, 0.04121654510498047, 0.041170879364013674, 0.0411786880493164, 0.04113974380493164, 0.04120787048339844, 0.041317279815673826, 0.041373695373535156, 0.04130508804321289, 0.04132761764526367, 0.04329564666748047, 0.041224193572998044, 0.04083091354370117, 0.04070406341552734, 0.04072857666015625, 0.040855552673339846, 0.04079411315917969, 0.04072447967529297, 0.04075020980834961, 0.04069059371948242, 0.040771102905273436, 0.04081913757324219, 0.04077772903442383, 0.04089651107788086, 0.040855552673339846, 0.0409354248046875, 0.04085964965820312, 0.04084326553344726, 0.04086793518066406, 0.04105628967285156, 0.04102336120605469, 0.040925182342529294, 0.04103168106079102, 0.04125286483764649, 0.041260639190673826, 0.041162689208984374, 0.04109257507324219, 0.041067615509033206, 0.0409640007019043, 0.040822784423828126, 0.04078742218017578, 0.0407737922668457, 0.040804737091064455, 0.04098457717895508, 0.04095180892944336, 0.04085964965820312, 0.04087795257568359, 0.04086732864379883, 0.040852096557617186, 0.04096540832519531, 0.04100780868530273, 0.04104307174682617, 0.04131923294067383, 0.04129110336303711, 0.04147071838378906, 0.04155187225341797, 0.04148223876953125, 0.041463294982910154, 0.04153539276123047, 0.041441600799560545, 0.04140060806274414, 0.0412938232421875, 0.04129536056518555, 0.04129824066162109, 0.04126534271240234, 0.04118096160888672, 0.04115420913696289, 0.04121247863769531, 0.041164798736572264, 0.041338878631591795, 0.041381889343261716, 0.04128153610229492, 0.04126310348510742, 0.04327043151855469, 0.04138134384155273, 0.040948032379150394, 0.04075990295410156, 0.04076927947998047, 0.040820545196533206, 0.0407815055847168, 0.044096160888671875, 0.04069580841064453, 0.040697856903076174, 0.04073062515258789, 0.040775360107421874, 0.040890113830566406, 0.04084940719604492, 0.04082912063598633, 0.04078400039672852, 0.040812801361083985, 0.04086950302124023, 0.04100339126586914, 0.04107843017578125, 0.0410906867980957, 0.041151199340820316, 0.04120560073852539, 0.0413812141418457, 0.04129017639160156, 0.041217758178710935, 0.04114684677124023, 0.04101548767089844, 0.04100080108642578, 0.04111990356445312, 0.041100959777832034, 0.04094569778442383, 0.040929023742675784, 0.04098310470581055, 0.04196352005004883, 0.04114176177978516, 0.04111315155029297, 0.04098700714111328, 0.041084800720214844, 0.04118710327148437, 0.041184158325195314, 0.04115456008911133, 0.04122009658813477, 0.041218143463134765, 0.041387935638427735, 0.04146585464477539, 0.04159814453125, 0.0415404167175293, 0.04151295852661133, 0.041391265869140624, 0.04136758422851562, 0.04136838531494141, 0.04126310348510742, 0.04122009658813477, 0.04129977416992187, 0.04117900848388672, 0.0412163200378418, 0.0413076171875, 0.04124931335449219, 0.041267200469970705, 0.04115660858154297, 0.0413573112487793, 0.04134108734130859, 0.04328243255615234, 0.041371105194091796, 0.04089705657958984, 0.04063235092163086, 0.040818462371826174, 0.04085299301147461, 0.04089516830444336, 0.04086783981323242, 0.04083846282958985, 0.04078662490844726, 0.04085916900634766, 0.04078435134887695, 0.04079801559448242, 0.040747230529785156, 0.04086492919921875, 0.04076367950439453, 0.04098716735839844, 0.041095169067382815, 0.041336830139160154, 0.041306110382080076, 0.041111553192138675, 0.04117094421386719, 0.04140851211547852, 0.041398273468017575, 0.04130099105834961, 0.041261215209960934, 0.041113983154296874, 0.04098710250854492, 0.040900672912597656, 0.040957889556884765, 0.041095169067382815, 0.04106607818603516, 0.04109353637695313, 0.04114236831665039, 0.04129119873046875, 0.04119315338134766, 0.04123747253417969, 0.041146175384521484, 0.04107651138305664, 0.041226463317871095, 0.041164798736572264, 0.04122803115844727, 0.04139033508300781, 0.04143513488769531, 0.04142419052124023, 0.04171004867553711, 0.041737823486328124, 0.04164451217651367, 0.041734432220458986, 0.041463680267333984, 0.04137350463867188, 0.04139145660400391, 0.041395038604736326, 0.041397918701171876, 0.04141910552978516, 0.04124467086791992, 0.04126022338867188, 0.041216831207275394, 0.04131951904296875, 0.04147702407836914, 0.041441280364990236, 0.04135286331176758, 0.04146790313720703, 0.04349507141113281, 0.0413680305480957, 0.04082259368896484, 0.040634334564208986, 0.04075734329223633, 0.040791999816894534, 0.04088147354125977, 0.04284880065917969, 0.04063868713378906, 0.04085968017578125, 0.04095974349975586, 0.04088604736328125, 0.04080643081665039, 0.0408191032409668, 0.040994815826416016, 0.04099484634399414, 0.04095734405517578, 0.04099116897583008, 0.041201793670654296, 0.04112524795532226, 0.04122060775756836, 0.041211040496826175, 0.04135011291503906, 0.04135718536376953, 0.0413779182434082, 0.04137139129638672, 0.04117248153686524, 0.0410775032043457, 0.040970241546630856, 0.041003009796142575, 0.04112179183959961, 0.04103968048095703, 0.040933441162109375, 0.040978561401367186, 0.040892414093017575, 0.040869888305664064, 0.04092067337036133, 0.04087039947509766, 0.04094537734985351, 0.040945854187011715, 0.04097820663452149, 0.04111177444458008, 0.04121395111083984, 0.04127129745483398, 0.04131033706665039, 0.04129324722290039, 0.04162953567504883, 0.04172009658813477, 0.04161990356445312, 0.04149440002441406, 0.04152681732177734, 0.041322975158691405, 0.04127859115600586, 0.041237377166748045, 0.04114950561523437, 0.04126611328125, 0.0412586898803711, 0.0412204475402832, 0.04126051330566406, 0.041259521484375, 0.041381855010986325, 0.04145084762573242, 0.041366207122802735, 0.043733089447021485, 0.04141007995605469, 0.040938270568847655, 0.040857921600341796, 0.040861473083496094, 0.04081039810180664, 0.04081478500366211, 0.04091872024536133, 0.040834880828857424, 0.04089907073974609, 0.04092278289794922, 0.04094521713256836, 0.0409403190612793, 0.04087398529052735, 0.04086982345581055, 0.04078598403930664, 0.040888160705566404, 0.040925342559814455, 0.0411212158203125, 0.041344799041748044, 0.04131305694580078, 0.04128992080688477, 0.04151059341430664, 0.04151103973388672, 0.04152048110961914, 0.04144604873657227, 0.04143513488769531, 0.04119507217407226, 0.04104227066040039, 0.041154144287109375, 0.04115302276611328, 0.041045761108398436, 0.04104217529296875, 0.04181350326538086, 0.04113049697875976, 0.04114636611938476, 0.04105152130126953, 0.0412410888671875, 0.04119769668579101, 0.0410503044128418, 0.04115372848510742, 0.04126988983154297, 0.04149459075927735, 0.041666496276855466, 0.04167679977416992, 0.04158259201049805, 0.041702560424804684, 0.0415362548828125, 0.04164751815795899, 0.04154812622070313, 0.04144342422485352, 0.04144153594970703, 0.04140614318847656, 0.04138016128540039, 0.04135644912719726, 0.04135110473632812, 0.04128646469116211, 0.041304126739501956, 0.04132799911499024, 0.041396480560302734, 0.041268993377685546, 0.041337505340576175, 0.041790817260742186, 0.04329471969604492, 0.04136345672607422, 0.04087113571166992, 0.040733470916748046, 0.04086697769165039, 0.041032352447509766, 0.04115670394897461, 0.040826465606689455, 0.04081919860839844, 0.040793407440185545, 0.04079654312133789, 0.040845535278320313, 0.0409334716796875, 0.04085145568847656, 0.04085939025878906, 0.04084467315673828, 0.04086179351806641, 0.040839969635009764, 0.04109270477294922, 0.040968448638916015, 0.040933536529541015, 0.04091292953491211, 0.04106441497802735, 0.041307552337646485, 0.04123503875732422, 0.04114361572265625, 0.041156639099121095, 0.04102006530761719, 0.04105420684814453, 0.0410742073059082, 0.04119776153564453, 0.041037696838378906, 0.04109507369995117, 0.04114246368408203, 0.041160190582275394, 0.04110009765625, 0.04105215835571289, 0.04116396713256836, 0.04106032180786133, 0.04102025604248047, 0.04125696182250976, 0.041488384246826174, 0.041404415130615234, 0.04149878311157226, 0.04152268981933594, 0.04137318420410156, 0.041378654479980466, 0.041411582946777346, 0.04146688079833984, 0.04137286376953125, 0.04134147262573242, 0.04121014404296875, 0.04127059173583984, 0.0413067512512207, 0.04140447998046875, 0.04119955062866211, 0.041295585632324217, 0.04116831970214844, 0.041214241027832034, 0.04136342239379883, 0.041290302276611325, 0.041302112579345705, 0.041250816345214845, 0.04327228927612305, 0.04131020736694336, 0.04089440155029297, 0.040779518127441405, 0.04078755187988281, 0.040671966552734376, 0.04069375991821289, 0.04080230331420898, 0.040845439910888674, 0.040929153442382814, 0.04094927978515625, 0.04088675308227539, 0.040894462585449216, 0.04088790512084961, 0.040874401092529294, 0.04093734359741211, 0.04092700958251953, 0.04098287963867187, 0.041070465087890626, 0.04132220840454102, 0.0410442886352539, 0.04105583953857422, 0.041140735626220705, 0.041262081146240234, 0.041253887176513675, 0.042757312774658204, 0.04095375823974609, 0.04106537628173828, 0.04110540771484375, 0.04112371063232422, 0.04095808029174805, 0.040959999084472655, 0.04106854248046875, 0.04117504119873047, 0.04110745620727539, 0.041011199951171876, 0.040945758819580076, 0.040976158142089845, 0.04120528030395508, 0.04108553695678711, 0.0410107536315918, 0.041296321868896486, 0.041443134307861326, 0.041375934600830076, 0.041527294158935545, 0.041678848266601565, 0.041637889862060545, 0.04153926467895508, 0.04154732894897461, 0.04148441696166992, 0.041512737274169924, 0.04143804931640625, 0.041350399017333984, 0.041339839935302734, 0.04138995361328125, 0.04124665451049805, 0.04138172912597656, 0.04127356719970703, 0.04137567901611328, 0.0413941764831543, 0.04133184051513672, 0.041384830474853515, 0.041404415130615234, 0.04336435317993164, 0.04125696182250976, 0.040890113830566406, 0.04073289489746094, 0.04072246551513672, 0.040818145751953125, 0.04090934371948242, 0.0410082893371582, 0.040991008758544924, 0.04099334335327148, 0.04100505447387695, 0.041064254760742186, 0.04090284729003906, 0.04091289520263672, 0.04095084762573242, 0.040960960388183594, 0.040796161651611325, 0.04091888046264648, 0.04096150588989258, 0.04090950393676758, 0.04102259063720703, 0.041210750579833984, 0.041221569061279294, 0.04129644775390625, 0.04132044982910156, 0.041336830139160154, 0.041148414611816404, 0.04103168106079102, 0.04095590209960937, 0.04106198501586914, 0.0410810546875, 0.04119577789306641, 0.041097152709960935, 0.041215774536132815, 0.041102718353271485, 0.041151294708251955, 0.041148448944091795, 0.041199169158935546, 0.041228736877441406, 0.04130118560791016, 0.04132742309570313, 0.0413675537109375, 0.04135289764404297, 0.0413856315612793, 0.0414826545715332, 0.041480449676513674, 0.04149168014526367, 0.04155196762084961, 0.041529727935791017, 0.04155833435058594, 0.04150179290771484, 0.04149955368041992, 0.04137984085083008, 0.04137779235839844, 0.041248542785644535, 0.04128496170043945, 0.04129267120361328, 0.04134912109375, 0.042031105041503904, 0.041737472534179684, 0.04144607925415039, 0.04140652847290039, 0.04135116958618164]",tokens/s,24.28319790453669,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,843.771904,558.825472,0.0,163.577856,154.631168,s,1,8.00877294921875,8.00877294921875,0.0,8.00877294921875,8.00877294921875,8.00877294921875,8.00877294921875,[8.00877294921875],,kWh,2.1754401229198566e-05,2.3922935190957997e-06,7.536950474007487e-06,3.168364522230185e-05,,MB,1196.183552,611.254272,0.0,201.326592,187.147776,s,30,0.19616284799575806,0.006538761599858603,0.00013730042177287423,0.006516160011291504,0.0066924351692199706,0.006713878369331359,0.0069213583135604865,"[0.007003903865814209, 0.00655238389968872, 0.006690783977508545, 0.006612927913665772, 0.006543263912200927, 0.00663702392578125, 0.006719264030456543, 0.006657216072082519, 0.006386528015136719, 0.006424960136413574, 0.0064652800559997555, 0.006520063877105713, 0.00667683219909668, 0.006707295894622802, 0.00646127986907959, 0.006395328044891358, 0.006508287906646729, 0.0066433920860290525, 0.006512256145477295, 0.006609920024871826, 0.006563680171966553, 0.006437983989715576, 0.006429791927337647, 0.006380735874176025, 0.006371295928955078, 0.006402527809143067, 0.0065049281120300294, 0.006523776054382324, 0.006439680099487304, 0.006380256175994873]",tokens/s,39151.14446220762,kWh,1.9014748516559677e-07,2.0969349838361798e-08,8.76385065928472e-08,2.9875534159680575e-07,tokens/kWh,856888444.6775599,MB,1209.9584,613.351424,0.0,203.423744,187.150336,s,30,10.121157592773438,0.33737191975911446,0.006944105613889463,0.33639990234375,0.34557210083007817,0.35074585571289063,0.35332661407470706,"[0.35142596435546875, 0.34991461181640626, 0.3447034606933594, 0.34299172973632813, 0.345089599609375, 0.34135153198242185, 0.3415032653808594, 0.33556829833984375, 0.330926025390625, 0.3306352844238281, 0.3398701477050781, 0.35410293579101565, 0.3389402160644531, 0.3327669677734375, 0.3322997741699219, 0.3302099914550781, 0.3411739501953125, 0.34000830078125, 0.3374844665527344, 0.3426356201171875, 0.3336680908203125, 0.3295469360351562, 0.331063232421875, 0.32873614501953125, 0.3288599853515625, 0.33218084716796875, 0.3349767761230469, 0.33723150634765625, 0.33184170532226565, 0.3294502258300781]",tokens/s,186.73753300210157,kWh,9.395116657307233e-06,1.0361209795953927e-06,3.490285502697509e-06,1.3921523139600134e-05,tokens/kWh,4525366.899028086,,s,1890,10.1066182656288,0.005347417071761278,0.00020955570937754628,0.005310992002487183,0.005513996982574463,0.005591457629203796,0.006038199048042297,"[0.005385119915008545, 0.005482079982757568, 0.006101247787475586, 0.0056341438293457035, 0.005629824161529541, 0.00672976016998291, 0.005650432109832764, 0.0061914238929748535, 0.005571616172790527, 0.005527455806732178, 0.005444352149963379, 0.005488063812255859, 0.005523200035095215, 0.005579360008239746, 0.0054943361282348635, 0.005511231899261474, 0.005515872001647949, 0.0054988799095153805, 0.005742591857910156, 0.0056212158203125, 0.0055874881744384765, 0.0055474238395690914, 0.005477024078369141, 0.005554111957550048, 0.0054776320457458495, 0.005543839931488037, 0.005601823806762695, 0.005607295989990235, 0.005623392105102539, 0.005591904163360596, 0.005577919960021972, 0.005604320049285889, 0.005590911865234375, 0.005745855808258057, 0.005571648120880127, 0.005567647933959961, 0.005589151859283447, 0.005579296112060547, 0.005615488052368164, 0.005543680191040039, 0.005529344081878662, 0.005546495914459228, 0.005638144016265869, 0.005679103851318359, 0.005566592216491699, 0.0055454721450805666, 0.005543647766113281, 0.005491072177886963, 0.005427487850189209, 0.005480000019073486, 0.005421504020690918, 0.005431295871734619, 0.005484799861907959, 0.005347008228302002, 0.005407872200012207, 0.005411776065826416, 0.005403872013092041, 0.0054362878799438475, 0.005437312126159668, 0.00544598388671875, 0.0054924159049987795, 0.0054160962104797365, 0.005360479831695557, 0.0053153600692749025, 0.005367712020874023, 0.0053853759765625, 0.00542416000366211, 0.005396383762359619, 0.005386240005493164, 0.00541923189163208, 0.005410592079162598, 0.005407040119171143, 0.005350240230560303, 0.005348256111145019, 0.005326752185821533, 0.005409023761749268, 0.005452672004699707, 0.005359615802764893, 0.005382719993591308, 0.005347519874572754, 0.0054431681632995605, 0.005519648075103759, 0.005527391910552979, 0.005406847953796387, 0.005470240116119384, 0.005400864124298096, 0.005375743865966797, 0.0053844480514526364, 0.0054906878471374515, 0.0053678078651428224, 0.005392384052276611, 0.005388095855712891, 0.005702943801879883, 0.005540575981140137, 0.005495999813079834, 0.005473279953002929, 0.005480447769165039, 0.0064899840354919434, 0.005605504035949707, 0.005601280212402344, 0.005525087833404541, 0.005640672206878662, 0.005521344184875488, 0.005522624015808105, 0.005513984203338623, 0.0054795842170715335, 0.00549571180343628, 0.006146048069000244, 0.0055658559799194336, 0.0065414719581604, 0.005906879901885986, 0.005848671913146973, 0.00554640007019043, 0.005784863948822021, 0.005606112003326416, 0.005646368026733398, 0.0055149121284484865, 0.005633503913879395, 0.005524352073669434, 0.005512159824371338, 0.005489727973937988, 0.005597119808197022, 0.005505023956298828, 0.006170783996582031, 0.005551360130310059, 0.005558879852294922, 0.005474720001220703, 0.005470304012298584, 0.005451072216033936, 0.005517600059509277, 0.005478816032409668, 0.005485983848571777, 0.005441952228546143, 0.005445280075073242, 0.005480735778808594, 0.005499135971069336, 0.005438560009002685, 0.005521791934967041, 0.005521952152252197, 0.005595136165618897, 0.0054988799095153805, 0.005502975940704345, 0.005529151916503906, 0.005487040042877197, 0.0054778242111206054, 0.00556828784942627, 0.005493535995483399, 0.005537600040435791, 0.005465792179107666, 0.005473087787628174, 0.005422944068908691, 0.005394559860229492, 0.005363423824310303, 0.0054579200744628905, 0.005443583965301513, 0.005449183940887451, 0.0054635839462280275, 0.005427648067474365, 0.0054494719505310055, 0.005501760005950928, 0.005416959762573242, 0.005444608211517334, 0.005440127849578857, 0.005429215908050537, 0.005460608005523682, 0.005424384117126465, 0.005442080020904541, 0.005469759941101074, 0.005509568214416504, 0.00546611213684082, 0.005488639831542969, 0.005396480083465576, 0.005404767990112305, 0.005439008235931396, 0.005396927833557129, 0.00540451192855835, 0.00544982385635376, 0.005392159938812256, 0.005439167976379395, 0.005423647880554199, 0.005486591815948487, 0.0055008320808410645, 0.005425248146057129, 0.005458975791931152, 0.005448448181152344, 0.005437664031982422, 0.0054720320701599125, 0.005490911960601807, 0.00547983980178833, 0.005371871948242187, 0.005417280197143555, 0.005408768177032471, 0.005438752174377441, 0.005416672229766845, 0.0054568958282470706, 0.005394591808319092, 0.005460063934326172, 0.005371647834777832, 0.005393472194671631, 0.005468671798706055, 0.0053632321357727055, 0.005403615951538086, 0.005375455856323242, 0.005413216114044189, 0.005373472213745117, 0.005374559879302979, 0.005378047943115235, 0.005383264064788818, 0.005393311977386474, 0.005502079963684082, 0.005413760185241699, 0.005424672126770019, 0.005460608005523682, 0.0055294399261474605, 0.0053944320678710935, 0.005398528099060058, 0.005414912223815918, 0.0053505277633667, 0.005347199916839599, 0.005383488178253174, 0.005421823978424072, 0.005342144012451172, 0.0053640642166137694, 0.005406688213348389, 0.0053654398918151856, 0.005386240005493164, 0.0053944320678710935, 0.005392096042633057, 0.005456319808959961, 0.005578783988952637, 0.005510335922241211, 0.005585536003112793, 0.005524608135223388, 0.005514111995697022, 0.005542912006378174, 0.005510015964508057, 0.005507199764251709, 0.005601280212402344, 0.005481887817382813, 0.0055426878929138185, 0.005481311798095703, 0.005440447807312012, 0.005486623764038086, 0.005530975818634033, 0.005425343990325928, 0.005460447788238525, 0.005443583965301513, 0.005418496131896973, 0.0054297599792480465, 0.0054694080352783205, 0.00546281623840332, 0.005350751876831055, 0.005356544017791748, 0.0053376641273498535, 0.005407167911529541, 0.005407872200012207, 0.005370751857757568, 0.005438560009002685, 0.005452191829681397, 0.005377535820007324, 0.005402944087982178, 0.005407423973083496, 0.005402175903320312, 0.005478943824768066, 0.005449120044708252, 0.005464288234710693, 0.005475647926330566, 0.005608672142028808, 0.005574143886566162, 0.005468416213989258, 0.00543123197555542, 0.005490752220153808, 0.005480447769165039, 0.005736735820770264, 0.005467936038970948, 0.005459904193878174, 0.0054991040229797365, 0.0055229439735412595, 0.005580448150634766, 0.005503615856170654, 0.005465248107910156, 0.005505087852478027, 0.0055071358680725095, 0.005550816059112549, 0.005564479827880859, 0.005586880207061768, 0.005504320144653321, 0.005472703933715821, 0.005478623867034912, 0.005494815826416016, 0.005451776027679443, 0.005399680137634278, 0.005429535865783691, 0.00540118408203125, 0.005423168182373047, 0.005481887817382813, 0.005480319976806641, 0.005481279850006103, 0.005490528106689453, 0.005432352066040039, 0.005464928150177002, 0.00547430419921875, 0.005417344093322754, 0.005477312088012695, 0.0054915199279785155, 0.005392384052276611, 0.005447679996490478, 0.00541596794128418, 0.005485536098480225, 0.0054336638450622554, 0.0054403839111328125, 0.005472447872161865, 0.005464799880981445, 0.005475296020507813, 0.005547071933746338, 0.0055305280685424806, 0.005451680183410645, 0.0055251197814941404, 0.0054198079109191895, 0.005342591762542725, 0.005366079807281494, 0.005405824184417724, 0.005332064151763916, 0.0053303041458129885, 0.005310880184173584, 0.00530841588973999, 0.005310336112976074, 0.005399680137634278, 0.005273663997650146, 0.005286367893218994, 0.005309919834136963, 0.0052861437797546386, 0.00532089614868164, 0.005371551990509033, 0.005295008182525635, 0.005314559936523438, 0.0053012480735778805, 0.005345600128173828, 0.005364255905151367, 0.0053003840446472165, 0.005357439994812012, 0.005298399925231933, 0.005302271842956543, 0.005287839889526367, 0.005326272010803223, 0.005319104194641113, 0.005298304080963135, 0.005308767795562744, 0.005342559814453125, 0.00532041597366333, 0.0053827519416809085, 0.005382336139678955, 0.005397408008575439, 0.005489376068115235, 0.005458111763000488, 0.005451776027679443, 0.005449728012084961, 0.005439487934112549, 0.005593440055847168, 0.005424287796020508, 0.005450047969818115, 0.005902048110961914, 0.005491168022155762, 0.005484543800354004, 0.005573919773101807, 0.005421855926513672, 0.0057404799461364744, 0.005465983867645264, 0.005545695781707763, 0.005481056213378906, 0.005449535846710205, 0.0054860801696777345, 0.005499648094177246, 0.00546995210647583, 0.005496831893920898, 0.005488160133361816, 0.005460095882415771, 0.005479936122894287, 0.005456352233886718, 0.005437632083892822, 0.005457695960998535, 0.005406752109527588, 0.0053678078651428224, 0.005408736228942871, 0.0053578557968139645, 0.005351168155670166, 0.005478400230407715, 0.005406720161437988, 0.005355519771575928, 0.005388288021087646, 0.005457791805267334, 0.005400352001190186, 0.005365695953369141, 0.005707903861999512, 0.005410367965698242, 0.005402815818786621, 0.005351967811584472, 0.005445631980895996, 0.005402815818786621, 0.005368927955627442, 0.005345344066619873, 0.005397151947021484, 0.005345280170440674, 0.005400576114654541, 0.005399936199188232, 0.00537667179107666, 0.005392352104187012, 0.005436863899230957, 0.005400479793548584, 0.00538486385345459, 0.005400191783905029, 0.005392191886901855, 0.005410943984985352, 0.0053805441856384275, 0.005370975971221924, 0.005393248081207275, 0.005345344066619873, 0.00540883207321167, 0.005353312015533448, 0.00532806396484375, 0.005325727939605713, 0.005306367874145508, 0.0053567042350769044, 0.005408927917480468, 0.005432000160217285, 0.005418303966522217, 0.005528255939483643, 0.005572319984436035, 0.005502495765686035, 0.0054906558990478515, 0.005441343784332276, 0.005464032173156739, 0.005448703765869141, 0.005412896156311035, 0.005449215888977051, 0.005425631999969482, 0.005400800228118897, 0.005486368179321289, 0.0054271998405456545, 0.005448031902313233, 0.00540227222442627, 0.005396607875823974, 0.005433152198791504, 0.005421023845672607, 0.005451871871948242, 0.005431295871734619, 0.005489920139312744, 0.005437983989715576, 0.005459839820861816, 0.005384543895721436, 0.005368896007537842, 0.005370687961578369, 0.0053290238380432125, 0.0054559998512268065, 0.005342400074005127, 0.005329279899597168, 0.005379424095153808, 0.00534000015258789, 0.005369344234466553, 0.0053582720756530765, 0.005349088191986084, 0.005348800182342529, 0.005296576023101807, 0.005337279796600342, 0.005322368144989014, 0.005427743911743164, 0.005406720161437988, 0.005327968120574951, 0.00531657600402832, 0.00534009599685669, 0.005349376201629639, 0.005320703983306885, 0.0053350400924682614, 0.005333055973052978, 0.005283775806427002, 0.005322751998901368, 0.005296351909637451, 0.005328671932220459, 0.005322751998901368, 0.005328896045684814, 0.005344575881958008, 0.0053582720756530765, 0.005292031764984131, 0.005289152145385743, 0.005255424022674561, 0.0052540478706359865, 0.005237855911254883, 0.005326528072357178, 0.005237631797790528, 0.0052379522323608395, 0.005208447933197021, 0.005235455989837647, 0.005205440044403076, 0.005191135883331299, 0.005225535869598389, 0.005219423770904541, 0.005212096214294434, 0.005229279994964599, 0.005285759925842285, 0.005189727783203125, 0.005191679954528809, 0.005201920032501221, 0.005238912105560303, 0.005179423809051514, 0.0050869441032409665, 0.005190336227416993, 0.005175295829772949, 0.005246079921722412, 0.005188543796539307, 0.005201920032501221, 0.005181439876556396, 0.005242496013641357, 0.005175136089324951, 0.005197535991668701, 0.005159743785858154, 0.005189407825469971, 0.005187104225158692, 0.0052368960380554195, 0.005198368072509766, 0.005215328216552734, 0.005167327880859375, 0.00518336009979248, 0.005192512035369873, 0.005198847770690918, 0.005159967899322509, 0.005213600158691406, 0.005179967880249024, 0.005187007904052734, 0.005247104167938233, 0.00519212818145752, 0.005195136070251465, 0.005212160110473632, 0.005246623992919922, 0.005219295978546142, 0.005201920032501221, 0.005265344142913819, 0.005263584136962891, 0.0052304320335388186, 0.005269504070281982, 0.005318655967712403, 0.005371903896331787, 0.005322751998901368, 0.005295135974884033, 0.0052912960052490235, 0.005373983860015869, 0.0053814401626586916, 0.005269824028015136, 0.00525110387802124, 0.005289567947387695, 0.005276063919067383, 0.005242879867553711, 0.005251071929931641, 0.005225920200347901, 0.005657375812530518, 0.005308191776275634, 0.0052791681289672855, 0.005291935920715332, 0.00524560022354126, 0.005223807811737061, 0.005311103820800781, 0.0052420802116394044, 0.005220799922943115, 0.005219808101654053, 0.0052674880027771, 0.005321599960327148, 0.005298304080963135, 0.005306208133697509, 0.005187583923339844, 0.005277696132659912, 0.005203551769256591, 0.005247392177581787, 0.005232736110687256, 0.005188672065734863, 0.005249824047088623, 0.005236320018768311, 0.005212160110473632, 0.005236447811126709, 0.005195839881896973, 0.005210815906524658, 0.005258880138397217, 0.005237215995788574, 0.00521779203414917, 0.005234272003173828, 0.005239615917205811, 0.005251264095306397, 0.005216063976287842, 0.0052163200378417965, 0.005206240177154541, 0.005223936080932618, 0.0053554558753967285, 0.00527184009552002, 0.005275648117065429, 0.005320352077484131, 0.005227968215942382, 0.005237664222717285, 0.005224448204040527, 0.005234176158905029, 0.005202079772949219, 0.0051654081344604495, 0.005230016231536865, 0.00524073600769043, 0.0051998400688171385, 0.005226880073547363, 0.005232223987579346, 0.005288576126098633, 0.005252255916595459, 0.005219552040100097, 0.00519481611251831, 0.005190303802490234, 0.0052137279510498045, 0.005269599914550781, 0.005208447933197021, 0.005238527774810791, 0.005237215995788574, 0.005231584072113037, 0.0054767999649047855, 0.005292416095733643, 0.005261312007904053, 0.0052425599098205565, 0.005225855827331543, 0.005227456092834473, 0.005270847797393799, 0.005227039813995362, 0.005228096008300781, 0.005200479984283447, 0.005257376194000244, 0.005224192142486572, 0.0053105602264404295, 0.00528380823135376, 0.005275680065155029, 0.005239488124847412, 0.005244448184967041, 0.0053415679931640625, 0.005380127906799316, 0.005375904083251953, 0.005415103912353516, 0.005440512180328369, 0.005368703842163086, 0.005445631980895996, 0.005692768096923828, 0.005517312049865723, 0.005444159984588623, 0.005382239818572998, 0.005439680099487304, 0.005379903793334961, 0.005441535949707031, 0.00535756778717041, 0.0053043198585510255, 0.005249343872070313, 0.005324480056762696, 0.005326848030090332, 0.005260831832885742, 0.005285855770111084, 0.0053005762100219725, 0.005301792144775391, 0.0053233919143676756, 0.005316256046295166, 0.005325151920318604, 0.005361120223999024, 0.005343776226043701, 0.00531660795211792, 0.005390336036682129, 0.0053303041458129885, 0.005312416076660156, 0.005349952220916748, 0.005314720153808594, 0.0052882561683654785, 0.005323455810546875, 0.005397791862487793, 0.005401951789855957, 0.005472640037536621, 0.0054906878471374515, 0.0054026880264282226, 0.005470143795013428, 0.005480447769165039, 0.005502912044525146, 0.005486656188964844, 0.005431295871734619, 0.005498591899871826, 0.005445919990539551, 0.005455872058868408, 0.00550707197189331, 0.005435391902923584, 0.005406720161437988, 0.00547430419921875, 0.005406720161437988, 0.005402624130249023, 0.005361663818359375, 0.005341055870056152, 0.005386367797851562, 0.005427135944366455, 0.005393727779388428, 0.005350272178649903, 0.005356927871704101, 0.005390719890594482, 0.005415487766265869, 0.0054002561569213865, 0.005360960006713867, 0.005375967979431152, 0.0053769278526306155, 0.005432479858398438, 0.005380000114440918, 0.005372255802154541, 0.005395199775695801, 0.005422880172729492, 0.005473440170288086, 0.005455904006958008, 0.005620704174041748, 0.005562079906463623, 0.005489920139312744, 0.005598239898681641, 0.005474016189575195, 0.0055025601387023925, 0.00544323205947876, 0.005473023891448975, 0.005435455799102784, 0.0053860158920288085, 0.005367968082427978, 0.005439487934112549, 0.005414912223815918, 0.00548367977142334, 0.005475168228149414, 0.005445248126983643, 0.005450111865997314, 0.005427103996276855, 0.005404767990112305, 0.005389632225036621, 0.005415616035461426, 0.005609471797943116, 0.005434368133544922, 0.0056514558792114256, 0.005383776187896728, 0.005429024219512939, 0.005763711929321289, 0.0054778242111206054, 0.005464191913604736, 0.005461984157562256, 0.005491168022155762, 0.005408192157745361, 0.00541270399093628, 0.005464320182800293, 0.005370336055755615, 0.005406720161437988, 0.005335103988647461, 0.005354656219482422, 0.005753632068634033, 0.005752831935882568, 0.006037312030792236, 0.007058879852294922, 0.009331456184387207, 0.008516736030578613, 0.005598176002502442, 0.005582752227783203, 0.005525504112243652, 0.005551616191864014, 0.005626495838165283, 0.005443327903747559, 0.005564576148986817, 0.0053983678817749025, 0.005375455856323242, 0.0057489280700683595, 0.005420608043670654, 0.005320576190948487, 0.005303040027618409, 0.005312672138214111, 0.00530841588973999, 0.005240960121154785, 0.005344927787780762, 0.005287231922149659, 0.005266335964202881, 0.005260287761688232, 0.005274335861206055, 0.005275936126708985, 0.005293087959289551, 0.005303264141082764, 0.005271455764770508, 0.005290080070495606, 0.005285183906555176, 0.005333951950073242, 0.005318048000335693, 0.0053005762100219725, 0.0053043198585510255, 0.005414912223815918, 0.005436768054962158, 0.005417920112609863, 0.005480095863342285, 0.005465568065643311, 0.005448224067687989, 0.005516416072845459, 0.005434304237365723, 0.005416319847106933, 0.005422912120819092, 0.0054815678596496584, 0.005578464031219483, 0.005566271781921386, 0.005486783981323243, 0.0055101442337036136, 0.005442240238189697, 0.0055248641967773435, 0.005387199878692627, 0.005279583930969238, 0.005275807857513428, 0.005268703937530517, 0.005286687850952148, 0.005302271842956543, 0.005285888195037842, 0.00529363203048706, 0.005280191898345947, 0.005373888015747071, 0.005297311782836914, 0.005303199768066406, 0.0053463678359985355, 0.005319007873535156, 0.005305056095123291, 0.005348991870880127, 0.005327104091644287, 0.005444799900054932, 0.005409599781036377, 0.005434688091278076, 0.005315072059631347, 0.005383935928344727, 0.005375999927520752, 0.005379839897155762, 0.005396736145019531, 0.0054102401733398435, 0.005357600212097168, 0.005339680194854736, 0.0053554878234863285, 0.005281599998474121, 0.005289247989654541, 0.005313055992126465, 0.005234623908996582, 0.005269343852996826, 0.005289663791656494, 0.005270624160766601, 0.005385280132293701, 0.005322559833526612, 0.005308608055114746, 0.005356224060058594, 0.0052962880134582516, 0.00525929594039917, 0.005311520099639893, 0.0053043198585510255, 0.0052822079658508304, 0.005320703983306885, 0.00532528018951416, 0.00530841588973999, 0.005351424217224121, 0.00544371223449707, 0.005355391979217529, 0.005345280170440674, 0.005320703983306885, 0.005267327785491944, 0.005281919956207275, 0.005271423816680909, 0.005242112159729004, 0.005261695861816406, 0.005200160026550293, 0.005310400009155274, 0.005183104038238525, 0.0052202239036560055, 0.005187967777252197, 0.005177760124206543, 0.005228544235229492, 0.005216288089752197, 0.005204224109649659, 0.005211872100830078, 0.005210112094879151, 0.0051998720169067385, 0.005251071929931641, 0.005175295829772949, 0.005189824104309082, 0.005171264171600342, 0.0051855998039245605, 0.005168223857879638, 0.005196352005004883, 0.005218592166900635, 0.005216000080108643, 0.005180448055267334, 0.005185535907745361, 0.0052557439804077144, 0.005198239803314209, 0.005111584186553955, 0.005177760124206543, 0.005200352191925049, 0.005244383811950684, 0.005208608150482178, 0.005232639789581299, 0.005227712154388428, 0.005281983852386474, 0.005276288032531738, 0.005185440063476563, 0.005172607898712159, 0.005184127807617188, 0.0052399678230285645, 0.005243840217590332, 0.005162112236022949, 0.005188479900360108, 0.005187583923339844, 0.005361663818359375, 0.005203968048095703, 0.005197824001312256, 0.005179391860961914, 0.005164608001708985, 0.005189087867736817, 0.005172192096710205, 0.005151872158050537, 0.00520851182937622, 0.005185823917388916, 0.005183231830596924, 0.005290112018585205, 0.005191967964172363, 0.005201632022857666, 0.005161119937896729, 0.005203904151916504, 0.005228640079498291, 0.005177440166473389, 0.005197887897491455, 0.005173183917999267, 0.005799935817718506, 0.005982207775115967, 0.006150335788726807, 0.006502367973327637, 0.005484384059906006, 0.005242879867553711, 0.005216576099395752, 0.005207744121551514, 0.005251071929931641, 0.005230591773986816, 0.005195775985717774, 0.00532419204711914, 0.005204544067382812, 0.005199903964996338, 0.005191296100616455, 0.005175680160522461, 0.0051662721633911135, 0.005181344032287598, 0.005210400104522705, 0.005204288005828858, 0.0051708159446716305, 0.005189663887023925, 0.00524560022354126, 0.005178880214691162, 0.005175583839416504, 0.00520851182937622, 0.005102880001068115, 0.005405792236328125, 0.005179168224334717, 0.005234848022460938, 0.005216095924377441, 0.00530841588973999, 0.005314720153808594, 0.005256192207336426, 0.005194591999053955, 0.005201920032501221, 0.005208064079284668, 0.005172863960266113, 0.005157248020172119, 0.0052013759613037105, 0.005180255889892578, 0.005170048236846924, 0.005271647930145264, 0.005184224128723144, 0.005164480209350586, 0.005177919864654541, 0.005305984020233154, 0.005155456066131592, 0.005177408218383789, 0.005189151763916016, 0.005183648109436035, 0.005332992076873779, 0.005547679901123047, 0.005200128078460694, 0.005174528121948243, 0.0052007360458374025, 0.005183487892150879, 0.005171199798583984, 0.005236800193786621, 0.005186592102050781, 0.005211040019989014, 0.005199967861175537, 0.005204063892364502, 0.005228352069854736, 0.005172416210174561, 0.005194623947143554, 0.0051916160583496095, 0.0051643519401550295, 0.005229248046875, 0.005248672008514405, 0.00522876787185669, 0.005238912105560303, 0.005233920097351074, 0.005288703918457031, 0.005236639976501465, 0.005275743961334229, 0.00525216007232666, 0.005259647846221924, 0.0052740478515625, 0.005265503883361816, 0.005292031764984131, 0.005298111915588379, 0.005255263805389404, 0.005237055778503418, 0.005325535774230957, 0.00527023983001709, 0.005263040065765381, 0.005399104118347168, 0.00526470422744751, 0.005214943885803223, 0.005255712032318115, 0.00525267219543457, 0.005304351806640625, 0.005279104232788086, 0.005278463840484619, 0.005297696113586426, 0.005301760196685791, 0.005337984085083008, 0.005298175811767578, 0.005242335796356201, 0.005259935855865479, 0.005253087997436523, 0.005346528053283691, 0.005372704029083252, 0.005326848030090332, 0.0053877439498901366, 0.005470367908477783, 0.005425631999969482, 0.00540012788772583, 0.005384543895721436, 0.005385695934295654, 0.005372447967529297, 0.005398528099060058, 0.005433343887329102, 0.005334752082824707, 0.005318943977355957, 0.005276703834533691, 0.005253952026367187, 0.005355679988861084, 0.005294047832489014, 0.005270976066589356, 0.005404672145843506, 0.0054514241218566895, 0.005579103946685791, 0.005494912147521972, 0.005501279830932617, 0.005471648216247559, 0.005396927833557129, 0.00547215986251831, 0.00549721622467041, 0.005449728012084961, 0.005486591815948487, 0.0055459198951721194, 0.005613632202148437, 0.00555017614364624, 0.005619616031646729, 0.005561952114105225, 0.005601920127868652, 0.005574175834655762, 0.005646592140197754, 0.005615615844726562, 0.005530784130096436, 0.005511936187744141, 0.005531904220581055, 0.005412992000579834, 0.005432864189147949, 0.005348544120788574, 0.005346303939819336, 0.005339231967926026, 0.0053779520988464355, 0.005401631832122803, 0.005489727973937988, 0.005679232120513916, 0.005488895893096924, 0.005836415767669678, 0.005541888236999512, 0.005681151866912842, 0.005550079822540284, 0.0055763840675354, 0.005570879936218262, 0.005522463798522949, 0.00553059196472168, 0.005590240001678467, 0.005464863777160644, 0.005476352214813233, 0.005457376003265381, 0.005461599826812744, 0.005333951950073242, 0.005334752082824707, 0.005365280151367188, 0.0053294401168823245, 0.005314527988433838, 0.005378111839294434, 0.00532089614868164, 0.005395967960357666, 0.005788127899169922, 0.005662144184112549, 0.005411424160003662, 0.005453311920166016, 0.005820223808288574, 0.005472576141357422, 0.005449408054351806, 0.005446335792541504, 0.005392384052276611, 0.005342336177825928, 0.005299071788787842, 0.005344480037689209, 0.005309216022491455, 0.005297952175140381, 0.005324575901031494, 0.005277599811553955, 0.005244639873504639, 0.005325215816497803, 0.00525324821472168, 0.005212448120117188, 0.005265408039093018, 0.005206016063690186, 0.0052276158332824705, 0.005229472160339355, 0.005252352237701416, 0.005249792098999023, 0.0052512001991271975, 0.005242239952087402, 0.005273920059204101, 0.005204127788543701, 0.005238560199737549, 0.005204351902008056, 0.005219232082366943, 0.005233632087707519, 0.005257376194000244, 0.005254687786102295, 0.005319007873535156, 0.005352640151977539, 0.005378528118133545, 0.005343552112579346, 0.005316544055938721, 0.0052954239845275876, 0.005274144172668457, 0.005319968223571777, 0.005356480121612548, 0.005315936088562011, 0.0052919998168945315, 0.005328800201416016, 0.005267231941223144, 0.00528495979309082, 0.005221471786499023, 0.0052358078956604, 0.005218016147613525, 0.005232736110687256, 0.005218272209167481, 0.005388224124908447, 0.005281792163848877, 0.005322656154632568, 0.005331039905548096, 0.005299200057983398, 0.005264383792877197, 0.005273600101470947, 0.005597184181213379, 0.005337088108062744, 0.005330560207366944, 0.005330624103546143, 0.005333695888519287, 0.005338912010192871, 0.005394815921783447, 0.005356863975524902, 0.005313055992126465, 0.005338560104370118, 0.005316671848297119, 0.005267744064331055, 0.005486976146697998, 0.005377888202667236, 0.005251071929931641, 0.005242976188659668, 0.00526643180847168, 0.005208992004394532, 0.0052733120918273925, 0.005224575996398926, 0.005506752014160156, 0.005480447769165039, 0.005286464214324951, 0.00527350378036499, 0.0052674560546875, 0.005281375885009766, 0.005347199916839599, 0.005392159938812256, 0.005399295806884766, 0.005440832138061524, 0.005452479839324951, 0.005523551940917968, 0.005597087860107422, 0.00569488000869751, 0.005451680183410645, 0.005475135803222656, 0.005433311939239502, 0.005459167957305908, 0.005495488166809082, 0.005416607856750489, 0.00539302396774292, 0.00540067195892334, 0.005404064178466797, 0.005378367900848388, 0.005433631896972656, 0.005492735862731934, 0.005302271842956543, 0.005337088108062744, 0.0055328960418701174, 0.005310624122619629, 0.005372384071350098, 0.0053446397781372074, 0.005325535774230957, 0.0053209600448608394, 0.0053348479270935055, 0.005264736175537109, 0.0053985600471496585, 0.005286848068237305, 0.0052936959266662595, 0.005296192169189453, 0.0052633600234985355, 0.005255231857299805, 0.005315743923187256, 0.00534607982635498, 0.005312511920928955, 0.0053249278068542485, 0.005447423934936523, 0.005414463996887207, 0.005356095790863037, 0.005536863803863526, 0.005359839916229248, 0.005411520004272461, 0.007044544219970703, 0.006045375823974609, 0.006243552207946777, 0.005506720066070557, 0.005565760135650635, 0.005483200073242187, 0.005687295913696289, 0.005615327835083008, 0.005410783767700196, 0.005356895923614502, 0.005379263877868652, 0.005340320110321045, 0.005326848030090332, 0.005400991916656494, 0.0053786239624023435, 0.005340479850769043, 0.005340544223785401, 0.005298592090606689, 0.005321311950683594, 0.005316703796386719, 0.005385568141937256, 0.005330719947814941, 0.005363935947418213, 0.005343776226043701, 0.005421055793762207, 0.005386240005493164, 0.005398431777954102, 0.005390272140502929, 0.005398687839508057, 0.005383456230163574, 0.00537395191192627, 0.005353631973266601, 0.005222591876983642, 0.005300127983093262, 0.005294239997863769, 0.005326496124267578, 0.005294496059417724, 0.00532480001449585, 0.0052856640815734865, 0.005420639991760254, 0.005282432079315186, 0.005279744148254394, 0.005258399963378907, 0.005382719993591308, 0.005261023998260498, 0.005284416198730469, 0.0052871999740600585, 0.005261504173278809, 0.005290527820587158, 0.005971968173980713, 0.005605375766754151, 0.00531609582901001, 0.005312160015106201, 0.005351871967315674, 0.005310751914978027, 0.005350815773010254, 0.005343071937561035, 0.005304255962371826, 0.005302303791046142, 0.005241759777069092, 0.005227935791015625, 0.005231200218200683, 0.0052674560546875, 0.005249023914337158, 0.005339136123657226, 0.0052667841911315915, 0.005233312129974365, 0.005236959934234619, 0.005393856048583985, 0.005333439826965332, 0.005273375988006592, 0.005238272190093994, 0.005238687992095948, 0.005198560237884522, 0.005306367874145508, 0.005260608196258545, 0.005194719791412354, 0.005246975898742676, 0.005543168067932129, 0.005202112197875977, 0.00521779203414917, 0.005196576118469238, 0.005232639789581299, 0.005222400188446045, 0.0052165122032165525, 0.005226240158081054, 0.005193535804748535, 0.005253151893615722, 0.005271743774414062, 0.0052193598747253414, 0.005213024139404297, 0.0051818561553955075, 0.005182591915130616, 0.0051851201057434085, 0.005266335964202881, 0.00510211181640625, 0.005234367847442627, 0.005218560218811035, 0.005222464084625244, 0.005232639789581299, 0.005163008213043213, 0.005207295894622803, 0.005207071781158448, 0.005199903964996338, 0.005275328159332276, 0.005241087913513184, 0.005211904048919678, 0.005255167961120606, 0.005220352172851562, 0.00518943977355957, 0.005215519905090332, 0.0051957440376281736, 0.005208255767822266, 0.005218688011169433, 0.005216447830200195, 0.005265215873718262, 0.00520252799987793, 0.005221568107604981, 0.005191904067993164, 0.005185919761657715, 0.0052197761535644534, 0.005140543937683106, 0.005210368156433105, 0.005170623779296875, 0.005240960121154785, 0.005239168167114258, 0.005197696208953858, 0.005196320056915283, 0.005221759796142578, 0.0051617598533630375, 0.0053431358337402346, 0.00518943977355957, 0.005226304054260254, 0.00522057580947876, 0.0052460160255432126, 0.005237696170806885, 0.005246528148651123, 0.00520851182937622, 0.005189375877380371, 0.005236192226409912, 0.005230751991271973, 0.005210752010345459, 0.005202015876770019, 0.005210015773773193, 0.0053678078651428224, 0.0052750401496887206, 0.005323359966278076, 0.00520524787902832, 0.005224256038665771, 0.005187744140625, 0.005528416156768799, 0.00524012804031372, 0.005212160110473632, 0.005208352088928223, 0.005176832199096679, 0.0051801280975341795, 0.005195903778076172, 0.005269375801086426, 0.005171135902404785, 0.005211967945098877, 0.005205728054046631, 0.005184000015258789, 0.005214240074157715, 0.005258431911468506, 0.005251904010772705, 0.005274623870849609, 0.005333504199981689, 0.0053318080902099605, 0.005237760066986084, 0.005211071968078613, 0.005295519828796387, 0.005238751888275147, 0.005222239971160889, 0.005251455783843994, 0.005211455821990967, 0.005174079895019531, 0.005193727970123291, 0.005221439838409424, 0.005209023952484131, 0.005176959991455078, 0.005216224193572998, 0.0051940159797668455, 0.005147840023040771, 0.005219264030456543, 0.005212096214294434, 0.0051918082237243654, 0.005433279991149903, 0.00543552017211914, 0.005270815849304199, 0.005206111907958984, 0.005184000015258789, 0.005173247814178467, 0.005151840209960937, 0.0053441600799560545, 0.005754432201385498, 0.005509568214416504, 0.005195199966430664, 0.005232831954956055, 0.005200352191925049, 0.005138400077819824, 0.0051916160583496095, 0.005144864082336426, 0.005201663970947266, 0.005145567893981934, 0.0051404800415039064, 0.005188896179199219, 0.0051463360786437986, 0.005142528057098389, 0.00517958402633667, 0.005148384094238281, 0.005155104160308838, 0.005152575969696045, 0.00524294376373291, 0.006780159950256348, 0.005196127891540527, 0.005148191928863525, 0.005163839817047119, 0.005189504146575928, 0.005199327945709229, 0.005145120143890381, 0.005130368232727051, 0.005043488025665283, 0.005290016174316407, 0.005169888019561767, 0.0051773438453674315, 0.005138432025909424, 0.005167520046234131, 0.005174880027770996, 0.0051948800086975095, 0.005177599906921387, 0.005151711940765381, 0.0051580162048339845, 0.005216800212860108, 0.005157983779907226, 0.005172128200531006, 0.005127808094024658, 0.0054254398345947264, 0.005259359836578369, 0.005240447998046875, 0.005615615844726562, 0.005218688011169433, 0.005197824001312256, 0.005211872100830078, 0.005353759765625, 0.005177440166473389, 0.005203360080718994, 0.005200255870819092, 0.00519385576248169, 0.005234687805175781, 0.00525494384765625, 0.005201759815216064, 0.005192160129547119, 0.005618815898895264, 0.0052091522216796875, 0.005164608001708985, 0.00520579195022583, 0.005298431873321533, 0.005203167915344239, 0.0052130880355834965, 0.005243135929107666, 0.00521120023727417, 0.005179168224334717, 0.005161439895629883, 0.005165503978729248, 0.005158815860748291, 0.005208159923553467, 0.0051847038269042965, 0.005151552200317383, 0.00522815990447998, 0.005171584129333496, 0.0052204480171203615, 0.005190688133239746, 0.00514246416091919, 0.0051550078392028804, 0.005165823936462402, 0.005249023914337158, 0.005250048160552978, 0.005157792091369629, 0.005171296119689942, 0.005172512054443359, 0.005199999809265137, 0.00517526388168335, 0.005175648212432861, 0.005191008090972901, 0.005106175899505615, 0.005195424079895019, 0.0051938238143920895, 0.00518006420135498, 0.005141952037811279, 0.005183872222900391, 0.005205567836761475, 0.00516761589050293, 0.005149856090545655, 0.005179520130157471, 0.005235424041748047, 0.005197824001312256, 0.0051979517936706545, 0.005170464038848877, 0.00517958402633667, 0.005202335834503174, 0.005219552040100097, 0.005173471927642823, 0.00516764783859253, 0.0052135357856750485, 0.005170080184936523, 0.00521295976638794, 0.00521014404296875, 0.005196512222290039, 0.0051814718246459964, 0.0052081279754638676, 0.005202112197875977, 0.00524835205078125, 0.005216063976287842, 0.005294911861419678, 0.005214303970336914, 0.0051794562339782715, 0.005226143836975098, 0.0052308797836303714, 0.005240543842315674, 0.005239295959472656, 0.005232639789581299, 0.005240511894226074, 0.005210112094879151, 0.005187583923339844, 0.005187583923339844, 0.005153823852539063, 0.005229631900787354, 0.005210015773773193, 0.00523414421081543, 0.00521017599105835, 0.005220831871032715, 0.005273695945739746, 0.005363615989685058, 0.005218560218811035, 0.0051996798515319825, 0.005197760105133057, 0.005255167961120606, 0.005197824001312256, 0.0055214080810546875, 0.0052408318519592285, 0.005230175971984863, 0.005196191787719726, 0.005283840179443359, 0.005203199863433838, 0.005190624237060547, 0.005195424079895019, 0.0051981439590454105, 0.005091360092163086, 0.005216576099395752, 0.005195775985717774, 0.005285312175750732, 0.005245503902435303, 0.005159967899322509, 0.005219071865081787, 0.00522876787185669, 0.005188896179199219, 0.005245664119720459, 0.00522979211807251, 0.005210847854614258, 0.005276991844177246, 0.0052492480278015135, 0.005240447998046875, 0.00538108777999878, 0.005277632236480713, 0.005268703937530517, 0.005260064125061035, 0.0052063679695129395, 0.005218175888061524, 0.005187136173248291, 0.005334303855895996, 0.005234943866729736, 0.0052005119323730465, 0.0052501120567321775, 0.005204991817474365, 0.005201536178588867, 0.00523635196685791, 0.00522873592376709, 0.005233215808868408, 0.005191679954528809, 0.005234464168548584, 0.005343008041381836, 0.005243328094482422, 0.005244927883148193, 0.005283199787139893, 0.005221216201782226, 0.005254079818725586, 0.005230463981628418, 0.005199935913085938, 0.005425119876861572, 0.005292992115020752, 0.0052715520858764645, 0.005326848030090332, 0.005277696132659912, 0.005341055870056152, 0.005300543785095215, 0.005302303791046142, 0.005317503929138183, 0.005348256111145019, 0.0053002238273620605, 0.005274687767028809, 0.005297088146209717, 0.0053053760528564455, 0.0053207998275756835, 0.005373087882995606, 0.005365664005279541, 0.005278624057769775, 0.0053105602264404295, 0.005423647880554199, 0.005273888111114502, 0.005319808006286621, 0.005444767951965332, 0.005309279918670654, 0.005326848030090332, 0.005414112091064453, 0.005270304203033448, 0.005295711994171143, 0.005301663875579834, 0.005284543991088867, 0.005261760234832763, 0.005265279769897461, 0.005320703983306885, 0.005340735912322998, 0.0053294081687927245, 0.005729407787322998, 0.005329343795776367, 0.005333504199981689, 0.0052386560440063476, 0.005269504070281982, 0.0052911038398742675, 0.005346240043640137, 0.005351391792297364, 0.005326752185821533, 0.0052466878890991215, 0.0052739839553833006, 0.005261312007904053, 0.005281087875366211, 0.005339263916015625, 0.005326848030090332, 0.0053417601585388184, 0.005304543972015381, 0.005320159912109375, 0.005333312034606933, 0.005270624160766601, 0.005374879837036133, 0.005339488029479981, 0.005266304016113281, 0.005294879913330078, 0.005305856227874756, 0.005300735950469971, 0.0053309440612792965, 0.005299680233001709, 0.005291903972625732, 0.005307360172271729, 0.005308256149291992, 0.005352352142333984, 0.005292223930358887, 0.005286816120147705, 0.005287903785705567, 0.005280928134918213, 0.005304895877838135, 0.005308671951293946, 0.005257120132446289, 0.005283840179443359, 0.005256608009338379, 0.005291872024536133, 0.005247680187225342, 0.0052583680152893065, 0.005225215911865234, 0.005281023979187012, 0.005272160053253174, 0.0052778878211975095, 0.005303840160369873, 0.005327104091644287, 0.0053143038749694825, 0.005296383857727051, 0.005285888195037842, 0.005312511920928955, 0.0052432317733764645, 0.005255904197692871, 0.0052848000526428225, 0.005246975898742676, 0.005255167961120606, 0.005273759841918945, 0.0052449598312377926, 0.005322559833526612, 0.005467807769775391, 0.005499231815338135, 0.005306367874145508, 0.005530816078186035, 0.005313344001770019, 0.0053506879806518555, 0.005361728191375733, 0.006898367881774902, 0.006851967811584473, 0.00540118408203125, 0.005359615802764893, 0.005310463905334473, 0.005351424217224121, 0.005328896045684814, 0.005296127796173096, 0.005311488151550293, 0.005299200057983398, 0.0053034558296203614, 0.005331424236297608, 0.005622399806976318, 0.005400576114654541, 0.005324831962585449, 0.005316319942474365, 0.00524505615234375, 0.005263232231140137, 0.005199391841888428, 0.0052219839096069335, 0.005224703788757324, 0.005206655979156494, 0.0052367358207702636, 0.005255167961120606, 0.005193024158477783, 0.005298143863677979, 0.005238880157470703, 0.0051943678855896, 0.005190720081329345, 0.005190080165863037, 0.005247488021850586, 0.005484799861907959, 0.005261280059814453, 0.005257279872894287, 0.005224160194396973, 0.0052080960273742675, 0.005247039794921875, 0.00523414421081543, 0.005248703956604004, 0.005264287948608399, 0.005252480030059814, 0.005241312026977539, 0.005298175811767578, 0.005275296211242676, 0.00514083194732666, 0.005173376083374023, 0.005189023971557617, 0.0052213120460510255, 0.0052015681266784665, 0.005188960075378418, 0.005208735942840576, 0.005203680038452148, 0.00526364803314209, 0.005210112094879151, 0.005203968048095703, 0.005184607982635498, 0.005200640201568603, 0.005175487995147705, 0.005189087867736817, 0.005181280136108399, 0.005222527980804443, 0.005191679954528809, 0.005186079978942871, 0.005242976188659668, 0.0053628478050231935, 0.005181375980377197, 0.005267936229705811, 0.0052288641929626465, 0.005191711902618408, 0.005214208126068115, 0.005220352172851562, 0.005222367763519287, 0.005206048011779785, 0.00531660795211792, 0.005226047992706298, 0.005226880073547363, 0.00570579195022583, 0.005610496044158936, 0.005876832008361817, 0.005631904125213623, 0.005756927967071533, 0.0053229122161865235, 0.005256063938140869, 0.005311200141906738, 0.0052492799758911135, 0.005229983806610107, 0.005252960205078125, 0.0052839360237121585, 0.0052432317733764645, 0.005194399833679199, 0.005217023849487305, 0.005196063995361328, 0.005169792175292969, 0.0052367358207702636, 0.0052977919578552244, 0.005202303886413575, 0.00518723201751709, 0.005310207843780518, 0.005212768077850342, 0.005187583923339844, 0.0053002238273620605, 0.005209343910217285, 0.005167871952056885, 0.005172800064086914, 0.005230591773986816, 0.005150847911834717, 0.0051511359214782716, 0.005136832237243653, 0.005199776172637939, 0.005164735794067383, 0.005210527896881104, 0.005156864166259765, 0.0051673917770385745, 0.00517091178894043, 0.005242879867553711, 0.005219552040100097, 0.005354400157928467, 0.005214208126068115, 0.005267327785491944, 0.005525599956512451, 0.005214240074157715, 0.005222271919250488, 0.005189631938934326, 0.005213791847229004, 0.005327263832092285, 0.005230336189270019, 0.005252927780151367, 0.005222879886627198, 0.005193120002746582, 0.005203936100006103, 0.005196159839630127, 0.005222527980804443, 0.005187679767608642, 0.005173247814178467, 0.005259263992309571, 0.005186783790588379, 0.005180448055267334, 0.005189087867736817, 0.005257567882537842, 0.005162943840026855, 0.0051651840209960935, 0.005185408115386963, 0.005195775985717774, 0.005154911994934082, 0.005197728157043457, 0.005221536159515381, 0.005177760124206543, 0.00519596815109253, 0.005189888000488281, 0.005179327964782715, 0.005205344200134277, 0.005179967880249024, 0.006047904014587402, 0.005228544235229492, 0.005185535907745361, 0.005238783836364746, 0.005244927883148193, 0.0052120318412780765, 0.005171167850494385, 0.0051914558410644535, 0.005202400207519531, 0.005183328151702881, 0.005173312187194824, 0.0052074241638183595, 0.005161600112915039, 0.005163167953491211, 0.005228384017944336, 0.005174399852752686, 0.005237631797790528, 0.005159135818481445]",tokens/s,187.00617262132317,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Deci/DeciLM-7B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Deci/DeciLM-7B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3143.491584,4639.883264,0.0,4244.635648,4125.520384,s,1,11.9792587890625,11.9792587890625,0.0,11.9792587890625,11.9792587890625,11.9792587890625,11.9792587890625,[11.9792587890625],,kWh,0.00014281466581666488,1.574390353824077e-05,5.3873098654011464e-05,0.00021243166800891713,,MB,3187.0976,4813.94688,0.0,4404.0192,4310.797312,s,10,1.1434629516601562,0.11434629516601562,0.00016764718123132437,0.11430782318115235,0.1145461784362793,0.11459340934753418,0.11463119407653809,"[0.11464064025878906, 0.1145203857421875, 0.11453568267822266, 0.11433599853515625, 0.11418809509277343, 0.11411901092529297, 0.11417081451416015, 0.1144258270263672, 0.11427964782714843, 0.1142468490600586]",tokens/s,2238.8132438250145,kWh,3.4241915051362132e-06,3.7762772297003015e-07,2.2542459894419778e-06,6.056065217548221e-06,tokens/kWh,42271671.58936918,MB,3191.300096,4816.044032,0.0,4406.116352,4310.799872,s,10,22.40705126953125,2.2407051269531246,0.01225969367185595,2.2395146484375,2.2519275390624998,2.259535791015625,2.265622392578125,"[2.243108154296875, 2.26714404296875, 2.25023681640625, 2.238847412109375, 2.239515625, 2.22377685546875, 2.222224609375, 2.235335205078125, 2.247348876953125, 2.239513671875]",tokens/s,28.116149350569124,kWh,6.422840164111306e-05,7.084238282746035e-06,3.3550254876357245e-05,0.00010486289480021634,tokens/kWh,600784.4826335085,,s,630,22.40389617156981,0.03556173995487274,0.0011464844853278735,0.03544276809692383,0.03588814811706543,0.036251442718505854,0.03793366279602052,"[0.03651932907104492, 0.035850849151611325, 0.03918447875976563, 0.035675838470458986, 0.03560620880126953, 0.035442623138427734, 0.03555145645141602, 0.03538927841186523, 0.035374942779541015, 0.03515667343139649, 0.03546316909790039, 0.03536844635009766, 0.0354411506652832, 0.035438591003417966, 0.03575603103637695, 0.03582361602783203, 0.03530697631835938, 0.03554483032226562, 0.035240062713623045, 0.03538332748413086, 0.035312255859375, 0.03546665573120117, 0.036733535766601565, 0.035608417510986326, 0.0355101432800293, 0.03543888092041016, 0.03530255889892578, 0.035668830871582034, 0.035518081665039065, 0.03567177581787109, 0.035582622528076174, 0.035483230590820314, 0.03551478576660156, 0.0357498893737793, 0.035422206878662106, 0.03555737686157227, 0.03540172958374024, 0.035456321716308595, 0.03542086410522461, 0.0353950080871582, 0.035574337005615235, 0.03539046478271484, 0.035714046478271484, 0.035464256286621094, 0.035703136444091794, 0.035412574768066404, 0.03551027297973633, 0.03536281585693359, 0.035434337615966795, 0.0354428482055664, 0.03650547027587891, 0.03544076919555664, 0.03534793472290039, 0.035374847412109375, 0.035526592254638674, 0.0354947509765625, 0.03556556701660156, 0.0358109130859375, 0.03537097549438477, 0.03529561614990234, 0.035538593292236326, 0.035601024627685544, 0.03519382476806641, 0.036413345336914066, 0.035544734954833984, 0.03553580856323242, 0.03579904174804688, 0.03585152053833008, 0.0370445442199707, 0.03593875122070313, 0.06162636947631836, 0.03573964691162109, 0.03575356674194336, 0.035598751068115234, 0.03659932708740234, 0.03576675033569336, 0.035950592041015625, 0.03573168182373047, 0.03546495819091797, 0.03563932800292969, 0.03566284942626953, 0.03549622344970703, 0.035524703979492187, 0.03518323135375977, 0.0350904312133789, 0.03521094512939453, 0.03521955108642578, 0.035319393157958984, 0.03522732925415039, 0.03551728057861328, 0.03550585556030274, 0.03547177505493164, 0.03544678497314453, 0.03559219360351563, 0.03560009765625, 0.036042400360107425, 0.03571913528442383, 0.035541664123535155, 0.03555123138427734, 0.03566796875, 0.03534364700317383, 0.035438591003417966, 0.035437217712402345, 0.03563248062133789, 0.035477310180664065, 0.03521628952026367, 0.03530752182006836, 0.035244033813476565, 0.03552179336547852, 0.03573763275146485, 0.03537753677368164, 0.03522524642944336, 0.03530207824707031, 0.03523993682861328, 0.03533580780029297, 0.03557756805419922, 0.03546588897705078, 0.035342464447021486, 0.035523681640625, 0.035377857208251956, 0.03525846481323242, 0.03538473510742188, 0.03571363067626953, 0.035823646545410155, 0.03546646499633789, 0.03546092987060547, 0.03595775985717774, 0.03531465530395508, 0.03604006576538086, 0.035307296752929686, 0.03528793716430664, 0.03610009765625, 0.035748992919921875, 0.035326847076416014, 0.03514518356323242, 0.035653537750244144, 0.03534214401245117, 0.03565241622924804, 0.03538489532470703, 0.03544313430786133, 0.035901439666748046, 0.035383201599121096, 0.035170398712158206, 0.0355491828918457, 0.035068958282470704, 0.0352239990234375, 0.0353408317565918, 0.035491199493408204, 0.03538188934326172, 0.03558720016479492, 0.03543081665039063, 0.035546878814697265, 0.03537583923339844, 0.0356495361328125, 0.03543654251098633, 0.03556966400146484, 0.036420894622802735, 0.03589177703857422, 0.0354439697265625, 0.03619318389892578, 0.03560857772827149, 0.03585539245605469, 0.03826937484741211, 0.03554668807983399, 0.03557993698120117, 0.035875232696533206, 0.035797534942626955, 0.03577622222900391, 0.03544707107543945, 0.03635577774047852, 0.03550396728515625, 0.03568483352661133, 0.03552870559692383, 0.035722496032714844, 0.035385982513427734, 0.0357367057800293, 0.03563849639892578, 0.03581219100952148, 0.03577148818969727, 0.03592585754394531, 0.03551846313476562, 0.03577036666870117, 0.03800668716430664, 0.03598140716552734, 0.03585433578491211, 0.03565363311767578, 0.03575603103637695, 0.03612992095947266, 0.0356545295715332, 0.0364101448059082, 0.03558732986450195, 0.03541823959350586, 0.035396095275878905, 0.03551859283447266, 0.03548364639282227, 0.035366912841796876, 0.03626803207397461, 0.0356495361328125, 0.03564271926879883, 0.03602908706665039, 0.03562905502319336, 0.035514110565185546, 0.03597558212280273, 0.03613679885864258, 0.035833854675292966, 0.03529523086547852, 0.03544598388671875, 0.03530575942993164, 0.0354411506652832, 0.03565283203125, 0.03560323333740235, 0.035520511627197264, 0.0356126708984375, 0.03550822448730469, 0.03564892959594727, 0.03701148986816406, 0.03565011215209961, 0.03531161499023437, 0.03548303985595703, 0.035839935302734376, 0.03554780960083008, 0.03520105743408203, 0.03684473419189453, 0.035226432800292966, 0.03539673614501953, 0.035133472442626955, 0.03514860916137695, 0.03509785461425781, 0.03518940734863281, 0.03531897735595703, 0.035146656036376955, 0.03517440032958984, 0.035166206359863283, 0.03522528076171875, 0.03493510437011719, 0.03486105728149414, 0.03511852645874024, 0.03526076889038086, 0.035425567626953126, 0.035414974212646486, 0.03570278549194336, 0.035381248474121094, 0.03553484725952148, 0.035418113708496096, 0.03539785766601562, 0.03545577621459961, 0.03575081634521484, 0.03560198211669922, 0.03585612869262695, 0.035523006439208984, 0.03547580718994141, 0.03540563201904297, 0.03639014434814453, 0.03567030334472656, 0.03817929458618164, 0.035677375793457033, 0.03548275375366211, 0.03568576049804688, 0.03557414245605469, 0.035403713226318356, 0.03523993682861328, 0.035999744415283204, 0.03643116760253906, 0.03557369613647461, 0.03532262420654297, 0.03544063949584961, 0.03522355270385742, 0.03539052963256836, 0.036979904174804686, 0.035332927703857424, 0.03526067352294922, 0.03523244857788086, 0.03530259323120117, 0.035251007080078126, 0.03525836944580078, 0.03536281585693359, 0.03549184036254883, 0.03527884674072266, 0.035168254852294925, 0.03553484725952148, 0.03559004974365235, 0.03545471954345703, 0.0356130256652832, 0.035438591003417966, 0.035618976593017576, 0.035636577606201175, 0.035291648864746096, 0.035415775299072264, 0.035364768981933595, 0.03535305786132813, 0.0354299201965332, 0.03542240142822266, 0.03553507232666016, 0.035544544219970706, 0.035660289764404295, 0.03555648040771484, 0.03553068923950195, 0.035496673583984374, 0.036357376098632814, 0.03562700653076172, 0.03552355194091797, 0.03553484725952148, 0.03553814315795899, 0.03525302505493164, 0.03562627029418945, 0.035465953826904296, 0.03522355270385742, 0.035198974609375, 0.035190910339355466, 0.035168128967285155, 0.03520307159423828, 0.035135486602783206, 0.03523174285888672, 0.03543040084838867, 0.03538336181640625, 0.03585760116577148, 0.03567919921875, 0.035448833465576174, 0.03537625503540039, 0.035557662963867184, 0.03605155181884766, 0.03603046417236328, 0.03544268798828125, 0.03495935821533203, 0.03518873596191406, 0.03519452667236328, 0.0350456657409668, 0.035186080932617186, 0.03501071929931641, 0.034845184326171875, 0.03485232162475586, 0.03505337524414062, 0.03504969787597656, 0.035068416595458986, 0.03523583984375, 0.035340286254882815, 0.035178497314453126, 0.03494633483886719, 0.035153888702392576, 0.03544579315185547, 0.034985183715820316, 0.03500902557373047, 0.034824127197265624, 0.03495328140258789, 0.03529523086547852, 0.035222751617431644, 0.035017505645751956, 0.0353521614074707, 0.03550864028930664, 0.035411903381347656, 0.035468639373779295, 0.035252063751220704, 0.03545695877075195, 0.03536767959594726, 0.035317951202392575, 0.03521945571899414, 0.035366912841796876, 0.03509862518310547, 0.035370559692382814, 0.03511270523071289, 0.035269313812255856, 0.035244033813476565, 0.03530332946777344, 0.03541763305664063, 0.03543657684326172, 0.03535295867919922, 0.03525833511352539, 0.035493408203125, 0.03556143951416016, 0.03524371337890625, 0.035507198333740234, 0.035552833557128904, 0.03534646224975586, 0.03522396850585938, 0.035462814331054686, 0.03528515243530273, 0.035393726348876955, 0.03529305648803711, 0.03572079849243164, 0.03543049621582031, 0.03497872161865234, 0.03512700653076172, 0.035147136688232425, 0.035300064086914065, 0.03513568115234375, 0.03544063949584961, 0.03524723052978516, 0.03556966400146484, 0.03549008178710938, 0.035342304229736325, 0.03536460876464844, 0.035647361755371094, 0.03544927978515625, 0.037517887115478515, 0.03634175872802734, 0.035581344604492186, 0.035258975982666016, 0.03526860809326172, 0.034961406707763674, 0.03526041412353516, 0.03557580947875977, 0.03513481521606445, 0.03504399871826172, 0.03511862564086914, 0.03738163375854492, 0.035347423553466796, 0.03481577682495117, 0.03507228851318359, 0.03523276901245117, 0.035019710540771486, 0.034928638458251955, 0.035176448822021485, 0.035020671844482425, 0.03517862319946289, 0.03503081512451172, 0.03519830322265625, 0.03500851058959961, 0.03513740921020508, 0.03491123199462891, 0.03507164764404297, 0.03501398468017578, 0.035195903778076174, 0.03504864120483398, 0.03499091339111328, 0.03484675216674805, 0.03501052856445312, 0.03473923110961914, 0.035424480438232424, 0.034688926696777346, 0.034855777740478516, 0.03498307037353516, 0.03554307174682617, 0.03484960174560547, 0.03492252731323242, 0.03504563140869141, 0.03529414367675781, 0.03532470321655273, 0.03540582275390625, 0.03511283111572266, 0.03545449447631836, 0.03516179275512695, 0.03580108642578125, 0.03530246353149414, 0.03512368011474609, 0.03522608184814453, 0.035362239837646484, 0.0353608627319336, 0.035455455780029295, 0.037754878997802735, 0.03544416046142578, 0.03561119842529297, 0.03556697463989258, 0.03553897476196289, 0.03534908676147461, 0.035692543029785154, 0.035743743896484374, 0.035514366149902346, 0.035211265563964846, 0.03536076736450195, 0.03538534545898438, 0.035437889099121093, 0.03529296112060547, 0.03506883239746094, 0.03508428955078125, 0.035358592987060546, 0.03525830459594727, 0.035415233612060545, 0.03538431930541992, 0.03539558410644531, 0.03529235076904297, 0.03534521484375, 0.035641056060791015, 0.035502368927001954, 0.03652403259277344, 0.03531980895996094, 0.035381248474121094, 0.03541356658935547, 0.03547321701049805, 0.03588774490356445, 0.035678207397460936, 0.035359905242919924, 0.03558281707763672, 0.03562905502319336, 0.03565356826782227, 0.03590969467163086, 0.0354856948852539, 0.03560771179199219, 0.035404640197753905, 0.035315711975097655, 0.03538739013671875, 0.035573760986328126, 0.0353177604675293, 0.03680188751220703, 0.03525904083251953, 0.035282943725585936, 0.03534988784790039, 0.035127777099609375, 0.03495116806030273, 0.03527695846557617, 0.035073631286621096, 0.035111328125, 0.035098079681396485, 0.03529372787475586, 0.03522326278686523, 0.035901439666748046, 0.035448833465576174, 0.035503456115722656, 0.03570959854125977, 0.035487743377685545, 0.03568230438232422, 0.03539756774902344, 0.03556358337402344, 0.03558127975463867, 0.03550598526000977, 0.03544150543212891, 0.03577446365356445, 0.035368896484375, 0.035665409088134765, 0.03552880096435547, 0.036657089233398436, 0.036350494384765626, 0.03583590316772461, 0.03545651245117187, 0.03557126235961914, 0.03525888061523438, 0.035279296875, 0.03560595321655274, 0.03533676910400391, 0.03541788864135742, 0.035408096313476564, 0.03523174285888672, 0.03961446380615234, 0.03623116683959961, 0.035545215606689454, 0.03554441452026367, 0.03561321640014648, 0.03530508804321289, 0.03550812911987305, 0.03502889633178711, 0.035184734344482424, 0.035339839935302736, 0.03524291229248047, 0.03522745513916015, 0.035415809631347654, 0.03528953552246094, 0.03554870223999024, 0.03529571151733398, 0.03540377426147461, 0.03560780715942383, 0.035576576232910155, 0.035514366149902346, 0.03565894317626953, 0.03621356964111328, 0.03949363327026367, 0.03598255920410156, 0.035535358428955076, 0.03554947280883789, 0.03567756652832031, 0.03527888107299805, 0.035431007385253906, 0.03556073760986328, 0.035310016632080075, 0.03554111862182617, 0.03577241516113281, 0.03537526321411133, 0.035465217590332034, 0.035211265563964846, 0.03600588989257812, 0.03557558441162109, 0.03532207870483398, 0.035581310272216796, 0.03587926483154297, 0.036020511627197264, 0.035659774780273434, 0.035878398895263675, 0.03546572875976563, 0.03548364639282227, 0.03539328002929688, 0.035837535858154294, 0.03554780960083008, 0.035533824920654294, 0.03535769653320313, 0.0367342414855957, 0.03571779251098633, 0.03561648178100586, 0.03525462341308594, 0.03513753509521484, 0.03543222427368164, 0.03532128143310547, 0.03560335922241211, 0.036015998840332034, 0.03541936111450195, 0.03530217742919922, 0.035606529235839846, 0.035850238800048825, 0.03602447891235352, 0.03529507064819336, 0.035534656524658204, 0.03552614212036133, 0.035515071868896485, 0.03549958419799805, 0.03569440078735352, 0.03608633422851563, 0.03528857421875, 0.03567865753173828, 0.036149375915527346, 0.03532099151611328, 0.03523622512817383, 0.03555100631713867, 0.03516384124755859, 0.03520000076293945, 0.035274177551269534, 0.03517279815673828, 0.03505936050415039, 0.03513190460205078, 0.035437759399414064, 0.03598416137695312, 0.035536319732666015, 0.03524358367919922, 0.03592515182495117, 0.03534627151489258, 0.03541740798950195, 0.03546182250976562, 0.03561827087402344, 0.035230239868164065, 0.035443840026855467, 0.03533913421630859, 0.03535478210449219, 0.03536576080322266, 0.03552150344848633]",tokens/s,28.120108894249366,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1030.61504,965.67296,0.0,570.425344,536.326656,s,1,8.4580712890625,8.4580712890625,0.0,8.4580712890625,8.4580712890625,8.4580712890625,8.4580712890625,[8.4580712890625],,kWh,3.686940990834652e-05,4.05941187368778e-06,1.2865010291990275e-05,5.3793832074024575e-05,,MB,1262.75584,1034.878976,0.0,624.951296,594.377728,s,10,0.20789846420288088,0.020789846420288086,0.000156543809506488,0.020772943496704104,0.020961072158813477,0.02104437599182129,0.02111101905822754,"[0.0211276798248291, 0.020942560195922853, 0.02068841552734375, 0.02059040069580078, 0.020784351348876955, 0.020589759826660156, 0.02076153564453125, 0.02089727973937988, 0.02071865653991699, 0.02079782485961914]",tokens/s,12313.703277296869,kWh,6.076229467082312e-07,6.701004061386872e-08,3.9350112330144745e-07,1.0681341106235472e-06,tokens/kWh,239670278.71673742,MB,1275.990016,1049.55904,0.0,639.63136,607.71072,s,10,11.622058959960938,1.162205895996094,0.011934721498417466,1.1644623413085937,1.176913916015625,1.1784804565429687,1.1797336889648438,"[1.1765657958984375, 1.1506571044921876, 1.1425592041015624, 1.1505360107421876, 1.15258935546875, 1.1618824462890625, 1.168671630859375, 1.1715081787109376, 1.1800469970703125, 1.167042236328125]",tokens/s,54.207262428319105,kWh,3.2591786253707696e-05,3.594396489437822e-06,1.304845407469995e-05,4.9234636817845455e-05,tokens/kWh,1279586.9751833975,,s,630,11.615323753356947,0.018437021830725292,0.0004221994916844008,0.018455056190490725,0.018704652023315428,0.018967825794219968,0.01997489570617676,"[0.01814137649536133, 0.018436960220336914, 0.018668479919433593, 0.021637664794921876, 0.018501983642578126, 0.01927382469177246, 0.01839286422729492, 0.018638431549072267, 0.018555423736572266, 0.018327680587768555, 0.018392959594726564, 0.019791040420532226, 0.01818623924255371, 0.018191295623779295, 0.018612064361572266, 0.01861039924621582, 0.018245695114135742, 0.01832310485839844, 0.020714847564697266, 0.019788543701171876, 0.018700000762939453, 0.018786592483520506, 0.018552831649780274, 0.018560831069946288, 0.018687328338623046, 0.018686464309692383, 0.01859574317932129, 0.018573759078979492, 0.018503231048583986, 0.01861881637573242, 0.018733055114746093, 0.01859107208251953, 0.018687679290771485, 0.018704479217529296, 0.018702335357666015, 0.01866172790527344, 0.018683647155761717, 0.01862326431274414, 0.0186200008392334, 0.01868227195739746, 0.018569311141967772, 0.018708383560180664, 0.018476800918579103, 0.018690496444702148, 0.018503135681152343, 0.018526336669921876, 0.018546239852905273, 0.018535072326660157, 0.01859174346923828, 0.01841971206665039, 0.018475072860717773, 0.018436159133911133, 0.018484512329101564, 0.018520288467407227, 0.01853273582458496, 0.018518016815185546, 0.018453567504882813, 0.018423807144165038, 0.01841862487792969, 0.018448383331298827, 0.01851087951660156, 0.01847769546508789, 0.018246240615844726, 0.018312543869018556, 0.01865180778503418, 0.018472959518432617, 0.018333696365356447, 0.018339296340942383, 0.018265663146972658, 0.01824787139892578, 0.018158143997192382, 0.01814860725402832, 0.018253952026367187, 0.018168352127075196, 0.01827257537841797, 0.01827187156677246, 0.018235647201538085, 0.01822537612915039, 0.018236928939819336, 0.018172319412231446, 0.01814886474609375, 0.018061088562011718, 0.017932607650756837, 0.017897344589233397, 0.017904159545898437, 0.0177674560546875, 0.017744863510131836, 0.017612800598144532, 0.01760630416870117, 0.01762156867980957, 0.017600288391113283, 0.01761075210571289, 0.01765817642211914, 0.017761184692382814, 0.01796339225769043, 0.01787107276916504, 0.01774015998840332, 0.01849235153198242, 0.018161855697631835, 0.018285247802734376, 0.018273759841918945, 0.018332191467285156, 0.01842995262145996, 0.018417503356933592, 0.018472576141357423, 0.01843452835083008, 0.018575424194335936, 0.018757055282592774, 0.018626367568969727, 0.01847372817993164, 0.018288639068603514, 0.018353759765625, 0.018434463500976564, 0.018537919998168947, 0.018538175582885744, 0.018408319473266602, 0.01855232048034668, 0.018626527786254884, 0.018547231674194337, 0.018550783157348632, 0.01864089584350586, 0.018515968322753908, 0.019095552444458007, 0.018570655822753905, 0.01895689582824707, 0.018382848739624022, 0.018041536331176757, 0.018332927703857423, 0.018342144012451173, 0.018625120162963867, 0.018433664321899416, 0.01838688087463379, 0.018428447723388672, 0.018414527893066406, 0.018557823181152344, 0.018448383331298827, 0.01845248031616211, 0.018509824752807616, 0.01835411262512207, 0.018421823501586915, 0.018565120697021483, 0.01840358352661133, 0.01851545524597168, 0.01838719940185547, 0.01831679916381836, 0.018274816513061523, 0.018544607162475586, 0.018613920211791993, 0.01849996757507324, 0.01862006378173828, 0.018917760848999023, 0.018393056869506836, 0.018556575775146484, 0.018534143447875975, 0.01841971206665039, 0.018281055450439454, 0.018175167083740236, 0.018112319946289063, 0.018340864181518556, 0.018423807144165038, 0.01856716728210449, 0.018294368743896484, 0.017826047897338868, 0.01760643196105957, 0.017516927719116213, 0.017448352813720702, 0.017452959060668946, 0.017502143859863283, 0.01749478340148926, 0.017512128829956054, 0.017539392471313475, 0.017547168731689454, 0.017939807891845704, 0.017789695739746095, 0.01778892707824707, 0.017614112854003907, 0.017615583419799803, 0.017639167785644533, 0.01763555145263672, 0.017545055389404297, 0.017825632095336913, 0.01796131134033203, 0.01790496063232422, 0.01801491165161133, 0.017912895202636718, 0.017820159912109376, 0.01772812843322754, 0.01799260711669922, 0.018246559143066405, 0.01741804885864258, 0.017850656509399414, 0.017981407165527343, 0.017821887969970703, 0.017759647369384766, 0.01788105583190918, 0.017848031997680664, 0.017801984786987305, 0.017777727127075194, 0.01802876853942871, 0.018083967208862305, 0.017859167098999023, 0.017731584548950196, 0.017741695404052734, 0.017727615356445313, 0.017735679626464843, 0.018192384719848635, 0.018007295608520508, 0.018130847930908203, 0.01792086410522461, 0.017840192794799804, 0.017778303146362303, 0.017754432678222656, 0.017734848022460937, 0.01777542304992676, 0.01799977684020996, 0.018331392288208008, 0.018027040481567384, 0.017932096481323243, 0.018182144165039063, 0.018317440032958984, 0.018465856552124023, 0.01850364875793457, 0.018402143478393553, 0.0183767032623291, 0.018401119232177736, 0.018423967361450196, 0.018915327072143554, 0.018581504821777343, 0.018607711791992186, 0.018714048385620116, 0.01973062324523926, 0.018649887084960938, 0.019044479370117186, 0.01852150344848633, 0.01854863929748535, 0.018510400772094728, 0.018463775634765624, 0.018576351165771485, 0.018486335754394533, 0.01843315124511719, 0.01847420883178711, 0.01857391929626465, 0.018582656860351564, 0.018587648391723634, 0.018570112228393554, 0.018606048583984375, 0.01865897560119629, 0.01846886444091797, 0.018542976379394532, 0.01845248031616211, 0.01845452880859375, 0.01855036735534668, 0.01822537612915039, 0.01846272087097168, 0.018472320556640626, 0.018452735900878908, 0.018336128234863282, 0.018264064788818358, 0.018151424407958985, 0.018257823944091797, 0.01803468894958496, 0.018148576736450196, 0.018166656494140624, 0.018067455291748045, 0.01781760025024414, 0.01780940818786621, 0.017831487655639647, 0.017868896484375, 0.0179531192779541, 0.017663999557495116, 0.01760256004333496, 0.017662176132202147, 0.01799068832397461, 0.018075584411621094, 0.01852694320678711, 0.018358367919921875, 0.01845849609375, 0.018352256774902344, 0.018533504486083985, 0.018500736236572266, 0.018941568374633788, 0.018478336334228514, 0.018442527770996094, 0.018327903747558594, 0.01837491226196289, 0.01843715286254883, 0.018349023818969728, 0.018380800247192384, 0.01845644760131836, 0.018333152770996095, 0.018385568618774415, 0.018382816314697265, 0.01847817611694336, 0.018551168441772462, 0.018683904647827147, 0.018634368896484375, 0.018545440673828125, 0.018419136047363283, 0.01851055908203125, 0.01839836883544922, 0.01840246391296387, 0.018357984542846678, 0.01827862358093262, 0.018330368041992187, 0.018365440368652345, 0.0182476806640625, 0.01835219192504883, 0.018302656173706053, 0.01815888023376465, 0.018254816055297853, 0.018376768112182616, 0.018245248794555663, 0.01813667106628418, 0.018274080276489257, 0.018305856704711913, 0.017779392242431642, 0.018277727127075195, 0.018225248336791993, 0.018135616302490234, 0.01786675262451172, 0.01803878402709961, 0.018112415313720702, 0.018244800567626954, 0.018092960357666017, 0.018167808532714845, 0.01800956726074219, 0.017949216842651366, 0.01789132881164551, 0.017740928649902343, 0.017985471725463865, 0.01824444770812988, 0.017868896484375, 0.018159616470336915, 0.018520063400268554, 0.018454143524169922, 0.019347360610961914, 0.01880873680114746, 0.018600032806396483, 0.018535072326660157, 0.019037471771240235, 0.01850831985473633, 0.018556640625, 0.018651296615600586, 0.018550912857055665, 0.01860812759399414, 0.018548128128051757, 0.018538591384887695, 0.019111776351928712, 0.018538848876953125, 0.018444000244140626, 0.018395360946655274, 0.0185532169342041, 0.01846272087097168, 0.01843404769897461, 0.018452127456665038, 0.018458976745605468, 0.018423807144165038, 0.01844428825378418, 0.018460800170898437, 0.018521760940551756, 0.018577056884765623, 0.018436288833618163, 0.018385343551635742, 0.019048383712768555, 0.0182969913482666, 0.018485088348388672, 0.018357280731201173, 0.01826095962524414, 0.01837238311767578, 0.018289119720458984, 0.018351871490478514, 0.01828598403930664, 0.0183253116607666, 0.0182685432434082, 0.018264415740966797, 0.019470399856567382, 0.020112512588500976, 0.018842496871948243, 0.01804697608947754, 0.018122400283813477, 0.018231903076171875, 0.018262847900390625, 0.018332639694213868, 0.018440160751342773, 0.018314783096313476, 0.019291711807250977, 0.019452608108520508, 0.018624319076538085, 0.01853081512451172, 0.0199781436920166, 0.0183306884765625, 0.018428863525390624, 0.018472959518432617, 0.018579456329345705, 0.0184520320892334, 0.01842835235595703, 0.018327360153198243, 0.018374847412109374, 0.018308128356933594, 0.018311424255371092, 0.01833238410949707, 0.018296831130981444, 0.018403295516967774, 0.018409503936767577, 0.018562431335449218, 0.01846131134033203, 0.018374656677246092, 0.018391040802001952, 0.01927168083190918, 0.018388992309570314, 0.01843404769897461, 0.018431711196899413, 0.018401567459106444, 0.018472959518432617, 0.01835212707519531, 0.01848121643066406, 0.019966943740844727, 0.01867286491394043, 0.01858665657043457, 0.018655967712402344, 0.018644960403442382, 0.018509855270385744, 0.018554943084716797, 0.018646272659301757, 0.01882307243347168, 0.018313472747802734, 0.018315807342529297, 0.018503679275512695, 0.01854863929748535, 0.018514015197753905, 0.018907360076904297, 0.018380479812622072, 0.018579551696777344, 0.018488576889038086, 0.018434816360473633, 0.018744895935058594, 0.018504127502441407, 0.01844428825378418, 0.018372608184814454, 0.01832262420654297, 0.018455360412597658, 0.01829478454589844, 0.018794048309326173, 0.018706880569458007, 0.01884547233581543, 0.018403104782104492, 0.01839084815979004, 0.018829343795776367, 0.01860259246826172, 0.01841766357421875, 0.01846681594848633, 0.018465824127197265, 0.01841219139099121, 0.019022207260131836, 0.019781312942504882, 0.018579168319702147, 0.01846940803527832, 0.018601984024047852, 0.01845248031616211, 0.01841152000427246, 0.01857151985168457, 0.018562816619873048, 0.018536447525024414, 0.018547807693481445, 0.018551712036132813, 0.01859686470031738, 0.018723424911499024, 0.01868671989440918, 0.018708127975463867, 0.018808832168579103, 0.01862838363647461, 0.01850595283508301, 0.01848320007324219, 0.018554880142211915, 0.018550783157348632, 0.01845577621459961, 0.0182772159576416, 0.018271360397338867, 0.01820719909667969, 0.018255552291870116, 0.018322431564331054, 0.01846646308898926, 0.018495487213134765, 0.018912511825561522, 0.018602880477905273, 0.01863257598876953, 0.01860406494140625, 0.018608095169067383, 0.01857535934448242, 0.01855062484741211, 0.01892572784423828, 0.01857961654663086, 0.018634592056274414, 0.01855238342285156, 0.01864067268371582, 0.018600608825683592, 0.01861625671386719, 0.018556320190429687, 0.018697887420654296, 0.018534400939941405, 0.018496768951416016, 0.01860736083984375, 0.018625024795532227, 0.01851753616333008, 0.018356767654418946, 0.018487295150756835, 0.018544607162475586, 0.018276704788208007, 0.01832124710083008, 0.018431840896606447, 0.018421024322509767, 0.01837948799133301, 0.018308767318725584, 0.019818944931030272, 0.018652639389038085, 0.019240800857543944, 0.018862688064575195, 0.018568672180175782, 0.01851651191711426, 0.018499584197998048, 0.018505727767944336, 0.018976768493652343, 0.019820383071899414, 0.01951145553588867, 0.018425567626953125, 0.018493728637695314, 0.018411359786987304, 0.02023644828796387, 0.01848464012145996, 0.018332160949707032, 0.018507648468017578, 0.018454751968383788, 0.01839891242980957, 0.018481056213378907, 0.01882383918762207, 0.018414655685424806, 0.019628799438476563, 0.019998655319213868, 0.02098966407775879, 0.01866156768798828, 0.01859190368652344, 0.01848512077331543, 0.018663551330566405, 0.018822399139404297, 0.018500288009643553, 0.018470848083496094, 0.018470975875854494, 0.01843814468383789, 0.018739423751831054, 0.018706207275390626, 0.018481151580810547, 0.018382623672485353, 0.018378976821899415, 0.018280607223510742, 0.01989347267150879, 0.018350719451904297, 0.018524160385131837, 0.01861417579650879, 0.018813024520874022, 0.018706623077392577, 0.01850553512573242, 0.01856719970703125, 0.018668800354003905, 0.018504032135009764, 0.018473344802856444, 0.018520063400268554, 0.01856265640258789, 0.018309120178222657, 0.018538496017456055, 0.01865497589111328, 0.01854841613769531, 0.01862713623046875, 0.018569215774536133, 0.018452064514160156, 0.01877382469177246, 0.018606719970703126, 0.01859152030944824, 0.01856940841674805, 0.01842918395996094, 0.0185097599029541, 0.018459680557250977, 0.018411296844482422, 0.01835775947570801, 0.018416128158569335, 0.018506784439086914, 0.018501760482788086, 0.018303840637207032, 0.018425504684448243, 0.018261663436889647, 0.01843235206604004, 0.018546815872192382, 0.018388511657714844, 0.018496192932128907, 0.01856233596801758, 0.018448896408081054, 0.018493631362915038, 0.01850067138671875, 0.018537343978881834, 0.018540351867675782, 0.018409311294555666, 0.018444639205932617, 0.01835215950012207, 0.018268159866333008, 0.018411136627197264, 0.018358720779418945, 0.01857974433898926, 0.01840867233276367, 0.01846668815612793, 0.018501792907714844, 0.018554912567138673, 0.018583936691284178, 0.018594015121459962, 0.01849123191833496, 0.018507232666015627, 0.01851651191711426, 0.018524160385131837, 0.01846873664855957, 0.018489471435546877, 0.01852604866027832, 0.01856528091430664, 0.01862406349182129, 0.018548608779907227, 0.01854707145690918, 0.01853785514831543, 0.018666303634643555, 0.01861030387878418, 0.018593568801879883, 0.018575071334838867, 0.019274080276489258, 0.018583263397216797]",tokens/s,54.23869479470379,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1572.00384,1861.156864,0.0,1465.909248,1358.169088,s,1,8.7545458984375,8.7545458984375,0.0,8.7545458984375,8.7545458984375,8.7545458984375,8.7545458984375,[8.7545458984375],,kWh,5.511534694168555e-05,6.072308957361274e-06,1.9198904248007853e-05,8.038656014705468e-05,,MB,1672.114176,1882.128384,0.0,1472.200704,1356.544512,s,10,0.4443775062561035,0.04443775062561035,0.0001839494016502686,0.04437872123718262,0.044650616455078125,0.044781340026855465,0.04488591888427734,"[0.04491206359863281, 0.04436921691894531, 0.044349407196044924, 0.044292064666748045, 0.04443267059326172, 0.044427104949951175, 0.044268577575683594, 0.04431660842895508, 0.04462156677246094, 0.04438822555541992]",tokens/s,5760.8676495984055,kWh,1.3502476793010703e-06,1.4890786084304384e-07,8.933430085805946e-07,2.3924985487247085e-06,tokens/kWh,107001109.83827247,MB,1680.297984,1882.128384,0.0,1472.200704,1409.94816,s,10,11.17914501953125,1.117914501953125,0.003509245620187121,1.1180933227539063,1.122057958984375,1.1229037963867188,1.1235804663085938,"[1.119744873046875, 1.1185635986328124, 1.1198131103515625, 1.1237496337890625, 1.1218699951171875, 1.111178955078125, 1.117623046875, 1.11444482421875, 1.1170386962890626, 1.1151182861328126]",tokens/s,56.354935811219704,kWh,3.26895830231972e-05,3.604478886908341e-06,1.578705922462118e-05,5.2081121134726705e-05,tokens/kWh,1209651.3789906262,,s,630,11.176699878692624,0.017740793458242265,0.00029612535460284,0.017657711982727052,0.0180024995803833,0.01817117118835449,0.019014234676361097,"[0.017864799499511717, 0.017692703247070313, 0.01762735939025879, 0.01767331123352051, 0.01756972885131836, 0.017605600357055665, 0.017727487564086913, 0.017770496368408203, 0.017764352798461915, 0.018087743759155273, 0.017756351470947264, 0.01768387222290039, 0.017666431427001954, 0.017735391616821288, 0.017735551834106446, 0.017759199142456054, 0.01789507293701172, 0.017671648025512694, 0.017693023681640625, 0.017569055557250978, 0.01766022491455078, 0.017584672927856447, 0.017732736587524413, 0.01768134307861328, 0.01761075210571289, 0.01756972885131836, 0.0176265926361084, 0.018002336502075195, 0.01767647933959961, 0.017547264099121093, 0.017704063415527344, 0.017658239364624025, 0.017627647399902344, 0.017588031768798827, 0.017600191116333007, 0.019212799072265627, 0.01799782371520996, 0.018074911117553712, 0.017809343338012696, 0.017707807540893555, 0.017703935623168944, 0.0176363525390625, 0.017624063491821287, 0.01773465538024902, 0.01764556884765625, 0.017625087738037108, 0.017657312393188476, 0.017555999755859374, 0.017524736404418945, 0.017620096206665038, 0.017707647323608397, 0.01756755256652832, 0.017543615341186522, 0.017555360794067384, 0.017578079223632814, 0.017584127426147463, 0.01762099266052246, 0.01760870361328125, 0.017638879776000975, 0.017523231506347655, 0.020685983657836915, 0.018528480529785157, 0.017826112747192382, 0.01776076889038086, 0.017960960388183594, 0.017564960479736328, 0.017708799362182618, 0.01763206481933594, 0.017829919815063478, 0.01798124885559082, 0.018213184356689453, 0.017772544860839845, 0.017496063232421876, 0.01770086479187012, 0.019812255859375, 0.017743967056274415, 0.01754521560668945, 0.017585344314575195, 0.017621824264526367, 0.017580032348632812, 0.017765888214111326, 0.017696704864501953, 0.017656288146972655, 0.01758415985107422, 0.017571903228759764, 0.017542943954467774, 0.017635488510131837, 0.0175534725189209, 0.017497695922851563, 0.017662303924560547, 0.017550432205200195, 0.01765456008911133, 0.017665632247924806, 0.017691232681274413, 0.0175861759185791, 0.01759846305847168, 0.017547264099121093, 0.017766016006469727, 0.017604991912841796, 0.017635072708129883, 0.01759187126159668, 0.017598207473754884, 0.017524991989135742, 0.017702720642089845, 0.017609600067138673, 0.01765376091003418, 0.01780940818786621, 0.017625087738037108, 0.017719039916992186, 0.017689952850341795, 0.017685407638549804, 0.017700544357299806, 0.01779715156555176, 0.01775644874572754, 0.017759424209594726, 0.01777542304992676, 0.017739072799682617, 0.017791296005249025, 0.01792243194580078, 0.01816166305541992, 0.01804204750061035, 0.017916736602783204, 0.01788217544555664, 0.01785331153869629, 0.01787091255187988, 0.018142784118652344, 0.018132991790771484, 0.01786444854736328, 0.01789673614501953, 0.017529024124145507, 0.017726240158081056, 0.017610240936279296, 0.017617408752441405, 0.017882335662841798, 0.01777129554748535, 0.017719295501708983, 0.017738912582397463, 0.01775292778015137, 0.017672191619873046, 0.01762054443359375, 0.017767936706542968, 0.017610719680786133, 0.017570751190185547, 0.017550975799560546, 0.0176376953125, 0.017705055236816408, 0.017719295501708983, 0.01752182388305664, 0.017559776306152342, 0.017588863372802733, 0.017612800598144532, 0.017969152450561524, 0.017917951583862304, 0.017778688430786133, 0.017795072555541993, 0.01767011260986328, 0.018067487716674803, 0.01765376091003418, 0.01761075210571289, 0.01845248031616211, 0.017936384201049805, 0.017609952926635742, 0.017720096588134764, 0.01785856056213379, 0.01780735969543457, 0.017780736923217775, 0.017870847702026366, 0.01780940818786621, 0.017827840805053712, 0.017803007125854493, 0.017862655639648437, 0.017858816146850587, 0.018030527114868165, 0.01793440055847168, 0.017950368881225587, 0.017889631271362304, 0.01778483200073242, 0.017675968170166017, 0.017754432678222656, 0.017905664443969727, 0.017966943740844725, 0.017690784454345704, 0.017731584548950196, 0.017762304306030274, 0.01778998374938965, 0.01765216064453125, 0.01776652717590332, 0.01762544059753418, 0.017592384338378907, 0.01809984016418457, 0.017715103149414064, 0.0177139835357666, 0.01764556884765625, 0.017630815505981445, 0.017637792587280272, 0.017663999557495116, 0.0175914249420166, 0.017693567276000976, 0.017924095153808595, 0.01796505546569824, 0.017903615951538086, 0.01804083251953125, 0.01802239990234375, 0.017931936264038086, 0.018085407257080077, 0.018436576843261717, 0.017955167770385742, 0.018520063400268554, 0.017944576263427735, 0.017878656387329103, 0.018010112762451173, 0.017951072692871092, 0.01804086494445801, 0.018077375411987305, 0.01790598487854004, 0.01788857650756836, 0.017867456436157225, 0.01764556884765625, 0.0175631046295166, 0.017591903686523438, 0.01755232048034668, 0.017656831741333007, 0.017802240371704102, 0.017817535400390626, 0.01774188804626465, 0.017612800598144532, 0.017679424285888673, 0.017585088729858398, 0.017526784896850587, 0.017725343704223632, 0.01757939147949219, 0.017593055725097655, 0.017563648223876953, 0.01762099266052246, 0.01764339256286621, 0.017725568771362304, 0.01762918472290039, 0.01761075210571289, 0.017582080841064454, 0.0176200008392334, 0.01784726333618164, 0.01790924835205078, 0.01785251235961914, 0.017886848449707032, 0.01790208053588867, 0.018223392486572267, 0.018255872726440428, 0.01807360076904297, 0.018043903350830077, 0.018037567138671873, 0.018145151138305664, 0.017888927459716798, 0.018495487213134765, 0.018255872726440428, 0.01823468780517578, 0.01816374397277832, 0.017996448516845703, 0.017889280319213868, 0.018147327423095702, 0.017835424423217772, 0.01776291275024414, 0.017864255905151366, 0.017883167266845704, 0.01790403175354004, 0.017874496459960938, 0.018106559753417968, 0.017873151779174805, 0.01803264045715332, 0.01763942337036133, 0.01761894416809082, 0.01763942337036133, 0.017710527420043944, 0.017685056686401367, 0.017625087738037108, 0.017655807495117186, 0.017682207107543944, 0.017637407302856446, 0.017701055526733397, 0.017643232345581055, 0.017583776473999023, 0.01772127914428711, 0.017600992202758788, 0.017665599822998045, 0.017828479766845703, 0.01799580764770508, 0.01789254379272461, 0.017879871368408202, 0.01864089584350586, 0.017898719787597658, 0.017756959915161134, 0.017712480545043947, 0.017676959991455077, 0.0177675838470459, 0.01750921630859375, 0.017657087326049804, 0.017627904891967774, 0.01761484718322754, 0.017669824600219725, 0.01764588737487793, 0.017582080841064454, 0.01802239990234375, 0.017716672897338866, 0.01776083183288574, 0.01840127944946289, 0.017709056854248048, 0.017563648223876953, 0.01769385528564453, 0.017718048095703126, 0.01763539123535156, 0.017718847274780274, 0.017626752853393556, 0.01765193557739258, 0.017635936737060546, 0.01767011260986328, 0.01763248062133789, 0.01763532829284668, 0.0176978874206543, 0.0175664005279541, 0.017688800811767578, 0.017628992080688476, 0.017567935943603515, 0.017504255294799806, 0.01754857635498047, 0.017586624145507813, 0.017481056213378907, 0.017761215209960938, 0.01753900718688965, 0.017598527908325196, 0.017564767837524413, 0.01758812713623047, 0.01756208038330078, 0.017535167694091795, 0.0175097599029541, 0.017572832107543946, 0.01755683135986328, 0.017558176040649413, 0.017508352279663086, 0.017501344680786134, 0.01747235107421875, 0.017522335052490234, 0.017455295562744142, 0.0175699520111084, 0.017571647644042968, 0.017598400115966795, 0.017530975341796876, 0.017590112686157226, 0.01754319953918457, 0.01757375907897949, 0.017521055221557617, 0.01760771179199219, 0.017504287719726563, 0.0175482234954834, 0.01763865661621094, 0.017676319122314453, 0.017803167343139647, 0.017767135620117187, 0.017647712707519532, 0.01760256004333496, 0.017747968673706056, 0.01787059211730957, 0.01755366325378418, 0.017573888778686524, 0.0175797119140625, 0.017548736572265626, 0.01749900817871094, 0.017528543472290038, 0.017613088607788086, 0.0175861759185791, 0.017557407379150392, 0.01759651184082031, 0.017528831481933595, 0.01762643241882324, 0.01763603210449219, 0.020391040802001954, 0.017656448364257813, 0.017705215454101562, 0.017596351623535157, 0.017549375534057617, 0.017657632827758788, 0.017557727813720704, 0.01763532829284668, 0.017498111724853514, 0.017604608535766602, 0.01753628730773926, 0.01769913673400879, 0.017588191986083985, 0.017680831909179687, 0.01800396728515625, 0.01826812744140625, 0.017889312744140625, 0.01802595138549805, 0.01790822410583496, 0.01778819274902344, 0.01787980842590332, 0.017821407318115233, 0.018286880493164064, 0.017895360946655274, 0.01787091255187988, 0.017726783752441407, 0.017746623992919923, 0.01776950454711914, 0.017630176544189455, 0.017645248413085936, 0.01762131118774414, 0.0176363525390625, 0.017622016906738282, 0.017663423538208007, 0.017558080673217773, 0.017715200424194336, 0.018067455291748045, 0.017674240112304687, 0.017625087738037108, 0.01764512062072754, 0.017682367324829102, 0.017526880264282226, 0.017524703979492188, 0.017529279708862304, 0.01785158348083496, 0.017531423568725585, 0.0177476806640625, 0.017543743133544922, 0.017510400772094727, 0.017528831481933595, 0.01762918472290039, 0.017669792175292968, 0.017529184341430665, 0.017542240142822265, 0.017561567306518554, 0.017578943252563477, 0.017524736404418945, 0.017475584030151366, 0.017539039611816406, 0.017755552291870116, 0.017561855316162108, 0.01760908889770508, 0.01925324821472168, 0.019099647521972657, 0.017713151931762695, 0.017677759170532225, 0.017588768005371094, 0.01766099166870117, 0.017606143951416017, 0.017539583206176757, 0.01762918472290039, 0.017520639419555666, 0.017491647720336914, 0.01746361541748047, 0.017483776092529296, 0.01753433609008789, 0.017604991912841796, 0.017488128662109376, 0.017508352279663086, 0.017589887619018554, 0.01880512046813965, 0.01765135955810547, 0.017561952590942384, 0.018274303436279296, 0.017886367797851563, 0.017492128372192384, 0.017535680770874022, 0.01756159973144531, 0.017537023544311522, 0.017528160095214844, 0.01775027275085449, 0.017830303192138672, 0.017713151931762695, 0.01784182357788086, 0.017694623947143554, 0.018373056411743163, 0.017571807861328125, 0.01756572723388672, 0.01762214469909668, 0.01749286460876465, 0.017952768325805665, 0.017530143737792967, 0.017799840927124024, 0.017604671478271484, 0.017645408630371093, 0.017627296447753907, 0.01757980728149414, 0.01751203155517578, 0.017582176208496093, 0.017473888397216798, 0.01771334457397461, 0.01758361625671387, 0.01754368019104004, 0.018017440795898437, 0.017774784088134765, 0.01753763198852539, 0.017631296157836915, 0.017649663925170898, 0.0175897274017334, 0.01757814407348633, 0.01757369613647461, 0.017582656860351563, 0.01769385528564453, 0.01769558334350586, 0.017661760330200196, 0.017811647415161135, 0.017767999649047853, 0.017793472290039063, 0.017915903091430666, 0.017991296768188475, 0.01801625633239746, 0.018569536209106445, 0.01821014404296875, 0.017996448516845703, 0.01799577522277832, 0.01804697608947754, 0.017958688735961913, 0.01786617660522461, 0.017807647705078124, 0.017731712341308593, 0.017514080047607423, 0.017677087783813477, 0.01761267280578613, 0.01766537666320801, 0.017697471618652344, 0.01823958396911621, 0.017909759521484374, 0.017739391326904295, 0.017615232467651367, 0.01764761543273926, 0.01757798385620117, 0.017534975051879884, 0.01755340766906738, 0.018585599899291993, 0.0178272647857666, 0.01760927963256836, 0.017541120529174805, 0.01761689567565918, 0.01760870361328125, 0.017780351638793945, 0.017667583465576172, 0.01763007926940918, 0.01762291145324707, 0.017684608459472655, 0.017833951950073243, 0.01792617607116699, 0.017756160736083985, 0.01776630401611328, 0.017582176208496093, 0.01817724800109863, 0.01766275215148926, 0.017575935363769533, 0.017534975051879884, 0.01753900718688965, 0.017450624465942383, 0.017588672637939454, 0.0175861759185791, 0.017532928466796875, 0.017548736572265626, 0.01767875289916992, 0.017921375274658202, 0.017603328704833984, 0.017672256469726564, 0.017616575241088867, 0.017688480377197266, 0.0175251522064209, 0.01761193656921387, 0.01765052795410156, 0.01761075210571289, 0.01756060791015625, 0.01750262451171875, 0.017611135482788087, 0.01754070472717285, 0.017611360549926756, 0.017570751190185547, 0.01798739242553711, 0.01769286346435547, 0.01749932861328125, 0.01761568069458008, 0.017588224411010742, 0.01757798385620117, 0.017537023544311522, 0.018276544570922853, 0.018802047729492188, 0.018051519393920898, 0.01785958480834961, 0.017728511810302734, 0.01779097557067871, 0.017657791137695313, 0.017590335845947266, 0.01763942337036133, 0.017506303787231444, 0.017582080841064454, 0.017653408050537108, 0.017612735748291017, 0.01762892723083496, 0.01765648078918457, 0.017530879974365234, 0.017557376861572264, 0.017537151336669922, 0.017551103591918946, 0.01772159957885742, 0.017665151596069337, 0.01749081611633301, 0.017596416473388672, 0.017534975051879884, 0.017711103439331053, 0.01776131248474121, 0.019202207565307616, 0.018428735733032227, 0.017860128402709962, 0.017736160278320312, 0.0176680965423584, 0.017633279800415038, 0.01782374382019043, 0.01761484718322754, 0.017620927810668947, 0.017495840072631837, 0.017481599807739258, 0.017487840652465822, 0.017760704040527344, 0.017616191864013673, 0.017918655395507813, 0.017556575775146483, 0.017505184173583984, 0.017496063232421876, 0.017530879974365234, 0.01767398452758789, 0.017524255752563476, 0.01749065589904785, 0.017510400772094727, 0.017588224411010742, 0.01756070327758789, 0.017590816497802735, 0.017665855407714842, 0.017537567138671876, 0.017540191650390623]",tokens/s,56.36726465215715,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,6641.700864,9387.835392,0.0,8992.587776,8404.320768,s,1,14.4562421875,14.4562421875,0.0,14.4562421875,14.4562421875,14.4562421875,14.4562421875,[14.4562421875],,kWh,0.00021343599917499127,2.3536222536619963e-05,9.489063146797538e-05,0.0003318628531795866,,MB,1564.639232,9402.515456,0.0,8992.587776,7879.473152,s,10,3.2490415954589844,0.32490415954589846,0.0005050043342131917,0.3249153594970703,0.3256789093017578,0.32569710540771485,0.3257116622924805,"[0.3243965454101562, 0.3239740295410156, 0.324656494140625, 0.32492047119140627, 0.32567486572265625, 0.32468362426757813, 0.32495562744140627, 0.3257153015136719, 0.32515438842773436, 0.32491024780273436]",tokens/s,787.9246617150048,kWh,9.496914052687149e-06,1.0473449039228742e-06,6.288436930387383e-06,1.6832695886997404e-05,tokens/kWh,15208496.70894072,MB,1567.612928,9402.515456,0.0,8992.587776,8125.43744,s,10,22.417164306640625,2.2417164306640625,0.0017633449796451378,2.241837158203125,2.243854833984375,2.244349658203125,2.244745517578125,"[2.24002197265625, 2.2391669921875, 2.240408935546875, 2.244844482421875, 2.239778076171875, 2.243744873046875, 2.241950927734375, 2.241723388671875, 2.243001220703125, 2.2425234375]",tokens/s,28.10346533496993,kWh,6.519396790940043e-05,7.190828362416798e-06,4.344726952481049e-05,0.00011583206579662773,tokens/kWh,543890.8437548918,,s,630,22.412986602783178,0.03557616921076699,0.0003406645445687422,0.03560273551940918,0.036014133071899414,0.03606753921508789,0.036162389793396,"[0.03562540817260742, 0.03507193756103515, 0.03492460632324219, 0.0348364143371582, 0.03492435073852539, 0.035019008636474606, 0.035110080718994144, 0.03505596923828125, 0.035095008850097656, 0.03513753509521484, 0.035233440399169924, 0.035176799774169924, 0.035243648529052735, 0.035092609405517575, 0.03514291381835938, 0.03523673629760742, 0.03522544097900391, 0.03515158462524414, 0.0351965446472168, 0.03526342391967773, 0.03540566253662109, 0.03540598297119141, 0.0354304313659668, 0.03528496170043945, 0.03542985534667969, 0.035356704711914065, 0.03539926528930664, 0.03548787307739258, 0.03551337432861328, 0.035467105865478514, 0.03546511840820313, 0.03556272125244141, 0.035601184844970706, 0.03555327987670898, 0.03572537612915039, 0.0356638069152832, 0.035743743896484374, 0.0357210578918457, 0.03566608047485351, 0.03566294479370117, 0.03573980712890625, 0.03578668975830078, 0.0358326416015625, 0.035778209686279296, 0.035804672241210936, 0.03578524780273438, 0.03587923049926758, 0.03590348815917969, 0.035907230377197265, 0.03583350372314453, 0.03600624084472656, 0.03598332977294922, 0.03585676956176758, 0.03581740951538086, 0.03594655990600586, 0.03599087905883789, 0.03606185531616211, 0.03599283218383789, 0.0360568962097168, 0.0360621452331543, 0.03611008071899414, 0.036063488006591794, 0.03609571075439453, 0.03567795181274414, 0.035084320068359376, 0.035041534423828125, 0.03491635131835937, 0.034917953491210935, 0.03499411010742187, 0.03516604614257812, 0.03508700942993164, 0.035074047088623043, 0.03501039886474609, 0.03528646469116211, 0.035100833892822265, 0.035007232666015624, 0.03511072158813477, 0.03514934539794922, 0.03516985702514648, 0.035138721466064456, 0.03505740737915039, 0.03516019058227539, 0.035168128967285155, 0.035422206878662106, 0.03539952087402344, 0.03530358505249023, 0.0353361930847168, 0.035401569366455075, 0.03543056106567383, 0.035423328399658206, 0.035376033782958984, 0.035479328155517575, 0.03541638565063476, 0.03542825698852539, 0.03543392181396485, 0.03552707290649414, 0.03564300918579102, 0.03570127868652344, 0.03564739227294922, 0.03573360061645508, 0.0356940803527832, 0.03566233444213867, 0.035640705108642576, 0.03569113540649414, 0.035694751739501956, 0.03575190353393555, 0.03575590515136719, 0.03575209426879883, 0.03580912017822266, 0.035827713012695314, 0.035768318176269534, 0.035835166931152344, 0.03587350463867187, 0.036026336669921874, 0.03584841537475586, 0.03591088104248047, 0.035983230590820314, 0.036059711456298826, 0.03602025604248047, 0.036014110565185546, 0.03606537628173828, 0.03611366271972656, 0.035937023162841794, 0.036073471069335936, 0.036183807373046876, 0.036286720275878905, 0.03581542587280274, 0.03510204696655273, 0.03499689483642578, 0.03492470550537109, 0.03498787307739258, 0.03496160125732422, 0.03504048156738281, 0.03502755355834961, 0.035194881439208986, 0.03518217468261719, 0.03520467376708984, 0.03530217742919922, 0.035162078857421876, 0.03527484893798828, 0.03535257720947266, 0.035282943725585936, 0.03530342483520508, 0.03520675277709961, 0.03521769714355469, 0.035244480133056644, 0.035272384643554686, 0.03525228881835937, 0.03534019088745117, 0.035384574890136716, 0.03533903884887695, 0.03523497772216797, 0.0353897590637207, 0.035573951721191405, 0.035536319732666015, 0.035438720703125, 0.035449886322021486, 0.03541801452636719, 0.035546974182128904, 0.03555123138427734, 0.0355978889465332, 0.03563708877563477, 0.03571785736083984, 0.03579276657104492, 0.035784702301025394, 0.035702880859375, 0.03574262237548828, 0.03569561767578125, 0.035741855621337894, 0.0357210578918457, 0.035861534118652345, 0.03585142517089844, 0.03589510345458984, 0.035831649780273436, 0.035783039093017575, 0.0357509765625, 0.0358397445678711, 0.035799808502197265, 0.036026336669921874, 0.0360450553894043, 0.03613491058349609, 0.035983360290527344, 0.03597312164306641, 0.03606118392944336, 0.036132865905761716, 0.036067329406738284, 0.036050430297851564, 0.03606915283203125, 0.03614998245239258, 0.03597321701049805, 0.0352525749206543, 0.03530758285522461, 0.03508736038208008, 0.035324737548828124, 0.03512543869018555, 0.03517030334472656, 0.0350860481262207, 0.035330337524414064, 0.03527376174926758, 0.03523231887817383, 0.03515228652954101, 0.03523788833618164, 0.035186912536621096, 0.03527459335327148, 0.03516851043701172, 0.03522937774658203, 0.03528297424316406, 0.035315582275390625, 0.035299423217773435, 0.0353175048828125, 0.03534668731689453, 0.03546268844604492, 0.035425888061523435, 0.035496448516845705, 0.03551785659790039, 0.03565667343139649, 0.03578396987915039, 0.03573657608032227, 0.03558572769165039, 0.035593696594238285, 0.035565921783447266, 0.035541217803955076, 0.035560672760009765, 0.03561510467529297, 0.03574758529663086, 0.03587696075439453, 0.035770942687988284, 0.035833854675292966, 0.03573676681518555, 0.03577859115600586, 0.03576502227783203, 0.03580723190307617, 0.0357531852722168, 0.035865345001220704, 0.035823070526123046, 0.03581542587280274, 0.03575235366821289, 0.035829025268554686, 0.03586547088623047, 0.03600598526000977, 0.03596278381347656, 0.0360299186706543, 0.03621532821655273, 0.03617724609375, 0.03606771087646484, 0.03609219360351563, 0.036007102966308595, 0.03611219024658203, 0.03602092742919922, 0.036025726318359375, 0.03596777725219726, 0.03617308807373047, 0.03575833511352539, 0.03521353530883789, 0.035116832733154295, 0.03498147201538086, 0.03622092819213867, 0.034977344512939455, 0.03497865676879883, 0.035135486602783206, 0.03513459014892578, 0.03510156631469727, 0.03513328170776367, 0.03510444641113281, 0.035200576782226566, 0.035144065856933596, 0.035098400115966794, 0.035130111694335935, 0.035198974609375, 0.035209217071533204, 0.03525577545166016, 0.03530124664306641, 0.035485599517822264, 0.03546803283691406, 0.03548124694824219, 0.03543075180053711, 0.03543225479125976, 0.03539068984985352, 0.03549657440185547, 0.03555977630615234, 0.03554508972167969, 0.03544473648071289, 0.03545395278930664, 0.03538547134399414, 0.03543743896484375, 0.03540172958374024, 0.035471359252929685, 0.035671199798583984, 0.035655998229980466, 0.03560428619384766, 0.03569452667236328, 0.03568310546875, 0.03564246368408203, 0.03563020706176758, 0.035718944549560545, 0.035689952850341794, 0.03564352035522461, 0.03574620819091797, 0.035768192291259764, 0.03577775955200195, 0.03576924896240234, 0.03570483016967774, 0.035899391174316404, 0.03594649505615234, 0.03594153594970703, 0.03589120101928711, 0.03610323333740234, 0.03616745758056641, 0.036001792907714845, 0.03590758514404297, 0.03594768142700195, 0.03587916946411133, 0.036014335632324215, 0.036011966705322265, 0.03604931259155274, 0.03595468902587891, 0.035331966400146485, 0.0350885124206543, 0.03501571273803711, 0.035064609527587894, 0.03504352188110352, 0.03502643203735351, 0.03508070373535156, 0.03522099304199219, 0.03517113494873047, 0.03523142242431641, 0.0351723518371582, 0.03522355270385742, 0.035133472442626955, 0.035279838562011716, 0.035293598175048825, 0.03528755187988281, 0.03521987152099609, 0.03529404830932617, 0.035794849395751956, 0.035310462951660154, 0.03536288070678711, 0.03544678497314453, 0.035491584777832035, 0.03544281768798828, 0.03541516876220703, 0.03555228805541992, 0.035565536499023435, 0.03561881637573242, 0.035552928924560544, 0.03558457565307617, 0.03557936096191406, 0.035598751068115234, 0.03561580657958984, 0.035811328887939455, 0.035695457458496095, 0.035828990936279295, 0.03578291320800781, 0.03578112030029297, 0.0356864013671875, 0.03586457443237305, 0.03592396926879883, 0.0358823356628418, 0.03584579086303711, 0.03592704010009766, 0.03582156753540039, 0.035847518920898436, 0.03580380630493164, 0.035917438507080075, 0.035926273345947266, 0.03602035140991211, 0.03592953491210937, 0.03593606567382813, 0.03596774291992187, 0.03603472137451172, 0.03593603134155274, 0.03591993713378906, 0.03597638320922852, 0.03602924728393555, 0.03604889678955078, 0.03597107315063477, 0.03599155044555664, 0.03613081741333008, 0.036001182556152346, 0.03529584121704102, 0.03508367919921875, 0.0349923210144043, 0.03510857772827149, 0.03506041717529297, 0.035059711456298825, 0.03503104019165039, 0.03506300735473633, 0.035068126678466795, 0.035099006652832034, 0.03511881637573242, 0.03513782501220703, 0.03518624114990234, 0.035197566986083985, 0.03515801620483398, 0.035209217071533204, 0.03522544097900391, 0.03523535919189453, 0.035426334381103514, 0.03543888092041016, 0.03541241455078125, 0.03537395095825195, 0.03543276977539062, 0.035399681091308595, 0.03539558410644531, 0.03550892639160156, 0.03559417724609375, 0.035622974395751954, 0.035556991577148436, 0.0355425910949707, 0.03555411148071289, 0.035598175048828125, 0.03560630416870117, 0.035581695556640626, 0.035625503540039065, 0.035761920928955075, 0.03574390411376953, 0.035696830749511715, 0.03561270523071289, 0.035880672454833985, 0.03592012786865234, 0.03586191940307617, 0.03584588623046875, 0.03583369445800781, 0.03576115036010742, 0.035805118560791015, 0.03587475204467774, 0.035902751922607425, 0.03589616012573242, 0.035878623962402344, 0.03584233474731445, 0.03590758514404297, 0.035917823791503906, 0.03596492767333984, 0.035953792572021484, 0.0360398063659668, 0.03597040176391601, 0.0361453742980957, 0.036133056640625, 0.036139007568359374, 0.03609193420410156, 0.03614716720581055, 0.03589225769042969, 0.0351798095703125, 0.034998943328857425, 0.03501590347290039, 0.03501536178588867, 0.03497129440307617, 0.03505811309814453, 0.035147232055664064, 0.0352562255859375, 0.03523798370361328, 0.035160511016845704, 0.03510691070556641, 0.035297279357910154, 0.035138782501220704, 0.035189537048339846, 0.035192832946777344, 0.03512934494018555, 0.03506380844116211, 0.035149822235107424, 0.03524367904663086, 0.03523004913330078, 0.03520476913452148, 0.0354015998840332, 0.03536943817138672, 0.03537510299682617, 0.035383296966552735, 0.03547545623779297, 0.035544864654541014, 0.03561049652099609, 0.035572063446044924, 0.035495712280273435, 0.035496158599853514, 0.035520511627197264, 0.035555233001708986, 0.03563859176635742, 0.035608993530273435, 0.03579308700561523, 0.03583347320556641, 0.035762752532958984, 0.035773696899414065, 0.03579366302490234, 0.03586374282836914, 0.03590841674804687, 0.035792064666748044, 0.03584236907958984, 0.03573113632202148, 0.03580601501464844, 0.03592192077636719, 0.03595395278930664, 0.035976959228515626, 0.03595734405517578, 0.03588768005371094, 0.036031806945800784, 0.03600230407714844, 0.036128799438476564, 0.03609196853637695, 0.03613180923461914, 0.036035488128662106, 0.036020030975341795, 0.03603478240966797, 0.03609135818481445, 0.03608425521850586, 0.03611856079101562, 0.03573161697387695, 0.035061153411865234, 0.034914752960205075, 0.03488083267211914, 0.03489247894287109, 0.03491430282592774, 0.03507820892333984, 0.03523366546630859, 0.035498046875, 0.0352911376953125, 0.03536627197265625, 0.03595737457275391, 0.03553279876708984, 0.035272705078125, 0.03527280044555664, 0.0352639045715332, 0.035315841674804685, 0.035314048767089844, 0.03534774398803711, 0.035270977020263675, 0.03541158294677734, 0.03543734359741211, 0.03546316909790039, 0.03539904022216797, 0.03550067138671875, 0.035518657684326174, 0.03560432052612305, 0.03559408187866211, 0.03596915054321289, 0.03554899215698242, 0.035459262847900394, 0.03546656036376953, 0.035599040985107425, 0.035676158905029294, 0.03568374252319336, 0.03562268829345703, 0.03565856170654297, 0.035776512145996094, 0.035781761169433594, 0.035746623992919925, 0.0358031997680664, 0.03574288177490234, 0.03572848129272461, 0.03573859024047851, 0.035727745056152345, 0.03569295883178711, 0.03586777496337891, 0.035865215301513674, 0.035875072479248045, 0.03588083267211914, 0.035932449340820315, 0.03587836837768555, 0.035869247436523435, 0.0359381103515625, 0.03597875213623047, 0.03595929718017578, 0.03590758514404297, 0.03587276840209961, 0.035985408782958986, 0.0359628791809082, 0.03590956878662109, 0.03603462219238281, 0.03607513427734375, 0.03585244750976563, 0.035157089233398435, 0.03495769500732422, 0.03497315216064453, 0.03508303833007813, 0.035135616302490236, 0.03511641693115235, 0.03512911987304688, 0.0353267822265625, 0.035291168212890626, 0.035196769714355466, 0.03520937728881836, 0.035249473571777344, 0.0352749137878418, 0.035314208984375, 0.035272350311279295, 0.035324256896972654, 0.035209217071533204, 0.03520512008666992, 0.03529654312133789, 0.03527753448486328, 0.035362495422363284, 0.0355140495300293, 0.035377792358398434, 0.03546316909790039, 0.03537299346923828, 0.03551852798461914, 0.03550592041015625, 0.035635456085205075, 0.035620864868164064, 0.035640670776367185, 0.03550255966186523, 0.03556726455688477, 0.035566078186035154, 0.035649566650390624, 0.03568435287475586, 0.03580332946777344, 0.0357743034362793, 0.03579804611206055, 0.03579747009277344, 0.03576607894897461, 0.03595481491088867, 0.0359714241027832, 0.03579923248291016, 0.03578675079345703, 0.03578675079345703, 0.03577971267700195, 0.03575900650024414, 0.03589673614501953, 0.03590361785888672, 0.036001598358154294, 0.036006366729736325, 0.0360654411315918, 0.03601408004760742, 0.03605052947998047, 0.03602793502807617, 0.03596992111206055, 0.03590758514404297, 0.03590553665161133, 0.03586867141723633, 0.03600953674316406, 0.03599200057983398, 0.03596646499633789]",tokens/s,28.10870372455261,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Qwen/Qwen-72B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Qwen/Qwen-72B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,11035.037696,14897.053696,0.0,14501.80608,13634.065408,s,1,18.76700390625,18.76700390625,0.0,18.76700390625,18.76700390625,18.76700390625,18.76700390625,[18.76700390625],,kWh,0.0003418410694708503,3.769941840282371e-05,0.00015415567887998305,0.0005336961667536571,,MB,2132.942848,14911.73376,0.0,14501.80608,12898.830848,s,10,6.400340270996093,0.6400340270996093,0.000510936811757298,0.6400107116699219,0.6404652282714843,0.6408448883056641,0.6411486163330078,"[0.640380859375, 0.640076904296875, 0.6392645874023437, 0.6397262573242187, 0.6396455078125, 0.640072998046875, 0.6412245483398438, 0.6403334350585937, 0.639666748046875, 0.6399484252929688]",tokens/s,399.9787341933906,kWh,1.8703781793228794e-05,2.0617305319390656e-06,1.2342978624375368e-05,3.310849094954323e-05,tokens/kWh,7732155.488153767,MB,2149.777408,14911.73376,0.0,14501.80608,13243.63264,s,10,36.24666357421875,3.624666357421875,0.0019243924507422602,3.625056518554688,3.626326318359375,3.6270015380859375,3.6275417138671875,"[3.62180712890625, 3.6256484375, 3.624166748046875, 3.6247294921875, 3.62556982421875, 3.624913818359375, 3.62617626953125, 3.62519921875, 3.62077587890625, 3.6276767578125]",tokens/s,17.380910072178384,kWh,0.0001057513950405222,1.1664895981548868e-05,7.05961988380235e-05,0.00018801248986009461,tokens/kWh,335084.12152235245,,s,630,36.24219458389281,0.05752729299030606,0.0005091918130798269,0.05749966430664062,0.05819214363098144,0.0583145601272583,0.05850669719696045,"[0.057396446228027344, 0.056642143249511716, 0.05666396713256836, 0.056575904846191405, 0.0567353286743164, 0.05672604751586914, 0.05664588928222656, 0.05652252960205078, 0.05687078475952148, 0.05687945556640625, 0.05689548873901367, 0.05690761566162109, 0.05681987380981445, 0.056798912048339846, 0.05684659194946289, 0.05698976135253906, 0.05698713684082031, 0.05707417678833008, 0.057112575531005856, 0.057024513244628906, 0.05711052703857422, 0.05706547164916992, 0.056973312377929686, 0.057124671936035154, 0.057254016876220705, 0.057314910888671876, 0.057395294189453126, 0.05736486434936523, 0.05748876953125, 0.05740771102905273, 0.05740585708618164, 0.057296897888183596, 0.05749350357055664, 0.05741292953491211, 0.05740409469604492, 0.05775779342651367, 0.05767113494873047, 0.057668033599853515, 0.057659393310546876, 0.057640960693359375, 0.05781708908081055, 0.05778841781616211, 0.05784783935546875, 0.057866207122802736, 0.057945537567138675, 0.05799168014526367, 0.05812752151489258, 0.05810179138183594, 0.05816153717041016, 0.05792409515380859, 0.05802793502807617, 0.05830460739135742, 0.05822259140014648, 0.05820620727539062, 0.05825273513793945, 0.05805318450927734, 0.05823283386230469, 0.05813862228393555, 0.05809971237182617, 0.05818527984619141, 0.058220993041992186, 0.058355712890625, 0.058365951538085936, 0.05756124877929687, 0.05680704116821289, 0.05668476867675781, 0.05650838470458985, 0.05690140914916992, 0.056957054138183597, 0.05675203323364258, 0.056703201293945314, 0.05694464111328125, 0.05726163101196289, 0.057270015716552734, 0.05722127914428711, 0.05710492706298828, 0.057215999603271485, 0.0573897590637207, 0.05743648147583008, 0.05748121643066406, 0.05725183868408203, 0.05706137466430664, 0.05708390426635742, 0.05700198364257812, 0.056997886657714845, 0.05716377639770508, 0.05706489562988281, 0.05735391998291016, 0.05735513687133789, 0.05737881469726563, 0.05749964904785156, 0.05773305511474609, 0.05764432144165039, 0.05763151931762695, 0.05765929412841797, 0.05775779342651367, 0.05783667373657227, 0.057736064910888674, 0.05797868728637695, 0.05782956695556641, 0.05758156967163086, 0.05753606414794922, 0.057600448608398434, 0.05777203369140625, 0.057679039001464844, 0.05781177520751953, 0.05767168045043945, 0.05774131011962891, 0.05779865646362305, 0.057773086547851564, 0.05781132888793945, 0.05774358367919922, 0.057903488159179686, 0.0578785285949707, 0.05784175872802735, 0.05791254425048828, 0.05816403198242188, 0.058138111114501956, 0.05809347152709961, 0.05815491104125976, 0.05809177780151367, 0.058232864379882815, 0.05839094543457031, 0.05826067352294922, 0.05821305465698242, 0.05826355361938477, 0.05762038421630859, 0.05673177719116211, 0.056608543395996094, 0.05649001693725586, 0.056704513549804686, 0.056711006164550784, 0.05681375885009766, 0.056939136505126955, 0.057100223541259765, 0.057173152923583985, 0.05691187286376953, 0.05700201416015625, 0.057008575439453125, 0.056984001159667966, 0.05693030548095703, 0.05702656173706055, 0.05718220901489258, 0.05736198425292969, 0.0571847038269043, 0.05716377639770508, 0.057060798645019534, 0.05707219314575195, 0.057006080627441405, 0.05703475189208984, 0.05706467056274414, 0.05721987152099609, 0.05737472152709961, 0.05730416107177734, 0.057265056610107425, 0.05740544128417969, 0.05737062454223633, 0.05732966232299805, 0.05743404769897461, 0.05737180709838867, 0.05744323348999023, 0.057591808319091796, 0.05758566284179688, 0.057501697540283205, 0.057575424194335936, 0.05769833755493164, 0.05792559814453125, 0.05795993423461914, 0.05817804718017578, 0.05816115188598633, 0.05801369476318359, 0.058071041107177736, 0.05809151840209961, 0.05811609649658203, 0.058141792297363284, 0.05813702392578125, 0.058055137634277346, 0.05820415878295898, 0.05808137512207031, 0.05802403259277344, 0.058005214691162106, 0.05790729522705078, 0.058103809356689455, 0.05804646301269531, 0.05827139282226562, 0.05847894287109375, 0.05849468612670899, 0.05839503860473633, 0.05848614501953125, 0.05733583831787109, 0.056799198150634764, 0.05689548873901367, 0.0566640625, 0.05678079986572265, 0.05696428680419922, 0.056767295837402344, 0.05671321487426758, 0.0567562255859375, 0.05692006301879883, 0.05684633636474609, 0.05705740737915039, 0.057076961517333984, 0.05726883316040039, 0.05714540863037109, 0.0571671371459961, 0.057197025299072266, 0.05715135955810547, 0.05709862518310547, 0.05740339279174805, 0.05719190216064453, 0.05713359832763672, 0.05717187118530274, 0.05739324951171875, 0.05731107330322266, 0.05744655990600586, 0.05743001556396484, 0.057425918579101565, 0.05747420883178711, 0.05751692962646485, 0.05749084854125976, 0.05748336029052734, 0.05752870559692383, 0.057493598937988284, 0.05747097778320313, 0.05764278411865234, 0.05760227203369141, 0.05749964904785156, 0.05785411071777344, 0.05766128158569336, 0.057643009185791017, 0.05767686462402344, 0.057753662109375, 0.057758590698242185, 0.05778636932373047, 0.05773926544189453, 0.057908927917480466, 0.05805702209472656, 0.05809766387939453, 0.05808892822265625, 0.058040863037109376, 0.05802361679077148, 0.057962814331054685, 0.058068992614746094, 0.05822387313842774, 0.05813529586791992, 0.058120159149169924, 0.05803830337524414, 0.05817958450317383, 0.05834515380859375, 0.05834988784790039, 0.05848608016967773, 0.058563262939453124, 0.057452545166015626, 0.05680105590820313, 0.056602272033691406, 0.056603199005126954, 0.05655276870727539, 0.05661721420288086, 0.05667011260986328, 0.05662511825561523, 0.056887550354003905, 0.056926528930664064, 0.05678694534301758, 0.05678672027587891, 0.05676873779296875, 0.05694668960571289, 0.05701827239990234, 0.05696112060546875, 0.057106433868408205, 0.05728015899658203, 0.057289249420166014, 0.05738476943969727, 0.05736038589477539, 0.05732352066040039, 0.057181953430175785, 0.05719065475463867, 0.05737859344482422, 0.05732694244384766, 0.057471263885498045, 0.05737123107910156, 0.057427967071533206, 0.057450496673583984, 0.057484928131103515, 0.0575200309753418, 0.05791587066650391, 0.05785599899291992, 0.05768396759033203, 0.057745407104492184, 0.057692161560058595, 0.057665534973144535, 0.057819168090820314, 0.0577371826171875, 0.05769772720336914, 0.05788934326171875, 0.0580055046081543, 0.05797478485107422, 0.05792144012451172, 0.05808547210693359, 0.05810790252685547, 0.05806284713745117, 0.05804032135009766, 0.058087425231933595, 0.05816320037841797, 0.058028030395507815, 0.057974945068359374, 0.058007392883300785, 0.05806204986572266, 0.05812918472290039, 0.058267520904541015, 0.05824313735961914, 0.058187488555908204, 0.058388126373291015, 0.05837894439697266, 0.05845372772216797, 0.05835776138305664, 0.057401344299316405, 0.0566558723449707, 0.056570945739746095, 0.05650527954101563, 0.05667839813232422, 0.05671094512939453, 0.05666838455200195, 0.05667795181274414, 0.05695052719116211, 0.05702067184448242, 0.057135425567626956, 0.05714956665039062, 0.05704294586181641, 0.05697052764892578, 0.05707440185546875, 0.05714739227294922, 0.057079105377197265, 0.05712966537475586, 0.057071872711181644, 0.05701919937133789, 0.057293758392333985, 0.05734604644775391, 0.05730012893676758, 0.05737673568725586, 0.05740019226074219, 0.057308513641357424, 0.05752284622192383, 0.05749308776855469, 0.057444766998291014, 0.057507839202880856, 0.057353374481201175, 0.057485279083251954, 0.05755311965942383, 0.057520576477050785, 0.057718528747558596, 0.057702880859375, 0.057621726989746096, 0.05771548843383789, 0.05779865646362305, 0.05773516845703125, 0.05784985733032227, 0.05773311996459961, 0.05777622222900391, 0.05778979110717773, 0.057939903259277344, 0.057813377380371095, 0.05779891204833985, 0.057896831512451175, 0.058048576354980466, 0.05817049789428711, 0.0579898567199707, 0.0580425910949707, 0.05790841674804687, 0.057996097564697265, 0.0579317741394043, 0.058054656982421876, 0.05833692932128906, 0.058208606719970704, 0.058191871643066405, 0.058431488037109375, 0.05856025695800781, 0.05850547027587891, 0.05854003143310547, 0.05740614318847656, 0.05675151824951172, 0.056648414611816404, 0.05658828735351563, 0.056704864501953126, 0.05674403381347656, 0.05707574462890625, 0.056938785552978516, 0.05699760055541992, 0.057086078643798825, 0.05690153503417969, 0.05709852981567383, 0.0573226547241211, 0.05755551910400391, 0.056952640533447264, 0.056926399230957034, 0.05701033782958984, 0.05732115173339844, 0.057333919525146486, 0.057236671447753906, 0.057215808868408206, 0.05728665542602539, 0.0572083511352539, 0.0572523193359375, 0.057606273651123044, 0.05738489532470703, 0.057391040802001955, 0.05728460693359375, 0.05736243057250977, 0.057565185546875, 0.05768806457519531, 0.0574213752746582, 0.057395648956298825, 0.05740544128417969, 0.05751603317260742, 0.057513568878173826, 0.05753283309936524, 0.05756694412231445, 0.05751631927490235, 0.057474239349365235, 0.05759673690795898, 0.057765151977539064, 0.05773731231689453, 0.05770207977294922, 0.05804048156738281, 0.05796726226806641, 0.057892990112304685, 0.0579317741394043, 0.05815500640869141, 0.05810940933227539, 0.05834121704101562, 0.058337535858154294, 0.058116222381591795, 0.0581286735534668, 0.058009056091308596, 0.058104385375976564, 0.058264896392822264, 0.05815372848510742, 0.058124225616455076, 0.05845590209960937, 0.05834915161132812, 0.05853241729736328, 0.058705921173095706, 0.05755644989013672, 0.05683705520629883, 0.05670902252197266, 0.05684838485717773, 0.056888896942138674, 0.05684415817260742, 0.056785472869873045, 0.056923423767089844, 0.05698015975952148, 0.05695695877075195, 0.05692195129394531, 0.05686492919921875, 0.05694367980957031, 0.056916927337646486, 0.057054561614990236, 0.05703132629394531, 0.057006080627441405, 0.057133056640625, 0.05706752014160156, 0.05701836776733398, 0.057194496154785154, 0.05718220901489258, 0.057134208679199217, 0.05710470581054688, 0.05725830459594727, 0.0572578239440918, 0.05745043182373047, 0.05751753616333008, 0.05732668685913086, 0.05730652618408203, 0.05723392105102539, 0.05736816024780273, 0.05739916610717773, 0.05757606506347656, 0.05773455810546875, 0.05786995315551758, 0.05770556640625, 0.0577042236328125, 0.057788063049316406, 0.05792124938964844, 0.057847713470458986, 0.05785468673706055, 0.05781462478637695, 0.057876895904541016, 0.057987071990966796, 0.057896961212158204, 0.05791129684448242, 0.0579788818359375, 0.05793920135498047, 0.05805110549926758, 0.058275360107421875, 0.05815740966796875, 0.05817808151245117, 0.05822009658813477, 0.05822079849243164, 0.05813407897949219, 0.05832134246826172, 0.05815849685668945, 0.05814707183837891, 0.05816953659057617, 0.05830627059936523, 0.05849542236328125, 0.058535743713378906, 0.05749967956542969, 0.05670406341552734, 0.056548255920410156, 0.05652275085449219, 0.056758014678955075, 0.05678515243530274, 0.05669411087036133, 0.05672412872314453, 0.05810966491699219, 0.05670694351196289, 0.05678531265258789, 0.057012222290039063, 0.05699174499511719, 0.05695078277587891, 0.05701017761230469, 0.05700771331787109, 0.057158046722412106, 0.05723244857788086, 0.05701670455932617, 0.05691587066650391, 0.05711529541015625, 0.05710831832885742, 0.05713116836547852, 0.05729718399047851, 0.057242942810058595, 0.057262561798095704, 0.05736441421508789, 0.0573493766784668, 0.057406208038330075, 0.057390079498291016, 0.057322494506835936, 0.05737267303466797, 0.05766105651855469, 0.05752617645263672, 0.057600479125976566, 0.057616382598876956, 0.05756073760986328, 0.05742364883422851, 0.05758627319335938, 0.05777814483642578, 0.05784928131103516, 0.05781356811523437, 0.05772252655029297, 0.05772457504272461, 0.05775843048095703, 0.0577658576965332, 0.05781897735595703, 0.057845664978027345, 0.057897216796875, 0.05800252914428711, 0.058022815704345705, 0.057987071990966796, 0.057984928131103515, 0.057962017059326174, 0.05796124649047851, 0.05788425445556641, 0.05813267135620117, 0.0581341438293457, 0.05810214233398438, 0.05815606307983399, 0.05814575958251953, 0.058210304260253906, 0.058218494415283206, 0.05731497573852539, 0.056812065124511715, 0.05687289428710938, 0.056721214294433595, 0.056735488891601564, 0.056901695251464844, 0.05692681503295898, 0.057079841613769534, 0.057003902435302733, 0.05693423843383789, 0.05721116638183594, 0.05718425750732422, 0.05704499053955078, 0.05699971389770508, 0.05698892974853516, 0.05711356735229492, 0.05715558242797852, 0.05722735977172851, 0.05758761596679687, 0.05724700927734375, 0.05725462341308594, 0.05715126419067383, 0.05720905685424805, 0.05728979110717773, 0.0573513298034668, 0.05756086349487305, 0.05767955017089844, 0.05744262313842773, 0.05732681655883789, 0.057477249145507815, 0.05749593734741211, 0.05738528060913086, 0.057460575103759765, 0.05745577621459961, 0.057598945617675784, 0.05774131011962891, 0.05764710235595703, 0.05757132720947265, 0.057708545684814455, 0.05779455947875976, 0.05781913757324219, 0.057915393829345706, 0.05792153549194336, 0.05781094360351562, 0.05809356689453125, 0.05798912048339844, 0.05793564987182617, 0.05783564758300781, 0.057853374481201175, 0.05788467025756836, 0.05827651214599609, 0.05814470291137695, 0.05820745468139649, 0.058261791229248044, 0.05822118377685547, 0.058253246307373045, 0.05817913436889648, 0.058194591522216794, 0.05822000122070312, 0.05834511947631836, 0.058284000396728514, 0.05842160034179687, 0.05850719833374023]",tokens/s,17.38305329556373,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.747328,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3551.899648,5264.83456,0.0,4869.586944,4520.068608,s,1,11.1682099609375,11.1682099609375,0.0,11.1682099609375,11.1682099609375,11.1682099609375,11.1682099609375,[11.1682099609375],,kWh,0.00012159290330833225,1.3405123182575374e-05,5.266587546599866e-05,0.00018766390195690627,,MB,1475.104768,5298.388992,0.0,4888.461312,4194.016256,s,10,1.7663264923095703,0.17663264923095703,0.00017453915418906213,0.17667317962646484,0.17684921112060545,0.17685737533569335,0.17686390670776367,"[0.17672930908203124, 0.1763408966064453, 0.17639692687988281, 0.17686553955078124, 0.1767637481689453, 0.17661705017089843, 0.17673196411132813, 0.17650860595703124, 0.1765250549316406, 0.17684739685058593]",tokens/s,1449.3356755650861,kWh,5.215853408333378e-06,5.752153263915551e-07,3.445940256750049e-06,9.237008991474982e-06,tokens/kWh,27714598.98288152,MB,1483.50976,5306.7776,0.0,4896.84992,4194.018816,s,10,16.84819104003906,1.6848191040039062,0.0057024234270836435,1.6827443847656252,1.6934502075195312,1.6938809265136718,1.6942255017089842,"[1.68153759765625, 1.68524853515625, 1.6789527587890625, 1.6933544921875, 1.679438720703125, 1.680484130859375, 1.691267333984375, 1.683951171875, 1.6943116455078124, 1.6796446533203124]",tokens/s,37.39273839564318,kWh,4.8736988856666766e-05,5.3754525794993705e-06,3.1029559545849735e-05,8.514200098201587e-05,tokens/kWh,739940.3264354473,,s,630,16.84579932975769,0.0267393640154884,0.0005493166513169489,0.02661803150177002,0.0270858341217041,0.027548237800598143,0.028608866176605227,"[0.027618944168090822, 0.02668582344055176, 0.026681312561035155, 0.026505247116088867, 0.026524991989135743, 0.02651411247253418, 0.02670182418823242, 0.026804224014282226, 0.026466304779052735, 0.02630985641479492, 0.02633123207092285, 0.02650592041015625, 0.026574560165405273, 0.026399007797241213, 0.02642736053466797, 0.026434816360473633, 0.026428192138671876, 0.02641663932800293, 0.02638489532470703, 0.026550271987915038, 0.026398719787597655, 0.026558464050292968, 0.026494911193847656, 0.026512895584106445, 0.02656928062438965, 0.026658016204833983, 0.026519615173339842, 0.026512096405029297, 0.026436800003051757, 0.026411840438842774, 0.02695782470703125, 0.026563808441162108, 0.026440383911132813, 0.0285548152923584, 0.02910611152648926, 0.027166847229003907, 0.026685087203979493, 0.02676924705505371, 0.027085792541503905, 0.026681343078613282, 0.026654111862182618, 0.026775392532348632, 0.02760576057434082, 0.026798080444335938, 0.027072511672973632, 0.026619903564453123, 0.026462207794189452, 0.0265031681060791, 0.026358783721923826, 0.026612735748291014, 0.026752031326293946, 0.026889184951782226, 0.02674406433105469, 0.026700351715087892, 0.02662419128417969, 0.02652774429321289, 0.02647859191894531, 0.02650435256958008, 0.02643849563598633, 0.026458112716674805, 0.02653539276123047, 0.02637177658081055, 0.026467168807983398, 0.028073951721191405, 0.027975839614868166, 0.028180032730102538, 0.026595775604248046, 0.02666700744628906, 0.02650726318359375, 0.02709065628051758, 0.02666249656677246, 0.02671891212463379, 0.026431488037109374, 0.026650623321533205, 0.0265765438079834, 0.02691312026977539, 0.026400768280029296, 0.02664816093444824, 0.02673196792602539, 0.0266331844329834, 0.026818208694458008, 0.026674976348876955, 0.026847679138183592, 0.02668351936340332, 0.026674911499023436, 0.026835231781005858, 0.02697417640686035, 0.026966047286987305, 0.02691276741027832, 0.026652671813964843, 0.026705919265747072, 0.026695327758789064, 0.026757471084594725, 0.026658815383911134, 0.026472448348999023, 0.026390527725219725, 0.026451967239379884, 0.026607616424560547, 0.02714022445678711, 0.026566335678100586, 0.026749088287353517, 0.02718921661376953, 0.02644780731201172, 0.026386367797851563, 0.02637433624267578, 0.026605567932128905, 0.027000831604003905, 0.026671104431152344, 0.026627904891967775, 0.026484928131103515, 0.02654412841796875, 0.026746591567993163, 0.026990304946899413, 0.026886528015136718, 0.026663103103637696, 0.026899871826171876, 0.026919519424438477, 0.026822240829467773, 0.026513824462890623, 0.026560480117797852, 0.026402847290039062, 0.02639664077758789, 0.02649295997619629, 0.026434688568115233, 0.02650815963745117, 0.02672230339050293, 0.027522880554199217, 0.026837024688720703, 0.02672630310058594, 0.026400447845458985, 0.026398464202880858, 0.026433536529541016, 0.027437952041625975, 0.027551679611206054, 0.02641312026977539, 0.026455232620239258, 0.026515520095825196, 0.02683545684814453, 0.026384288787841798, 0.026501472473144532, 0.02647596740722656, 0.026552831649780274, 0.026429088592529296, 0.026941471099853516, 0.0263822078704834, 0.02645622444152832, 0.026435935974121094, 0.026556415557861326, 0.02660963249206543, 0.026454048156738283, 0.026318431854248047, 0.026456192016601564, 0.02640105628967285, 0.026427040100097655, 0.026241376876831056, 0.026372095108032227, 0.026216064453125, 0.026351295471191406, 0.026219072341918944, 0.02667532730102539, 0.026888191223144533, 0.026927104949951174, 0.026844800949096678, 0.02664486312866211, 0.026613407135009766, 0.026415456771850587, 0.026543935775756835, 0.026538143157958983, 0.02645199966430664, 0.02636390495300293, 0.026597375869750976, 0.026480640411376953, 0.0263372802734375, 0.026259456634521484, 0.02650307273864746, 0.027050079345703124, 0.026802047729492188, 0.028313728332519533, 0.028058847427368163, 0.026680095672607422, 0.02671001625061035, 0.026975456237792968, 0.026639135360717773, 0.0266092472076416, 0.026618207931518555, 0.026697792053222657, 0.026681343078613282, 0.026609567642211913, 0.026488319396972656, 0.027512575149536134, 0.02698854446411133, 0.02675884819030762, 0.026687807083129882, 0.026871135711669922, 0.028630943298339845, 0.027058528900146483, 0.026816991806030272, 0.0270533447265625, 0.026924800872802735, 0.02686649513244629, 0.02757232093811035, 0.026943168640136718, 0.02701958465576172, 0.026796031951904296, 0.026613088607788087, 0.02691337585449219, 0.02695721626281738, 0.027054048538208007, 0.02694828796386719, 0.026884096145629883, 0.02678771209716797, 0.027073663711547853, 0.02673971176147461, 0.026671104431152344, 0.02674892807006836, 0.026674623489379882, 0.026554943084716797, 0.026609376907348634, 0.02981475257873535, 0.027050304412841796, 0.026682783126831054, 0.026774112701416015, 0.0265031681060791, 0.02685087966918945, 0.02697056007385254, 0.02706800079345703, 0.02715484809875488, 0.026658815383911134, 0.02675916862487793, 0.026756704330444334, 0.026420736312866212, 0.02660153579711914, 0.026643295288085938, 0.026473888397216795, 0.026993247985839845, 0.02656870460510254, 0.02633932876586914, 0.027229728698730468, 0.026581472396850585, 0.026401792526245117, 0.026266624450683593, 0.026343423843383788, 0.02691833686828613, 0.02664041519165039, 0.0269333438873291, 0.027288000106811525, 0.026611711502075194, 0.02670182418823242, 0.0265482234954834, 0.026957504272460936, 0.026413375854492188, 0.026482688903808595, 0.026839456558227538, 0.026427328109741213, 0.026478208541870118, 0.02632761573791504, 0.02639030456542969, 0.02634979248046875, 0.02633318328857422, 0.026294271469116212, 0.026425344467163086, 0.026564607620239256, 0.02672230339050293, 0.026470111846923827, 0.026431072235107423, 0.027030208587646484, 0.02672230339050293, 0.03003948783874512, 0.027544031143188475, 0.026929248809814454, 0.02643494415283203, 0.02641164779663086, 0.026443775177001954, 0.02691481590270996, 0.026503103256225586, 0.02642336082458496, 0.02651955223083496, 0.026451967239379884, 0.026558464050292968, 0.02676736068725586, 0.02721177673339844, 0.026492927551269533, 0.026462207794189452, 0.02637740707397461, 0.026434015274047852, 0.026445600509643556, 0.02691539192199707, 0.02677497673034668, 0.02652627182006836, 0.02656051254272461, 0.026526784896850585, 0.026617855072021485, 0.02647750473022461, 0.026467391967773438, 0.02655740737915039, 0.026511327743530273, 0.02647769546508789, 0.02658188819885254, 0.0263701114654541, 0.026478143692016603, 0.02645644760131836, 0.02641001510620117, 0.026917055130004884, 0.026850080490112303, 0.026516639709472656, 0.026781984329223633, 0.027480640411376954, 0.026533119201660155, 0.0264749755859375, 0.026517791748046873, 0.026832895278930666, 0.02692095947265625, 0.026556415557861326, 0.02655436706542969, 0.026396543502807616, 0.027586271286010742, 0.026786144256591798, 0.026727584838867186, 0.02662633514404297, 0.026575584411621094, 0.026558464050292968, 0.026512863159179688, 0.02676380729675293, 0.026884096145629883, 0.026845184326171875, 0.026684768676757814, 0.026755744934082032, 0.026839040756225587, 0.026742624282836913, 0.026590911865234376, 0.026544607162475586, 0.026445823669433592, 0.026445632934570314, 0.02662188720703125, 0.026444032669067384, 0.026388032913208008, 0.026765535354614258, 0.026811744689941405, 0.02648931121826172, 0.026495391845703126, 0.026341087341308595, 0.02642099189758301, 0.026761760711669923, 0.02654742431640625, 0.026823455810546876, 0.026556415557861326, 0.02672790336608887, 0.02657539176940918, 0.02722105598449707, 0.026657312393188477, 0.02674492835998535, 0.026851200103759767, 0.026806175231933595, 0.02674127960205078, 0.026626047134399415, 0.026502592086791992, 0.026362432479858398, 0.026430944442749023, 0.026636768341064453, 0.02649075126647949, 0.026501312255859374, 0.026410144805908205, 0.026325855255126953, 0.02659724807739258, 0.026634368896484375, 0.026570751190185548, 0.026630144119262695, 0.02615910339355469, 0.02642323112487793, 0.026558528900146483, 0.026372095108032227, 0.02712335968017578, 0.026552671432495116, 0.026791263580322265, 0.02778531265258789, 0.027348608016967774, 0.026999359130859376, 0.026700031280517577, 0.027456096649169922, 0.027567167282104493, 0.02659347152709961, 0.026501855850219726, 0.026595327377319337, 0.026605567932128905, 0.026646272659301758, 0.026418912887573243, 0.02632143974304199, 0.026408960342407226, 0.026462207794189452, 0.026471839904785157, 0.02647920036315918, 0.026755071640014647, 0.026541471481323242, 0.026857280731201173, 0.026780448913574218, 0.026673152923583986, 0.02672198486328125, 0.026827072143554686, 0.026563615798950196, 0.026637279510498046, 0.02973654365539551, 0.02759516716003418, 0.026998207092285155, 0.02673721694946289, 0.02669158363342285, 0.027451391220092772, 0.028436479568481447, 0.02668339157104492, 0.02671615982055664, 0.026805376052856444, 0.02758902359008789, 0.026834495544433595, 0.02701510429382324, 0.027048927307128906, 0.02697420883178711, 0.026842239379882813, 0.027706239700317385, 0.02671615982055664, 0.02675712013244629, 0.02656870460510254, 0.026648576736450196, 0.02674483108520508, 0.0265645751953125, 0.026530975341796874, 0.026552448272705077, 0.026599903106689454, 0.02671843147277832, 0.026880064010620118, 0.026779647827148437, 0.027191200256347657, 0.027273311614990234, 0.026463775634765624, 0.026431936264038086, 0.02650691223144531, 0.026345888137817384, 0.026445472717285156, 0.02653139114379883, 0.0267640323638916, 0.026836992263793946, 0.026848447799682616, 0.026573600769042968, 0.027165792465209962, 0.02655462455749512, 0.026552736282348634, 0.026478784561157227, 0.02647859191894531, 0.026436992645263672, 0.027634016036987306, 0.026972448348999024, 0.026496192932128907, 0.026720863342285156, 0.026715808868408204, 0.026863552093505858, 0.026827360153198244, 0.026744863510131837, 0.026629791259765626, 0.026614112854003905, 0.02660313606262207, 0.02644416046142578, 0.026642431259155275, 0.02654991912841797, 0.026433887481689452, 0.026445823669433592, 0.026531679153442383, 0.026646688461303712, 0.026729631423950195, 0.026667871475219727, 0.026711360931396484, 0.026576799392700197, 0.026794591903686524, 0.02685152053833008, 0.02708620834350586, 0.026921600341796876, 0.027862815856933593, 0.02673072052001953, 0.026985824584960936, 0.026821279525756837, 0.026685440063476562, 0.027033599853515625, 0.02675916862487793, 0.026847232818603517, 0.0269835205078125, 0.026997663497924804, 0.026816511154174806, 0.02752921676635742, 0.02649087905883789, 0.02668329620361328, 0.026505311965942382, 0.026597312927246094, 0.026695903778076173, 0.026613344192504884, 0.02643382453918457, 0.02639587211608887, 0.026342144012451173, 0.026445823669433592, 0.026552255630493165, 0.026740127563476563, 0.026878623962402343, 0.026666656494140625, 0.026504543304443358, 0.02652592086791992, 0.027060319900512695, 0.026546880722045897, 0.026447872161865234, 0.02873129653930664, 0.026760639190673827, 0.026585439682006835, 0.026693952560424804, 0.02676531219482422, 0.027047935485839843, 0.026944927215576172, 0.027755104064941406, 0.02793788719177246, 0.02672643280029297, 0.026495872497558595, 0.027031328201293944, 0.026505151748657228, 0.026535968780517578, 0.026520959854125975, 0.026854272842407226, 0.026816511154174806, 0.02671001625061035, 0.02653113555908203, 0.02652025604248047, 0.026644479751586913, 0.026501279830932617, 0.026500959396362305, 0.026703744888305663, 0.026816064834594727, 0.026812736511230468, 0.026761472702026366, 0.026570335388183593, 0.026583456039428712, 0.02652899169921875, 0.02650601577758789, 0.027711488723754882, 0.035465217590332034, 0.027033599853515625, 0.02670796775817871, 0.02679193687438965, 0.02654617691040039, 0.02716057586669922, 0.026706016540527344, 0.02667510414123535, 0.02672591972351074, 0.026575328826904297, 0.027076223373413085, 0.026585216522216796, 0.026429151535034178, 0.02661020851135254, 0.026410623550415038, 0.026507648468017578, 0.027133951187133788, 0.026431488037109374, 0.02644528007507324, 0.02673308753967285, 0.026557855606079102, 0.026558496475219726, 0.026655168533325196, 0.02658915138244629, 0.026494367599487305, 0.026479360580444335, 0.026466304779052735, 0.026533472061157228, 0.026476160049438476, 0.026669376373291014, 0.02670025634765625, 0.027389951705932617, 0.02717923164367676, 0.026647808074951172, 0.026702272415161134, 0.02737571144104004, 0.02672230339050293, 0.026599359512329102, 0.02662816047668457, 0.026468351364135743, 0.026527103424072267, 0.026718015670776366, 0.026626752853393554, 0.027408512115478515, 0.026408960342407226, 0.02647859191894531, 0.026663103103637696, 0.02720915222167969, 0.02659881591796875, 0.02652003288269043, 0.026465824127197266, 0.02658198356628418, 0.026617855072021485, 0.02741596794128418, 0.0266713924407959, 0.02644950485229492, 0.026249824523925783, 0.026406784057617188, 0.02625702476501465, 0.026264192581176758, 0.026349504470825194, 0.02643881607055664, 0.02645084762573242, 0.026513408660888672, 0.02627299118041992, 0.026460960388183595, 0.02652569580078125, 0.02629631996154785, 0.026320831298828125, 0.026277408599853516, 0.02659996795654297, 0.027694719314575195, 0.027961023330688478, 0.027642175674438475, 0.026812736511230468, 0.02643155288696289, 0.026425344467163086, 0.026477951049804688, 0.027073408126831056, 0.026683135986328124, 0.026410688400268556, 0.026536191940307617, 0.026431295394897462, 0.026566911697387695, 0.02701702308654785, 0.02671836853027344, 0.026308416366577148, 0.02651772880554199, 0.02654204750061035, 0.026488672256469725, 0.02658527946472168, 0.026605567932128905, 0.02634480094909668, 0.026388736724853517]",tokens/s,37.39804729165451,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4920.836096,8001.61792,0.0,7606.370304,6988.678144,s,1,13.69842578125,13.69842578125,0.0,13.69842578125,13.69842578125,13.69842578125,13.69842578125,[13.69842578125],,kWh,0.0001918337320374121,2.1153670764710996e-05,7.7197283980035e-05,0.00029018468678215813,,MB,2802.900992,8018.395136,0.0,7608.467456,6915.136,s,10,2.203863525390625,0.2203863525390625,0.001473780131591768,0.2205274887084961,0.22190032501220702,0.2221533332824707,0.22235573989868163,"[0.21690550231933595, 0.22057408142089843, 0.22184410095214843, 0.22048089599609375, 0.22132266235351564, 0.2190339813232422, 0.22027871704101562, 0.21995756530761718, 0.22105967712402344, 0.22240634155273437]",tokens/s,1161.596428502192,kWh,6.414966657065093e-06,7.074569528675595e-07,4.248040838044473e-06,1.1370464447977125e-05,tokens/kWh,22514471.697375912,MB,2807.226368,8020.492288,0.0,7610.564608,6915.13856,s,10,19.13417175292969,1.913417175292969,0.013176174665317692,1.9097753295898436,1.9265622436523437,1.9365471130371092,1.9445350085449218,"[1.9243433837890624, 1.9071304931640625, 1.8996845703125, 1.9014300537109374, 1.917813720703125, 1.946531982421875, 1.906928955078125, 1.904153076171875, 1.9137353515625, 1.912420166015625]",tokens/s,32.9253864831405,kWh,5.557457008085287e-05,6.1297175610456665e-06,3.68125475659516e-05,9.851683520785015e-05,tokens/kWh,639484.6105955701,,s,630,19.131715995788586,0.030367803167918372,0.000670981992929344,0.030225759506225587,0.030812736701965333,0.031045494365692136,0.032240393447875985,"[0.030822303771972655, 0.03043654441833496, 0.03013724708557129, 0.030142463684082032, 0.03018547248840332, 0.03038617515563965, 0.030471328735351563, 0.030905088424682616, 0.030527584075927733, 0.030470144271850585, 0.030658432006835937, 0.03050111961364746, 0.03050864028930664, 0.03044528007507324, 0.030552032470703126, 0.03032678413391113, 0.03029052734375, 0.030351200103759766, 0.03053526306152344, 0.030386751174926757, 0.03100876808166504, 0.03033228874206543, 0.03027008056640625, 0.030711904525756836, 0.030354848861694338, 0.031936704635620115, 0.03046227264404297, 0.030305791854858398, 0.030296064376831053, 0.03053004837036133, 0.030698720932006835, 0.03054217529296875, 0.030609855651855467, 0.03060326385498047, 0.030676992416381835, 0.030664575576782226, 0.030681503295898437, 0.030602079391479492, 0.030694271087646486, 0.030658367156982422, 0.030623935699462892, 0.03056003189086914, 0.030381599426269532, 0.030229183197021486, 0.030265344619750976, 0.03033907127380371, 0.030840831756591795, 0.03030531120300293, 0.030247007369995117, 0.030179519653320313, 0.030460607528686522, 0.0336814079284668, 0.03030611228942871, 0.030324672698974608, 0.030245119094848633, 0.030273536682128906, 0.03040870475769043, 0.03037558364868164, 0.0304051513671875, 0.030496063232421874, 0.030406976699829103, 0.030563520431518554, 0.030473215103149414, 0.03134864044189453, 0.031150655746459963, 0.030840032577514647, 0.030401311874389648, 0.032004096984863284, 0.03170211219787598, 0.03014678382873535, 0.030212799072265626, 0.02991916847229004, 0.030052000045776368, 0.030179264068603516, 0.030453567504882813, 0.03029212760925293, 0.030185983657836913, 0.030187519073486328, 0.02993356704711914, 0.02998271942138672, 0.030054399490356445, 0.02997657585144043, 0.030007295608520508, 0.030111743927001954, 0.03015670394897461, 0.030107744216918947, 0.030089216232299806, 0.030104896545410157, 0.030039039611816407, 0.03012076759338379, 0.03020275115966797, 0.029936927795410156, 0.029952735900878907, 0.030320640563964843, 0.030130176544189452, 0.030267391204833984, 0.03035545539855957, 0.030535104751586915, 0.030408863067626954, 0.030347679138183595, 0.030598239898681642, 0.03032566452026367, 0.03034726333618164, 0.030029504776000977, 0.03002809524536133, 0.03002947235107422, 0.030077280044555663, 0.030615007400512696, 0.030080928802490234, 0.030038272857666016, 0.03013875198364258, 0.030007423400878905, 0.030103424072265624, 0.03045964813232422, 0.0302957763671875, 0.03047478485107422, 0.03042665672302246, 0.03028611183166504, 0.03000339126586914, 0.03005232048034668, 0.030127168655395508, 0.03006287956237793, 0.02998159980773926, 0.030258975982666015, 0.02994790458679199, 0.029879711151123048, 0.031131391525268556, 0.030425216674804686, 0.030215456008911134, 0.03011452865600586, 0.029929216384887696, 0.029819135665893556, 0.030179168701171873, 0.030058656692504883, 0.02991231918334961, 0.03005516815185547, 0.03016067123413086, 0.030159072875976564, 0.029968095779418946, 0.030052255630493165, 0.030148223876953126, 0.029971200942993163, 0.030078975677490235, 0.03034726333618164, 0.030344287872314454, 0.030411327362060547, 0.030610944747924803, 0.030232831954956053, 0.030345312118530275, 0.030038015365600586, 0.03005695915222168, 0.02993152046203613, 0.029970432281494142, 0.030003103256225586, 0.030068479537963867, 0.03006230354309082, 0.029997695922851564, 0.030048255920410157, 0.02997452735900879, 0.03002732849121094, 0.030208032608032228, 0.030366111755371093, 0.030148319244384766, 0.03025644874572754, 0.030374879837036132, 0.03029750442504883, 0.030392927169799806, 0.03033497619628906, 0.03036774444580078, 0.030229663848876952, 0.030141120910644532, 0.030277088165283204, 0.030218591690063478, 0.030112096786499024, 0.030066816329956055, 0.030035839080810547, 0.030034975051879884, 0.03001852798461914, 0.029945632934570313, 0.0299704647064209, 0.03008492851257324, 0.030103935241699218, 0.030168479919433593, 0.030114400863647462, 0.030121503829956056, 0.030013216018676757, 0.03009164810180664, 0.03007315254211426, 0.03002908706665039, 0.030905344009399413, 0.030178720474243165, 0.03000992012023926, 0.030005279541015624, 0.02995782470703125, 0.03002566337585449, 0.03006287956237793, 0.03003606414794922, 0.03003392028808594, 0.02990460777282715, 0.029985055923461915, 0.029917184829711913, 0.029888511657714844, 0.030041791915893554, 0.030042400360107423, 0.03010166358947754, 0.029908287048339845, 0.030048095703125, 0.029975263595581055, 0.030302207946777345, 0.02983526420593262, 0.029906944274902345, 0.02993302345275879, 0.03006719970703125, 0.02988649559020996, 0.030017440795898437, 0.03023676872253418, 0.029913087844848633, 0.03010908889770508, 0.029872896194458008, 0.02985763168334961, 0.03015475273132324, 0.030021631240844726, 0.030013439178466796, 0.03003392028808594, 0.029954048156738283, 0.029925376892089843, 0.03030201530456543, 0.029892799377441406, 0.029924896240234374, 0.029841888427734376, 0.030607072830200196, 0.02990675163269043, 0.03006483268737793, 0.03008950424194336, 0.03010576057434082, 0.03010304069519043, 0.030086944580078125, 0.030032447814941406, 0.02999203109741211, 0.03013657569885254, 0.030196384429931642, 0.030459903717041017, 0.031066112518310547, 0.03077280044555664, 0.030685632705688477, 0.030717023849487303, 0.030695903778076173, 0.030781568527221678, 0.03080633544921875, 0.031340543746948245, 0.03077939224243164, 0.030692352294921874, 0.031217439651489258, 0.03060563278198242, 0.03073843193054199, 0.030754911422729493, 0.030605215072631836, 0.030822399139404297, 0.031264192581176756, 0.030827104568481447, 0.030702880859375, 0.03062396812438965, 0.030826335906982423, 0.0307553596496582, 0.03164374351501465, 0.03384108734130859, 0.03127833557128906, 0.030849311828613283, 0.0309703369140625, 0.03058460807800293, 0.030541183471679688, 0.030448543548583985, 0.030138368606567382, 0.030086847305297853, 0.02999737548828125, 0.0300677433013916, 0.029959135055541992, 0.02997216033935547, 0.031531328201293944, 0.030031200408935546, 0.030106271743774414, 0.029908992767333983, 0.03020307159423828, 0.029993087768554687, 0.030141120910644532, 0.03056025505065918, 0.030062623977661133, 0.030039583206176758, 0.030310848236083984, 0.03010748863220215, 0.029990463256835936, 0.030257568359375, 0.03007516860961914, 0.03029596710205078, 0.02995199966430664, 0.030106912612915038, 0.029938400268554686, 0.03011075210571289, 0.030098304748535157, 0.030136415481567383, 0.030031328201293946, 0.030336736679077148, 0.03030918312072754, 0.030480384826660156, 0.03008892822265625, 0.03002191925048828, 0.030115808486938477, 0.030998592376708985, 0.030174816131591797, 0.03008473587036133, 0.030044544219970704, 0.030113920211791993, 0.030208255767822264, 0.030230783462524415, 0.03027244758605957, 0.031049728393554688, 0.030468095779418947, 0.030724096298217773, 0.030672672271728516, 0.030265567779541015, 0.03898777770996094, 0.030620800018310548, 0.030339391708374023, 0.030519872665405273, 0.030664831161499023, 0.030812032699584962, 0.03079987144470215, 0.031133792877197267, 0.03067686462402344, 0.03070476722717285, 0.030978687286376955, 0.030775583267211915, 0.030732479095458985, 0.030830400466918945, 0.03099033546447754, 0.030760959625244142, 0.030834911346435546, 0.030897439956665038, 0.03071027183532715, 0.03096575927734375, 0.030633983612060548, 0.03054182434082031, 0.030430335998535157, 0.030472415924072266, 0.039051937103271483, 0.030514335632324217, 0.030573183059692383, 0.030064863204956056, 0.03122585678100586, 0.03178700828552246, 0.030265344619750976, 0.029904895782470704, 0.029997055053710937, 0.03006035232543945, 0.030074527740478516, 0.030060096740722655, 0.03003081512451172, 0.030263296127319338, 0.032933887481689454, 0.0308570556640625, 0.03024502372741699, 0.030313983917236328, 0.030605472564697266, 0.03052169609069824, 0.03062291145324707, 0.03104031944274902, 0.030724096298217773, 0.03078144073486328, 0.030484384536743164, 0.03071753692626953, 0.03047065544128418, 0.03038412857055664, 0.030275936126708983, 0.03058243179321289, 0.030418943405151368, 0.03049990463256836, 0.03060736083984375, 0.030327743530273437, 0.031317440032958985, 0.030605472564697266, 0.030624319076538085, 0.030680767059326174, 0.030482879638671877, 0.030552032470703126, 0.03058691215515137, 0.031021055221557618, 0.03071731185913086, 0.030406879425048827, 0.030382495880126953, 0.03054515266418457, 0.030262144088745117, 0.03045568084716797, 0.03050841522216797, 0.030313087463378907, 0.030183040618896484, 0.030456192016601564, 0.03022233581542969, 0.030117887496948242, 0.029992895126342775, 0.030023712158203125, 0.030392192840576173, 0.030036127090454102, 0.030547744750976564, 0.030446048736572265, 0.030472192764282226, 0.029979711532592775, 0.029931711196899413, 0.03026812744140625, 0.03006447982788086, 0.030164703369140625, 0.02995974349975586, 0.029950464248657226, 0.0299521598815918, 0.030066272735595704, 0.02984796714782715, 0.030086463928222656, 0.02994246482849121, 0.030011392593383788, 0.030269439697265626, 0.03002572822570801, 0.02998678398132324, 0.02994588851928711, 0.030238624572753905, 0.0303985595703125, 0.030498815536499024, 0.029962240219116212, 0.030074464797973634, 0.029877792358398436, 0.03020684814453125, 0.030439424514770507, 0.030021215438842775, 0.030107200622558592, 0.029924064636230468, 0.030002399444580077, 0.030008224487304686, 0.030084352493286132, 0.030132991790771484, 0.03115007972717285, 0.03039254379272461, 0.030240192413330078, 0.03013657569885254, 0.030203392028808593, 0.030568864822387694, 0.02998067283630371, 0.030299232482910155, 0.02989148712158203, 0.030036224365234374, 0.030199552536010744, 0.030048128128051757, 0.03044950485229492, 0.03027382469177246, 0.030097408294677733, 0.029989887237548828, 0.0299550724029541, 0.030138208389282228, 0.03009347152709961, 0.02996633529663086, 0.029925376892089843, 0.029937664031982423, 0.03033225631713867, 0.030371807098388673, 0.029986751556396483, 0.02996944046020508, 0.03009507179260254, 0.030046207427978516, 0.02993356704711914, 0.030090528488159178, 0.030098047256469727, 0.03000534439086914, 0.02999235153198242, 0.03020057678222656, 0.030054239273071288, 0.02995814323425293, 0.03001532745361328, 0.029922496795654296, 0.03015603256225586, 0.03029782485961914, 0.030357503890991212, 0.03023052787780762, 0.030244319915771485, 0.030073375701904298, 0.03003392028808594, 0.030293888092041015, 0.03086297607421875, 0.030444032669067384, 0.030516544342041017, 0.030388927459716795, 0.03057459259033203, 0.0306824951171875, 0.030376575469970704, 0.030449344635009767, 0.030359872817993162, 0.030466047286987305, 0.030430591583251954, 0.030405248641967773, 0.030662656784057617, 0.030585920333862305, 0.030593984603881835, 0.030357503890991212, 0.03046713638305664, 0.03030521583557129, 0.030051328659057616, 0.03011417579650879, 0.03002022361755371, 0.03142451286315918, 0.030699520111083983, 0.030697471618652345, 0.03062518310546875, 0.030601823806762695, 0.03227443313598633, 0.03215705490112305, 0.03084351921081543, 0.03073993682861328, 0.03083500862121582, 0.03080828857421875, 0.0307957763671875, 0.030875648498535156, 0.030674943923950194, 0.03087936019897461, 0.030652799606323243, 0.030731712341308594, 0.030605888366699217, 0.030717952728271485, 0.030631935119628906, 0.03055356788635254, 0.030577152252197266, 0.030381439208984375, 0.03015951919555664, 0.030130176544189452, 0.02994175910949707, 0.02993561553955078, 0.03000912094116211, 0.029873632431030275, 0.029934335708618164, 0.030117887496948242, 0.030361600875854492, 0.03014860725402832, 0.03001468849182129, 0.02995894432067871, 0.029894847869873047, 0.029949760437011717, 0.03061555290222168, 0.030058496475219725, 0.029998752593994142, 0.029953567504882813, 0.029986879348754884, 0.030024288177490234, 0.030163103103637696, 0.02983526420593262, 0.029970111846923827, 0.029947296142578125, 0.030020511627197266, 0.02993756866455078, 0.030053535461425782, 0.030012351989746094, 0.030062591552734375, 0.030040063858032227, 0.029995008468627928, 0.030033599853515624, 0.030001472473144532, 0.030094751358032225, 0.029936223983764648, 0.030046207427978516, 0.02998476791381836, 0.031067840576171873, 0.03129558372497559, 0.03016726493835449, 0.030892160415649413, 0.030498687744140624, 0.03034726333618164, 0.03032294464111328, 0.030519039154052734, 0.030016576766967774, 0.030159807205200194, 0.03016089630126953, 0.030198047637939453, 0.030164703369140625, 0.030070783615112305, 0.0301844482421875, 0.02997865676879883, 0.03004863929748535, 0.03009596824645996, 0.030192863464355468, 0.03019651222229004, 0.03014790344238281, 0.030474943161010744, 0.030104703903198242, 0.02998566436767578, 0.030091424942016602, 0.03006447982788086, 0.02994598388671875, 0.030065696716308595, 0.029937631607055665, 0.0299815673828125, 0.03002070426940918, 0.030276512145996092, 0.030304256439208983, 0.030395551681518553, 0.030331743240356444, 0.030165023803710937, 0.030512351989746094, 0.03002649688720703, 0.030087104797363283, 0.030068031311035158, 0.030231359481811524, 0.030062528610229493, 0.030103424072265624, 0.02997260856628418, 0.029949888229370118, 0.029931295394897462, 0.029911327362060546, 0.029959775924682616, 0.02983760070800781, 0.029991039276123045, 0.031324159622192385, 0.030003200531005858, 0.030707296371459962, 0.03062006378173828, 0.0304005126953125, 0.03034316825866699, 0.030212160110473632, 0.03453459167480469, 0.03081907272338867, 0.030760959625244142, 0.03075836753845215, 0.03082604789733887, 0.030753471374511718, 0.030969728469848634, 0.031072639465332032, 0.031031328201293944]",tokens/s,32.92961280308994,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,892.899328,697.237504,0.0,301.989888,282.769408,s,1,7.8845537109375,7.8845537109375,0.0,7.8845537109375,7.8845537109375,7.8845537109375,7.8845537109375,[7.8845537109375],,kWh,2.6668313554152216e-05,2.9345323346169523e-06,8.527506822009201e-06,3.8130352710778365e-05,,MB,1210.179584,751.763456,0.0,341.835776,318.94528,s,17,0.18908361434936521,0.01112256554996266,0.00021222526015458952,0.011087583541870118,0.011266982078552245,0.011395865440368651,0.011730078697204589,"[0.011250687599182128, 0.011087583541870118, 0.010914239883422851, 0.011123167991638183, 0.010944448471069335, 0.010919199943542481, 0.01109347152709961, 0.011029439926147461, 0.010915871620178222, 0.01092950439453125, 0.011813632011413574, 0.011291423797607421, 0.011043744087219238, 0.011073887825012207, 0.01122047996520996, 0.011225503921508789, 0.011207327842712403]",tokens/s,23016.272536227887,kWh,3.2520368245172314e-07,3.5864289637114663e-08,1.7807704174389274e-07,5.391450138327304e-07,tokens/kWh,474825869.537623,MB,1223.41376,776.92928,0.0,367.0016,318.94784,s,17,10.032159484863282,0.5901270285213696,0.009695058014600227,0.5856761474609375,0.603881689453125,0.605954833984375,0.6079085449218751,"[0.5876029052734375, 0.5784276733398438, 0.58346044921875, 0.5853865966796875, 0.5813885498046875, 0.5830097045898438, 0.5856761474609375, 0.5824195556640624, 0.5794144287109375, 0.5796782836914063, 0.5980902099609375, 0.5937981567382813, 0.60071484375, 0.6029066162109376, 0.60839697265625, 0.596444091796875, 0.6053442993164062]",tokens/s,106.75667602931809,kWh,1.6719968791390176e-05,1.8439400282366233e-06,6.465554816367315e-06,2.5029463635994106e-05,tokens/kWh,2517033.5615742733,,s,1071,10.023562960624696,0.009359069057539398,0.00023659001780922944,0.009291839599609376,0.009652000427246094,0.009727535724639892,0.00992654056549072,"[0.009441120147705078, 0.009519743919372558, 0.009531871795654297, 0.009578495979309083, 0.009511712074279786, 0.009505760192871094, 0.009513248443603515, 0.009453280448913574, 0.009475647926330566, 0.009494976043701172, 0.009564160346984863, 0.009590784072875976, 0.00951315212249756, 0.009540448188781738, 0.009609248161315918, 0.009548416137695312, 0.00953171157836914, 0.009487839698791503, 0.0093721923828125, 0.009383935928344727, 0.009310208320617675, 0.009261055946350098, 0.009338879585266113, 0.009224191665649414, 0.009157759666442871, 0.009227168083190919, 0.009193216323852539, 0.009130208015441895, 0.009162464141845704, 0.009165087699890136, 0.009129983901977539, 0.009117024421691895, 0.009492480278015136, 0.009187999725341798, 0.009238431930541992, 0.009207136154174804, 0.0095731201171875, 0.009299967765808105, 0.009393600463867188, 0.009531807899475098, 0.009349120140075684, 0.009314463615417481, 0.009275391578674316, 0.009349311828613281, 0.00923142433166504, 0.009291839599609376, 0.009191488265991211, 0.009220447540283203, 0.009228416442871094, 0.009142335891723632, 0.00932259178161621, 0.009174912452697755, 0.009164511680603027, 0.00916051197052002, 0.009193344116210938, 0.009110239982604981, 0.009107456207275391, 0.009095040321350098, 0.009164095878601074, 0.009139360427856446, 0.009116640090942383, 0.009173695564270019, 0.009113216400146484, 0.009472000122070312, 0.009307200431823731, 0.009222304344177246, 0.009228927612304687, 0.009197088241577148, 0.009124608039855956, 0.009143327713012695, 0.009237248420715332, 0.009236576080322266, 0.009213983535766601, 0.009174880027770995, 0.00921788787841797, 0.00922163200378418, 0.009152607917785644, 0.009104063987731933, 0.00913584041595459, 0.00915017604827881, 0.009198080062866211, 0.009080896377563477, 0.009639776229858399, 0.009250176429748536, 0.009261759757995605, 0.009193568229675294, 0.009160351753234864, 0.009167200088500976, 0.009193471908569336, 0.009162752151489258, 0.009150464057922364, 0.009131872177124023, 0.009138079643249511, 0.009123647689819336, 0.009144767761230469, 0.009113280296325684, 0.009163071632385254, 0.00911695957183838, 0.009173055648803712, 0.009161375999450683, 0.009175040245056153, 0.009119744300842286, 0.009123583793640137, 0.009156864166259765, 0.009070591926574707, 0.009137248039245606, 0.009118592262268067, 0.009130016326904297, 0.009101311683654785, 0.009119744300842286, 0.009127455711364746, 0.009140704154968262, 0.009193471908569336, 0.009154272079467774, 0.009089311599731446, 0.009142271995544434, 0.009215999603271484, 0.009219103813171387, 0.009141216278076172, 0.009190624237060546, 0.009183839797973633, 0.009130175590515138, 0.009156607627868652, 0.009119168281555175, 0.00911622428894043, 0.00909721565246582, 0.0089584321975708, 0.009136256217956542, 0.009289600372314454, 0.009232383728027344, 0.00920576000213623, 0.009230463981628419, 0.009176959991455078, 0.009164128303527832, 0.009181856155395508, 0.009113439559936523, 0.009137439727783202, 0.00915135955810547, 0.0092674560546875, 0.009299712181091309, 0.00920201587677002, 0.009281184196472168, 0.009172863960266113, 0.009250304222106934, 0.009171584129333495, 0.009154175758361817, 0.009185728073120117, 0.009195648193359374, 0.009178943634033203, 0.009252863883972168, 0.009307647705078125, 0.009179648399353027, 0.00922812843322754, 0.009154815673828125, 0.009154208183288574, 0.009160608291625976, 0.009134431838989258, 0.009149696350097656, 0.009157312393188477, 0.009158720016479492, 0.009410367965698243, 0.009133343696594238, 0.009216383934020996, 0.009216544151306152, 0.009256863594055175, 0.00909119987487793, 0.009084383964538574, 0.009173503875732422, 0.00915071964263916, 0.009236224174499512, 0.009211903572082519, 0.009090271949768067, 0.00939628791809082, 0.00919212818145752, 0.011000032424926759, 0.010997568130493164, 0.00922374439239502, 0.009212351799011231, 0.00919961643218994, 0.009168448448181153, 0.009226688385009766, 0.009172927856445313, 0.009267264366149903, 0.009136128425598144, 0.009203712463378906, 0.009197376251220702, 0.009169343948364259, 0.00921945571899414, 0.00931503963470459, 0.009111807823181153, 0.009297056198120117, 0.009331551551818848, 0.009430784225463867, 0.00953983974456787, 0.009352224349975586, 0.009341919898986816, 0.00933071994781494, 0.009326560020446777, 0.009295999526977539, 0.00931827163696289, 0.009269248008728028, 0.009244671821594238, 0.009310272216796875, 0.009352479934692383, 0.00927359962463379, 0.009341279983520507, 0.00930992031097412, 0.009233792304992676, 0.009287872314453126, 0.009278271675109863, 0.009256928443908692, 0.00919961643218994, 0.009244832038879394, 0.009240544319152832, 0.00944480037689209, 0.009550527572631836, 0.009289471626281739, 0.009166144371032714, 0.009183648109436036, 0.00915833568572998, 0.009120256423950195, 0.009262271881103516, 0.009144960403442383, 0.009152671813964845, 0.009195008277893067, 0.009236191749572754, 0.009185728073120117, 0.009183103561401368, 0.009213791847229005, 0.00927023983001709, 0.009197343826293945, 0.009279071807861328, 0.009218560218811036, 0.009170944213867188, 0.00917081642150879, 0.009271167755126953, 0.009287455558776855, 0.009189279556274414, 0.009228544235229492, 0.00921724796295166, 0.009245696067810059, 0.00925932788848877, 0.009230015754699707, 0.009249823570251465, 0.009335776329040528, 0.00957852840423584, 0.009371616363525391, 0.009451519966125489, 0.009517056465148926, 0.009476096153259277, 0.009395936012268066, 0.009296064376831054, 0.008969023704528808, 0.00919753646850586, 0.009218015670776368, 0.009156864166259765, 0.009148415565490722, 0.009193599700927734, 0.009129695892333985, 0.00913590431213379, 0.009257504463195801, 0.009359487533569336, 0.009340160369873047, 0.009383935928344727, 0.009416319847106934, 0.009413536071777345, 0.009477760314941406, 0.00935315227508545, 0.009296256065368652, 0.009298048019409179, 0.009176416397094727, 0.009160991668701171, 0.00918553638458252, 0.009226240158081055, 0.009295040130615235, 0.01005241584777832, 0.009474047660827637, 0.00918943977355957, 0.009191007614135742, 0.009154560089111329, 0.009119744300842286, 0.00917689609527588, 0.00906704044342041, 0.009104607582092285, 0.009098015785217285, 0.009115648269653321, 0.00910905647277832, 0.009908703804016114, 0.009187295913696289, 0.009149920463562012, 0.009110048294067382, 0.009121888160705567, 0.009105312347412109, 0.009132032394409179, 0.009138239860534667, 0.009201408386230469, 0.00915187168121338, 0.009040703773498535, 0.009132032394409179, 0.00912713623046875, 0.009188128471374511, 0.00909216022491455, 0.009177472114562988, 0.00910598373413086, 0.009121472358703614, 0.009127584457397461, 0.009144288063049316, 0.009533535957336426, 0.00917360019683838, 0.009146240234375, 0.009265279769897461, 0.009261055946350098, 0.00918297576904297, 0.00914457607269287, 0.009107456207275391, 0.008892512321472168, 0.009139264106750488, 0.009157855987548829, 0.00912559986114502, 0.009142111778259278, 0.009184479713439941, 0.009091232299804687, 0.009070528030395509, 0.009123744010925293, 0.009094079971313477, 0.009149696350097656, 0.009157024383544921, 0.00927939224243164, 0.009169343948364259, 0.009141759872436523, 0.009187264442443848, 0.009230463981628419, 0.00912396812438965, 0.00912822437286377, 0.009181216239929199, 0.009166560173034667, 0.009178432464599609, 0.009217056274414062, 0.009236224174499512, 0.009200896263122559, 0.009191935539245605, 0.009275391578674316, 0.009185728073120117, 0.009209856033325196, 0.009264575958251952, 0.009246848106384277, 0.009198016166687011, 0.009334783554077148, 0.009237888336181641, 0.0092674560546875, 0.009309823989868165, 0.009890239715576172, 0.010084671974182129, 0.010168383598327637, 0.009383744239807128, 0.009398719787597657, 0.009307423591613769, 0.009333151817321778, 0.00923852825164795, 0.009226240158081055, 0.009205056190490722, 0.009169440269470214, 0.009164735794067383, 0.009150272369384765, 0.009203167915344238, 0.009221183776855468, 0.009201055526733398, 0.00919382381439209, 0.009132448196411133, 0.009146080017089844, 0.00915839958190918, 0.009189760208129883, 0.009180416107177734, 0.009239168167114257, 0.009403840065002441, 0.009362272262573242, 0.009399968147277832, 0.009279040336608887, 0.009047840118408203, 0.009216480255126953, 0.009375743865966797, 0.009324543952941895, 0.009206784248352052, 0.009227392196655273, 0.009168767929077148, 0.00915875244140625, 0.009140128135681153, 0.009166848182678223, 0.009137696266174316, 0.009150783538818359, 0.009129695892333985, 0.009158944129943848, 0.00919372844696045, 0.009271200180053712, 0.009185279846191406, 0.009147775650024415, 0.009124480247497558, 0.009209759712219238, 0.00914675235748291, 0.009174816131591797, 0.009169055938720702, 0.00918505573272705, 0.009165120124816895, 0.009268671989440917, 0.00937600040435791, 0.00932249641418457, 0.009250816345214843, 0.009292832374572754, 0.009232895851135254, 0.009218527793884277, 0.009254912376403808, 0.009197407722473145, 0.009291935920715332, 0.009309184074401856, 0.00935206413269043, 0.00929321575164795, 0.009313088417053222, 0.009231743812561035, 0.009341055870056153, 0.009282079696655274, 0.009303936004638672, 0.009277440071105958, 0.009297087669372558, 0.009251263618469239, 0.009366080284118653, 0.00929964828491211, 0.009326399803161621, 0.010393792152404786, 0.009527296066284179, 0.009426079750061035, 0.009467071533203125, 0.009370752334594726, 0.00938646411895752, 0.009621696472167969, 0.009910623550415039, 0.009276736259460449, 0.00930851173400879, 0.009285056114196777, 0.009228608131408692, 0.009221535682678222, 0.009239583969116212, 0.009046112060546875, 0.009224191665649414, 0.009247584342956543, 0.009262399673461914, 0.00924947166442871, 0.00935097599029541, 0.009183232307434081, 0.009279935836791993, 0.009236224174499512, 0.009267200469970703, 0.009215935707092286, 0.009261119842529296, 0.009237664222717285, 0.009216192245483399, 0.009148223876953125, 0.009400608062744141, 0.00920633602142334, 0.009137760162353516, 0.009152928352355956, 0.009200703620910645, 0.009178048133850097, 0.009332736015319825, 0.009250847816467286, 0.009303104400634766, 0.009274271965026856, 0.009221343994140626, 0.009218079566955566, 0.009151231765747071, 0.009196672439575195, 0.009169792175292969, 0.009158656120300293, 0.009136063575744629, 0.009193535804748534, 0.009385248184204102, 0.00951363182067871, 0.009457728385925294, 0.009479968070983887, 0.00954595184326172, 0.009484255790710449, 0.009280832290649414, 0.009274080276489257, 0.009213536262512208, 0.009152095794677734, 0.009136608123779297, 0.009132512092590332, 0.009119135856628418, 0.009287520408630371, 0.009249664306640626, 0.009448800086975098, 0.009152895927429199, 0.009211935997009278, 0.009108991622924804, 0.00911616039276123, 0.009119744300842286, 0.009113183975219727, 0.009143903732299804, 0.00915129566192627, 0.009217120170593262, 0.009249695777893066, 0.009315872192382812, 0.009196000099182129, 0.00922812843322754, 0.009152223587036133, 0.008919232368469239, 0.009159775733947753, 0.00913481616973877, 0.009115648269653321, 0.009250240325927734, 0.009122367858886718, 0.009110976219177246, 0.009165056228637695, 0.009088543891906739, 0.009095968246459961, 0.009112607955932618, 0.00916988754272461, 0.00909273624420166, 0.00912934398651123, 0.00909769630432129, 0.009400863647460938, 0.009184415817260743, 0.009105728149414063, 0.009296159744262695, 0.009146400451660156, 0.009177311897277832, 0.009129983901977539, 0.009178367614746094, 0.009159616470336914, 0.00910051155090332, 0.009183103561401368, 0.009143263816833495, 0.009117440223693847, 0.009103487968444825, 0.009227231979370117, 0.009210783958435059, 0.009136128425598144, 0.009119744300842286, 0.009152223587036133, 0.009170975685119629, 0.009191679954528808, 0.009258848190307617, 0.009234592437744141, 0.00927948760986328, 0.009319519996643067, 0.009251168251037598, 0.009235360145568848, 0.00919654369354248, 0.00915113639831543, 0.009154560089111329, 0.00912508773803711, 0.009200415611267089, 0.009203935623168945, 0.009259807586669921, 0.009311455726623535, 0.009267135620117188, 0.009205599784851075, 0.009283583641052246, 0.009236319541931153, 0.009248607635498047, 0.009263008117675782, 0.009229951858520508, 0.009175935745239257, 0.00928876781463623, 0.009333184242248535, 0.009421216011047364, 0.009220095634460449, 0.009181440353393555, 0.008955936431884765, 0.009155167579650878, 0.009172863960266113, 0.00912831974029541, 0.009140000343322754, 0.009149760246276855, 0.009177760124206543, 0.009137151718139648, 0.009208255767822266, 0.009228863716125489, 0.009193120002746582, 0.00913868808746338, 0.009205632209777832, 0.009265055656433105, 0.00925487995147705, 0.009240672111511231, 0.009259008407592773, 0.009157792091369629, 0.009124640464782716, 0.009233823776245117, 0.009202431678771973, 0.0091810884475708, 0.009173024177551269, 0.009170911788940429, 0.009119744300842286, 0.009156319618225097, 0.009223648071289062, 0.009132863998413086, 0.009095040321350098, 0.009462240219116212, 0.009227423667907715, 0.009239040374755859, 0.009233504295349122, 0.009200544357299804, 0.009188447952270508, 0.009177087783813476, 0.009198495864868164, 0.00920150375366211, 0.009160320281982421, 0.009202143669128419, 0.009254976272583008, 0.009154560089111329, 0.009215871810913085, 0.00919155216217041, 0.009166848182678223, 0.009150464057922364, 0.009185279846191406, 0.009189375877380371, 0.009183232307434081, 0.00921945571899414, 0.009251456260681152, 0.00912179183959961, 0.00920691204071045, 0.009200511932373047, 0.009246591567993165, 0.00919155216217041, 0.009216064453125, 0.009196864128112794, 0.009310879707336427, 0.009218208312988281, 0.009227392196655273, 0.009222623825073242, 0.009207008361816406, 0.009712063789367675, 0.009254207611083984, 0.009795743942260743, 0.009865792274475097, 0.009285632133483887, 0.009357312202453612, 0.009312416076660156, 0.009256447792053223, 0.009240703582763672, 0.00918342399597168, 0.009482272148132324, 0.009247872352600097, 0.009163935661315918, 0.009164608001708985, 0.009186623573303222, 0.009340607643127441, 0.009631839752197266, 0.009532223701477051, 0.009664511680603028, 0.009518783569335937, 0.009568832397460937, 0.00960700798034668, 0.009658271789550782, 0.009527104377746582, 0.009420991897583008, 0.009352736473083496, 0.009273471832275391, 0.009299936294555665, 0.009254752159118652, 0.009210399627685547, 0.009238656044006349, 0.009230208396911622, 0.00927235221862793, 0.009171839714050292, 0.009291711807250977, 0.009253215789794923, 0.009258079528808593, 0.009278176307678222, 0.009311519622802734, 0.00940124797821045, 0.009690208435058594, 0.00943769645690918, 0.009469504356384278, 0.009482784271240234, 0.009363231658935546, 0.009333087921142577, 0.00971564769744873, 0.00959494400024414, 0.01222646427154541, 0.009582752227783204, 0.009506752014160156, 0.009435296058654785, 0.009484352111816407, 0.009466848373413085, 0.009514911651611328, 0.009677536010742187, 0.009695424079895019, 0.009576736450195312, 0.009573472023010255, 0.009767552375793458, 0.009623295783996582, 0.009593024253845214, 0.00963539218902588, 0.009289919853210449, 0.009488351821899414, 0.00953980827331543, 0.009556480407714844, 0.009444607734680175, 0.009626111984252929, 0.009687392234802247, 0.009553824424743652, 0.009488479614257812, 0.009549823760986328, 0.009623552322387695, 0.009576607704162597, 0.009549568176269532, 0.00957027244567871, 0.009565919876098632, 0.009599455833435059, 0.009574175834655762, 0.009549983978271484, 0.009531392097473144, 0.009623552322387695, 0.00956441593170166, 0.00953116798400879, 0.009396224021911622, 0.009409631729125977, 0.009356063842773437, 0.009285728454589843, 0.009285599708557128, 0.009252256393432617, 0.009228320121765136, 0.009218688011169434, 0.009301983833312988, 0.00933407974243164, 0.009355968475341796, 0.009344863891601562, 0.009458047866821289, 0.009485759735107423, 0.009516672134399414, 0.00945257568359375, 0.009411935806274413, 0.009445792198181152, 0.009445183753967286, 0.009373087882995606, 0.009387807846069336, 0.009292736053466796, 0.009275391578674316, 0.009250720024108887, 0.009255007743835449, 0.009229887962341308, 0.009290176391601563, 0.00930611228942871, 0.0093306884765625, 0.009314111709594727, 0.009348544120788574, 0.009270015716552735, 0.009357503890991211, 0.009324352264404296, 0.009371456146240234, 0.009397760391235351, 0.009423551559448242, 0.00937388801574707, 0.009406271934509278, 0.00933676815032959, 0.009309632301330566, 0.009107232093811035, 0.009341152191162109, 0.009281536102294922, 0.009289407730102539, 0.009388192176818848, 0.009586688041687011, 0.009562335968017577, 0.009539520263671875, 0.009508864402770996, 0.009760736465454101, 0.009424927711486817, 0.009527296066284179, 0.00951296043395996, 0.009496383666992188, 0.009507007598876953, 0.009539584159851074, 0.009474047660827637, 0.009704607963562012, 0.00956015968322754, 0.00957027244567871, 0.009527296066284179, 0.009457823753356933, 0.009474687576293946, 0.009389216423034669, 0.009355968475341796, 0.009482527732849121, 0.009487903594970704, 0.009455967903137207, 0.009480192184448242, 0.009633055686950684, 0.009757504463195801, 0.009775008201599121, 0.009813471794128418, 0.009796128273010254, 0.009836544036865234, 0.009747743606567383, 0.009751263618469239, 0.009764479637145997, 0.009726335525512695, 0.00966761589050293, 0.009671648025512696, 0.009674912452697753, 0.009564127922058106, 0.009558912277221679, 0.009673407554626465, 0.009572575569152832, 0.00963798427581787, 0.009613056182861329, 0.009506303787231446, 0.009655296325683594, 0.009534367561340332, 0.00937782382965088, 0.009411135673522949, 0.009514847755432128, 0.009459903717041016, 0.009423199653625489, 0.009369407653808594, 0.009361472129821777, 0.009306240081787109, 0.00931827163696289, 0.009323776245117188, 0.00933743953704834, 0.009283743858337402, 0.009046751976013184, 0.00929043197631836, 0.009300992012023926, 0.009338815689086915, 0.009357119560241699, 0.009307552337646484, 0.009325407981872558, 0.009338687896728516, 0.009430303573608398, 0.009556672096252442, 0.009533663749694825, 0.00949830436706543, 0.009485695838928223, 0.009497535705566407, 0.009424384117126466, 0.00950704002380371, 0.009471296310424804, 0.00937059211730957, 0.009501728057861328, 0.009376735687255859, 0.00931430435180664, 0.009466976165771484, 0.009438112258911132, 0.009631584167480469, 0.009609375953674316, 0.00970137596130371, 0.0097259521484375, 0.009760767936706542, 0.009626655578613282, 0.0096243839263916, 0.009652383804321289, 0.009664447784423828, 0.009612895965576173, 0.009629759788513183, 0.009867615699768066, 0.009678912162780762, 0.009560064315795898, 0.009570303916931153, 0.009640095710754394, 0.00957151985168457, 0.009639967918395996, 0.009605536460876465, 0.009709312438964843, 0.009531488418579101, 0.009592448234558106, 0.009632320404052734, 0.009783488273620605, 0.009629216194152832, 0.009619584083557129, 0.009637951850891113, 0.009672063827514649, 0.00953983974456787, 0.00974505615234375, 0.009654272079467773, 0.009631839752197266, 0.009631775856018066, 0.0096561918258667, 0.009891776084899903, 0.009705151557922363, 0.009695615768432617, 0.009646080017089843, 0.009619296073913574, 0.009580320358276367, 0.009387552261352539, 0.009571871757507323, 0.00954259204864502, 0.009570303916931153, 0.009616864204406739, 0.009633376121520995, 0.009563072204589844, 0.009637663841247559, 0.00980399990081787, 0.00971996784210205, 0.009673824310302734, 0.00973420810699463, 0.009714176177978515, 0.00968505573272705, 0.009832736015319823, 0.009769887924194335, 0.00974124813079834, 0.00963817596435547, 0.009620927810668945, 0.009652000427246094, 0.009595392227172851, 0.009666048049926757, 0.00983296012878418, 0.009906175613403321, 0.009645312309265137, 0.009647071838378906, 0.009608991622924805, 0.00956931209564209, 0.009542079925537109, 0.009580191612243652, 0.009565119743347167, 0.009533056259155273, 0.009523103713989258, 0.009553631782531739, 0.00959558391571045, 0.009508864402770996, 0.009627872467041016, 0.00958176040649414, 0.009605440139770507, 0.009827648162841797, 0.009531392097473144, 0.009602016448974609, 0.009775103569030762, 0.009743712425231933, 0.00975648021697998, 0.009703519821166993, 0.00971615982055664, 0.009611552238464355, 0.009633824348449707, 0.009553919792175293, 0.009684991836547852, 0.00959881591796875, 0.009648096084594727, 0.009967071533203125, 0.00972873592376709, 0.009686783790588379, 0.009695039749145509, 0.00966089630126953, 0.009671680450439453, 0.009570560455322265, 0.009536224365234375, 0.00957148838043213, 0.009573216438293458, 0.009396224021911622, 0.009569664001464844, 0.009794015884399414, 0.009613375663757324, 0.009635071754455566, 0.009579360008239746, 0.009517056465148926, 0.009524415969848633, 0.009589056015014648, 0.009756768226623535, 0.009558431625366211, 0.00950607967376709, 0.00952393627166748, 0.009533375740051269, 0.009460800170898438, 0.009475071907043458, 0.009396224021911622, 0.009354816436767579, 0.00934342384338379, 0.009373696327209472, 0.009359616279602051, 0.009339712142944335, 0.009299936294555665, 0.009350079536437989, 0.00934921646118164, 0.009348671913146972, 0.00936793613433838, 0.00936518383026123, 0.009327199935913086, 0.00945907211303711, 0.009437824249267578, 0.009371359825134278, 0.009442591667175293, 0.009554847717285157, 0.009412416458129883, 0.009418527603149414, 0.009394559860229492, 0.009335935592651367, 0.009360095977783204, 0.009463616371154785, 0.009453023910522461, 0.00933347225189209, 0.00935321617126465, 0.009331839561462402, 0.009311103820800781, 0.00925705623626709, 0.009309439659118653, 0.009275296211242675, 0.009280159950256347, 0.00926636791229248, 0.009337599754333497, 0.009734272003173828, 0.010217503547668457, 0.009963680267333984, 0.009442496299743652, 0.00940492820739746, 0.009531552314758301, 0.009367168426513672, 0.009397791862487794, 0.009421664237976074, 0.009486207962036133, 0.009568384170532226, 0.009587807655334473, 0.009715840339660645, 0.009635231971740722, 0.010001055717468262, 0.009494688034057617, 0.009433088302612304, 0.009413727760314941, 0.009316864013671875, 0.009295616149902344, 0.00933135986328125, 0.009349087715148925, 0.009504799842834472, 0.00960848045349121, 0.009599007606506348, 0.009550080299377442, 0.009557600021362305, 0.009503583908081054, 0.00946940803527832, 0.009418815612792968, 0.00945359992980957, 0.009472255706787109, 0.00946604824066162, 0.009444352149963378, 0.009419520378112793, 0.009489824295043945, 0.009501536369323731, 0.009584639549255371, 0.00963913631439209, 0.009653183937072754, 0.009776224136352539, 0.009736479759216309, 0.00965011215209961, 0.009746047973632812, 0.009716352462768555, 0.009689375877380372, 0.00977468776702881, 0.009836128234863281, 0.009655263900756836, 0.009587807655334473, 0.009663104057312012, 0.009560192108154297, 0.009584735870361329, 0.009666463851928712, 0.009636896133422852, 0.009648799896240234, 0.009621343612670898, 0.009597599983215332, 0.009662272453308106, 0.00953705596923828, 0.009572959899902344, 0.009577919960021972, 0.009724448204040528, 0.009619359970092773, 0.009658528327941894, 0.009721887588500977, 0.00977302360534668, 0.009667776107788085, 0.009624223709106445, 0.009670656204223632, 0.009730048179626465, 0.009702752113342285, 0.009754719734191895, 0.009625472068786622, 0.009676704406738281]",tokens/s,106.84823392711571,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1558.757376,1861.156864,0.0,1465.909248,1358.169088,s,1,8.9897099609375,8.9897099609375,0.0,8.9897099609375,8.9897099609375,8.9897099609375,8.9897099609375,[8.9897099609375],,kWh,5.637756384169279e-05,6.211118228914528e-06,2.0644460960006494e-05,8.323314303061381e-05,,MB,1643.012096,1882.128384,0.0,1472.200704,1356.544512,s,10,0.4452473640441894,0.04452473640441894,0.000230934573435603,0.044446897506713864,0.04470375595092774,0.04491909484863281,0.04509136596679687,"[0.04513443374633789, 0.04442534255981445, 0.044324703216552734, 0.044585472106933595, 0.04443088150024414, 0.04446291351318359, 0.04465590286254883, 0.04453984069824219, 0.04426470565795899, 0.044423168182373046]",tokens/s,5749.612926952506,kWh,1.3550279394869336e-06,1.4942703200066107e-07,8.961992766296999e-07,2.4006542481172944e-06,tokens/kWh,106637596.89708221,MB,1651.46624,1882.128384,0.0,1472.200704,1409.94816,s,10,11.271697265625,1.1271697265624998,0.018579239538430867,1.1324287109375,1.15225517578125,1.153508386230469,1.1545109545898438,"[1.1547615966796876, 1.111167724609375, 1.1320802001953125, 1.1327772216796874, 1.1519766845703125, 1.1377652587890625, 1.1368804931640626, 1.1102855224609376, 1.1004166259765624, 1.1035859375]",tokens/s,55.89220373415232,kWh,3.237725702801701e-05,3.5706078520038297e-06,1.5798369634570826e-05,5.174623451459166e-05,tokens/kWh,1217479.8918409212,,s,630,11.268769048690773,0.017886934997921898,0.0004886429251878175,0.017771887779235843,0.01840783042907715,0.01857043237686157,0.018941482048034673,"[0.018433664321899416, 0.018131328582763673, 0.018302112579345702, 0.01831942367553711, 0.018147775650024414, 0.018092384338378908, 0.018077247619628905, 0.0179238395690918, 0.0179168643951416, 0.017909503936767577, 0.018465791702270508, 0.018035711288452147, 0.01802239990234375, 0.01822719955444336, 0.018233343124389647, 0.018263200759887695, 0.018498271942138673, 0.01828860855102539, 0.01826812744140625, 0.018292928695678713, 0.018290687561035156, 0.018284543991088868, 0.018455808639526367, 0.018359039306640627, 0.01818204879760742, 0.018145376205444336, 0.01825382423400879, 0.018358272552490236, 0.018330848693847657, 0.018278911590576173, 0.01834832000732422, 0.018462656021118164, 0.01850783920288086, 0.018591583251953123, 0.01856492805480957, 0.018301279067993163, 0.018294271469116212, 0.01831065559387207, 0.018435007095336915, 0.01837868881225586, 0.018342016220092773, 0.018309120178222657, 0.01821286392211914, 0.01837808036804199, 0.018365087509155272, 0.018550783157348632, 0.018481151580810547, 0.018245311737060548, 0.018372928619384766, 0.018571296691894532, 0.01852207946777344, 0.018329599380493163, 0.01840742492675781, 0.018327552795410155, 0.01836841583251953, 0.0184116153717041, 0.020241504669189454, 0.018825759887695314, 0.01843849563598633, 0.017876031875610352, 0.017907743453979493, 0.017949567794799805, 0.018101472854614258, 0.019706239700317384, 0.017653087615966796, 0.01745167922973633, 0.017554719924926757, 0.017468128204345703, 0.017345792770385744, 0.017400575637817384, 0.017555456161499023, 0.01745510482788086, 0.017375232696533204, 0.017441856384277345, 0.017454015731811524, 0.01746124839782715, 0.01737094306945801, 0.017497695922851563, 0.018227807998657225, 0.01897881507873535, 0.01772297668457031, 0.01759859275817871, 0.0174881591796875, 0.01740185546875, 0.017373184204101562, 0.017508352279663086, 0.017590272903442384, 0.017483232498168945, 0.017393760681152344, 0.017585599899291992, 0.017953792572021485, 0.017864479064941406, 0.01753264045715332, 0.01751705551147461, 0.0174400634765625, 0.01744879913330078, 0.01737603187561035, 0.017469375610351563, 0.017360671997070313, 0.017690975189208983, 0.01762099266052246, 0.017502208709716797, 0.017467391967773437, 0.017485824584960938, 0.017831199645996092, 0.01754185676574707, 0.017559551239013673, 0.0175119686126709, 0.017581632614135742, 0.017893312454223632, 0.01783875274658203, 0.017813823699951173, 0.01775152015686035, 0.017729536056518554, 0.01768502426147461, 0.017740896224975586, 0.01778361511230469, 0.017735679626464843, 0.01764156723022461, 0.01778838348388672, 0.01762505531311035, 0.01751206398010254, 0.017428735733032226, 0.017519296646118163, 0.017484832763671875, 0.017649951934814452, 0.017761983871459962, 0.017817375183105468, 0.017662496566772462, 0.01787215995788574, 0.017810144424438477, 0.0175914249420166, 0.017557952880859377, 0.01784876823425293, 0.017641471862792968, 0.017542623519897462, 0.01753273582458496, 0.017749984741210936, 0.0175882568359375, 0.017563648223876953, 0.017638111114501955, 0.017489919662475584, 0.0174202880859375, 0.017505407333374023, 0.017604768753051756, 0.017684415817260744, 0.017590848922729493, 0.017475807189941406, 0.017456703186035157, 0.0175927677154541, 0.017880640029907226, 0.01823583984375, 0.0183474235534668, 0.018221664428710937, 0.018705440521240235, 0.018252351760864257, 0.0183525447845459, 0.018416704177856444, 0.018613183975219726, 0.01829203224182129, 0.018320064544677734, 0.018134912490844726, 0.01815283203125, 0.01830988883972168, 0.018139135360717772, 0.01807535934448242, 0.018112127304077148, 0.018117279052734376, 0.01822105598449707, 0.018089984893798827, 0.01801420783996582, 0.01798921585083008, 0.01802684783935547, 0.018038848876953124, 0.018018207550048827, 0.017981536865234377, 0.018114303588867186, 0.018036991119384765, 0.017897056579589843, 0.017856927871704103, 0.01793769645690918, 0.017805023193359373, 0.01789187240600586, 0.018076128005981445, 0.018232351303100587, 0.018244192123413085, 0.01835865592956543, 0.018327232360839843, 0.018790719985961914, 0.018585887908935547, 0.01828233528137207, 0.018249887466430664, 0.01824358367919922, 0.01825382423400879, 0.018388736724853517, 0.018350175857543945, 0.018569375991821287, 0.01884297561645508, 0.01866819190979004, 0.018275903701782226, 0.01829478454589844, 0.01826041603088379, 0.018229248046875, 0.018318944931030274, 0.01838688087463379, 0.018430431365966796, 0.018650848388671874, 0.018850080490112303, 0.01822105598449707, 0.018206720352172853, 0.018203840255737305, 0.017910591125488283, 0.017886655807495117, 0.017885759353637697, 0.01776416015625, 0.017709247589111327, 0.017762304306030274, 0.0177458553314209, 0.017657920837402342, 0.01765171241760254, 0.01781350326538086, 0.0181712646484375, 0.018135679244995116, 0.017913856506347657, 0.01782921600341797, 0.017889951705932616, 0.01773695945739746, 0.018146047592163084, 0.017671743392944337, 0.01846259117126465, 0.017652288436889648, 0.018015520095825195, 0.017611200332641602, 0.01765318489074707, 0.017382240295410155, 0.017532928466796875, 0.017483135223388672, 0.01807833671569824, 0.017571647644042968, 0.01748601531982422, 0.017460224151611328, 0.017521663665771483, 0.017452768325805664, 0.017518367767333985, 0.018015743255615235, 0.017642080307006838, 0.017667680740356444, 0.017762144088745116, 0.01751139259338379, 0.017573888778686524, 0.017639039993286133, 0.017713535308837892, 0.01848646354675293, 0.018223743438720703, 0.021685823440551758, 0.018537088394165038, 0.018241535186767577, 0.018126848220825196, 0.0182061767578125, 0.01810259246826172, 0.018261856079101562, 0.01879599952697754, 0.018278656005859376, 0.018562911987304687, 0.01877689552307129, 0.018374176025390626, 0.018432384490966798, 0.018344032287597657, 0.018319360733032225, 0.01840947151184082, 0.018406463623046876, 0.01825267219543457, 0.018075712203979494, 0.018110368728637697, 0.018276447296142577, 0.01839321517944336, 0.01823321533203125, 0.0183306884765625, 0.018526880264282227, 0.018126848220825196, 0.018122495651245116, 0.017789695739746095, 0.017766176223754884, 0.017709056854248048, 0.017751712799072266, 0.017727296829223634, 0.01787673568725586, 0.01772991943359375, 0.017666463851928712, 0.017655487060546874, 0.01764793586730957, 0.017977344512939454, 0.01820979118347168, 0.018195327758789064, 0.01833718490600586, 0.018291423797607422, 0.018280448913574218, 0.01831888008117676, 0.018177663803100586, 0.018223968505859375, 0.018305023193359374, 0.01819647979736328, 0.01840742492675781, 0.01832979202270508, 0.01820038414001465, 0.018339839935302735, 0.018751520156860352, 0.018386911392211915, 0.019165184020996092, 0.01835612869262695, 0.018217056274414063, 0.018187295913696288, 0.018202688217163084, 0.018258848190307618, 0.01803468894958496, 0.018173824310302733, 0.018450559616088866, 0.017936384201049805, 0.01781724739074707, 0.017903968811035155, 0.018753536224365236, 0.017752063751220702, 0.017750015258789064, 0.01763705635070801, 0.01749839973449707, 0.017688608169555663, 0.01768841552734375, 0.017879199981689454, 0.017739776611328126, 0.01779427146911621, 0.017916032791137695, 0.01768649673461914, 0.01768684768676758, 0.017669952392578125, 0.017506879806518556, 0.017498111724853514, 0.017479232788085938, 0.017520511627197265, 0.017436256408691408, 0.017578975677490234, 0.017426431655883787, 0.01763532829284668, 0.017524736404418945, 0.017661951065063478, 0.017672191619873046, 0.0177475528717041, 0.01771356773376465, 0.01756159973144531, 0.01764352035522461, 0.01798963165283203, 0.01803059196472168, 0.01809542465209961, 0.01816032028198242, 0.018155519485473632, 0.018025632858276366, 0.018137727737426758, 0.018221376419067382, 0.018190143585205078, 0.018151519775390625, 0.018310495376586914, 0.018437856674194335, 0.01814214324951172, 0.022026239395141603, 0.01843564796447754, 0.01823711967468262, 0.018156288146972656, 0.018277408599853516, 0.01849350357055664, 0.018315296173095703, 0.018279296875, 0.018765888214111327, 0.018593727111816408, 0.01884364891052246, 0.018498752593994142, 0.018327455520629882, 0.01840764808654785, 0.018301376342773436, 0.018281728744506835, 0.018490560531616212, 0.01826259231567383, 0.01831484794616699, 0.01823299217224121, 0.018136064529418947, 0.018075647354125975, 0.018150720596313476, 0.018121408462524413, 0.018382848739624022, 0.01849750328063965, 0.018589727401733397, 0.018357696533203124, 0.0184202880859375, 0.018241535186767577, 0.018124799728393554, 0.017897472381591797, 0.01786675262451172, 0.01783955192565918, 0.017818208694458007, 0.017838048934936523, 0.017804351806640625, 0.017725439071655275, 0.017660863876342775, 0.017958560943603517, 0.017626880645751953, 0.017711711883544923, 0.0176312313079834, 0.01772537612915039, 0.017649280548095704, 0.017521087646484374, 0.01800115203857422, 0.01811315155029297, 0.018142623901367186, 0.01861222457885742, 0.018684288024902344, 0.01820483207702637, 0.018243776321411134, 0.018142847061157225, 0.01812905693054199, 0.018156991958618165, 0.018180896759033203, 0.018077823638916017, 0.01808572769165039, 0.01799171257019043, 0.01803004837036133, 0.018307615280151367, 0.018182144165039063, 0.01840742492675781, 0.018261119842529296, 0.018074495315551758, 0.01815705680847168, 0.01811916732788086, 0.017901439666748047, 0.01779520034790039, 0.017795072555541993, 0.017838079452514647, 0.017706207275390625, 0.01803651237487793, 0.017799392700195312, 0.017890079498291016, 0.017682207107543944, 0.017682655334472656, 0.017532352447509766, 0.01745542335510254, 0.017581056594848633, 0.017377439498901366, 0.017377056121826173, 0.017406719207763672, 0.017283071517944337, 0.017319936752319336, 0.01791542434692383, 0.018665952682495116, 0.017520639419555666, 0.0174653434753418, 0.017661951065063478, 0.017550559997558595, 0.0174354248046875, 0.017764352798461915, 0.017429920196533204, 0.017422943115234374, 0.017506303787231444, 0.017496063232421876, 0.017591999053955077, 0.01748204803466797, 0.01744428825378418, 0.017465919494628907, 0.017569215774536132, 0.01744748878479004, 0.01746329689025879, 0.01860745620727539, 0.017494688034057616, 0.017421375274658202, 0.01743929672241211, 0.01742448043823242, 0.017359136581420898, 0.01740563201904297, 0.017487232208251952, 0.017674367904663087, 0.017715423583984376, 0.0176595516204834, 0.01822153663635254, 0.018229408264160155, 0.017934431076049806, 0.017761695861816407, 0.017701696395874024, 0.017780479431152345, 0.017764608383178712, 0.017794111251831054, 0.017777599334716798, 0.01789129638671875, 0.017690656661987304, 0.017732767105102538, 0.018058080673217774, 0.01761894416809082, 0.017561151504516603, 0.017580352783203124, 0.01741632080078125, 0.01755267143249512, 0.017545951843261718, 0.01760870361328125, 0.01764659118652344, 0.017443296432495117, 0.017406496047973632, 0.017514495849609374, 0.017573888778686524, 0.01746086311340332, 0.017567743301391603, 0.017529056549072265, 0.017409824371337892, 0.017345535278320313, 0.01741721534729004, 0.017330175399780275, 0.017441823959350587, 0.017777631759643555, 0.01781488037109375, 0.017603231430053712, 0.017489599227905273, 0.017426624298095703, 0.017418367385864258, 0.01738751983642578, 0.01740390396118164, 0.017518592834472657, 0.01762918472290039, 0.017333343505859376, 0.017375808715820312, 0.01739401626586914, 0.017327360153198242, 0.01739379119873047, 0.017336959838867187, 0.017331743240356447, 0.017359071731567383, 0.01734681510925293, 0.017354751586914064, 0.01737913513183594, 0.017451200485229492, 0.018071296691894532, 0.01765350341796875, 0.017499776840209962, 0.017516576766967773, 0.017371295928955078, 0.01744054412841797, 0.017371583938598632, 0.017609184265136718, 0.01739151954650879, 0.017494112014770507, 0.017469440460205078, 0.01737932777404785, 0.017372928619384765, 0.0175229434967041, 0.017482847213745118, 0.017582815170288087, 0.017360639572143555, 0.017449407577514647, 0.017494016647338868, 0.01742959976196289, 0.0175645751953125, 0.01741619110107422, 0.017369087219238282, 0.01743212890625, 0.017510015487670897, 0.017482559204101564, 0.017479167938232423, 0.01738598442077637, 0.017451007843017577, 0.017464895248413086, 0.01745052719116211, 0.017403871536254882, 0.017465599060058595, 0.017448671340942384, 0.01761484718322754, 0.017519968032836914, 0.017423007965087892, 0.017364320755004884, 0.017498687744140626, 0.017373311996459962, 0.017350528717041017, 0.017426528930664063, 0.01737932777404785, 0.017340063095092773, 0.017350400924682617, 0.017375839233398437, 0.017391263961791994, 0.017305696487426758, 0.017253984451293947, 0.017402240753173828, 0.017335840225219726, 0.01732275199890137, 0.01739571189880371, 0.01741004753112793, 0.017391616821289063, 0.017393760681152344, 0.01742019271850586, 0.017422143936157226, 0.017413951873779296, 0.017375616073608397, 0.017530879974365234, 0.01754521560668945, 0.02065577507019043, 0.01744316864013672, 0.017487871170043946, 0.017467391967773437, 0.017451007843017577, 0.017356800079345702, 0.017384672164916994, 0.01740060806274414, 0.01741414451599121, 0.017383424758911133, 0.017395679473876952, 0.017338399887084962, 0.017364992141723632, 0.017348159790039064, 0.017337791442871092, 0.01737126350402832, 0.017355648040771485, 0.017284223556518555, 0.017331071853637695, 0.017358400344848632, 0.01739107131958008, 0.017340639114379882, 0.01740444755554199, 0.018694368362426758, 0.018763776779174804, 0.017467391967773437, 0.017411487579345703, 0.017504863739013672, 0.017491455078125, 0.01807756805419922, 0.01774790382385254, 0.017465631484985353, 0.017426271438598633, 0.017433151245117188, 0.017474624633789064]",tokens/s,55.90672745868311,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1948.069888,2886.664192,0.0,2491.416576,2425.650176,s,1,9.972828125,9.972828125,0.0,9.972828125,9.972828125,9.972828125,9.972828125,[9.972828125],,kWh,8.597890382082672e-05,9.476869520918352e-06,3.169669202399139e-05,0.00012715246536573646,,MB,1733.013504,3115.25376,0.0,2705.32608,2606.127616,s,10,0.4999646072387696,0.04999646072387696,0.0003091115439414927,0.04991495895385742,0.05013563652038574,0.050508970832824705,0.05080763828277588,"[0.05088230514526367, 0.049979934692382814, 0.04994582366943359, 0.04974195098876953, 0.049794143676757815, 0.05005267333984375, 0.04999516677856446, 0.049836673736572266, 0.049851840972900394, 0.04988409423828125]",tokens/s,5120.362447531037,kWh,1.5538485170211853e-06,1.713596663676243e-07,1.0287892391062423e-06,2.7539974224950518e-06,tokens/kWh,92955787.79738672,MB,1737.27744,3115.25376,0.0,2705.32608,2606.130176,s,10,13.563270996093749,1.3563270996093748,0.010747341118720936,1.3583674926757812,1.3651344482421874,1.3715413452148437,1.3766668627929688,"[1.3593658447265624, 1.3383299560546875, 1.34439794921875, 1.3485174560546875, 1.3503255615234375, 1.357369140625, 1.363710693359375, 1.360469970703125, 1.362836181640625, 1.3779482421875]",tokens/s,46.44897238884642,kWh,4.0362754617560954e-05,4.451614435356935e-06,1.988122748889506e-05,6.469559654181294e-05,tokens/kWh,973791.1599483116,,s,630,13.561056861877454,0.021525487082345147,0.0003614071250362197,0.021439663887023924,0.021907794761657717,0.022065814304351807,0.022571212730407717,"[0.021614528656005858, 0.021540992736816405, 0.022388736724853517, 0.022577152252197266, 0.02147430419921875, 0.021477760314941405, 0.021293312072753905, 0.02134422492980957, 0.02127065658569336, 0.02156732749938965, 0.02171539115905762, 0.021896383285522462, 0.021689151763916014, 0.02219161605834961, 0.0218176326751709, 0.021835199356079103, 0.021738271713256836, 0.021774335861206053, 0.021761056900024413, 0.021926271438598634, 0.021784639358520506, 0.021797407150268556, 0.021772287368774415, 0.021751808166503905, 0.021770240783691407, 0.021727231979370116, 0.021792671203613282, 0.021799007415771485, 0.02179836845397949, 0.021647296905517577, 0.02166796875, 0.021680608749389648, 0.021583423614501954, 0.021592512130737304, 0.021531936645507812, 0.021443199157714844, 0.021420127868652345, 0.021278751373291015, 0.02127449607849121, 0.02119219207763672, 0.02136284828186035, 0.021162080764770507, 0.02121558380126953, 0.021136735916137694, 0.02431862449645996, 0.021315584182739256, 0.02123084831237793, 0.02120569610595703, 0.021321792602539063, 0.021304927825927734, 0.021481151580810546, 0.021306079864501955, 0.021157535552978515, 0.021276735305786134, 0.021289247512817383, 0.021202943801879884, 0.021208127975463866, 0.02117932891845703, 0.021263551712036134, 0.021310272216796874, 0.021231103897094726, 0.021238271713256835, 0.02122956848144531, 0.02152038383483887, 0.02135603141784668, 0.02136729621887207, 0.021376352310180664, 0.02155708885192871, 0.022068031311035158, 0.021384511947631836, 0.021309247970581056, 0.021463359832763672, 0.021213760375976564, 0.021170175552368165, 0.021169919967651368, 0.021838048934936523, 0.021268512725830076, 0.021214847564697267, 0.021125503540039062, 0.02114963150024414, 0.021045215606689455, 0.02126652717590332, 0.02111267280578613, 0.021182144165039062, 0.021053920745849608, 0.021082111358642578, 0.021129215240478515, 0.021368831634521485, 0.021981184005737304, 0.021161600112915038, 0.02107139205932617, 0.021042015075683592, 0.021064960479736328, 0.021193471908569336, 0.02119606399536133, 0.021287647247314453, 0.02105887985229492, 0.02107375907897949, 0.02118332862854004, 0.021096128463745117, 0.021132736206054686, 0.021213920593261718, 0.021242015838623046, 0.021198848724365234, 0.02115692710876465, 0.021161184310913086, 0.021113824844360352, 0.021118719100952147, 0.02110073661804199, 0.021154624938964844, 0.02140719985961914, 0.02112156867980957, 0.021090303421020508, 0.02126028823852539, 0.021114879608154297, 0.021157888412475585, 0.02109644889831543, 0.021135072708129882, 0.021217344284057617, 0.02123798370361328, 0.021082111358642578, 0.021194751739501954, 0.021403648376464843, 0.02125619125366211, 0.021295103073120117, 0.0212674560546875, 0.021819456100463867, 0.021786687850952148, 0.021790719985961913, 0.02168627166748047, 0.021440511703491212, 0.02124799919128418, 0.02129715156555176, 0.021172224044799806, 0.02126438331604004, 0.02130534362792969, 0.021196800231933592, 0.021295103073120117, 0.021346303939819337, 0.02123366355895996, 0.021182464599609374, 0.02125212860107422, 0.021178335189819337, 0.021219327926635743, 0.02124300765991211, 0.02125449562072754, 0.021188512802124023, 0.021203584671020508, 0.02143404769897461, 0.021291328430175782, 0.021336063385009766, 0.02143436813354492, 0.021773887634277345, 0.021363136291503906, 0.02144486427307129, 0.02147030448913574, 0.021340351104736328, 0.02131715202331543, 0.02143123245239258, 0.022231039047241212, 0.021415872573852537, 0.02123097610473633, 0.02121798324584961, 0.021224832534790038, 0.021157503128051758, 0.021117952346801756, 0.02111884880065918, 0.0212542724609375, 0.02114684867858887, 0.021174463272094726, 0.021117536544799805, 0.02127408027648926, 0.02120944023132324, 0.021215423583984375, 0.021243904113769533, 0.02167788887023926, 0.02152876853942871, 0.021302848815917968, 0.021234111785888674, 0.021202016830444335, 0.02123664093017578, 0.02131155204772949, 0.021135200500488283, 0.021245216369628905, 0.021371519088745117, 0.02127686309814453, 0.021388959884643555, 0.021364543914794924, 0.021317216873168947, 0.021408544540405274, 0.02122547149658203, 0.0212807674407959, 0.021393247604370117, 0.021352575302124022, 0.021356576919555663, 0.021322879791259765, 0.021377119064331054, 0.021177120208740234, 0.02116511917114258, 0.02119366455078125, 0.021114879608154297, 0.021174272537231444, 0.021131263732910157, 0.021075456619262696, 0.021117439270019533, 0.021053440093994142, 0.021319679260253906, 0.021180416107177736, 0.021132352828979493, 0.021164800643920897, 0.021192895889282228, 0.021456895828247072, 0.021325824737548828, 0.021395040512084962, 0.021253696441650392, 0.021228384017944336, 0.021303295135498047, 0.02123526382446289, 0.021324127197265626, 0.021194847106933593, 0.02129715156555176, 0.021303295135498047, 0.02132524871826172, 0.021369407653808594, 0.021256032943725585, 0.021297183990478516, 0.021311616897583006, 0.02127401542663574, 0.021385791778564454, 0.021338144302368165, 0.02145414352416992, 0.02184671974182129, 0.021460512161254882, 0.021307519912719727, 0.021428575515747072, 0.02135443115234375, 0.021319103240966798, 0.021396095275878907, 0.021379072189331053, 0.021618431091308593, 0.021836095809936524, 0.021812320709228516, 0.021871456146240233, 0.021820480346679688, 0.021846975326538086, 0.02168012809753418, 0.021800960540771484, 0.021813119888305664, 0.021796287536621092, 0.021772703170776366, 0.021694751739501954, 0.022171648025512695, 0.022556671142578123, 0.02221014404296875, 0.02210652732849121, 0.021972768783569335, 0.02185772705078125, 0.021830432891845702, 0.021750911712646485, 0.021623680114746094, 0.021753856658935547, 0.02148703956604004, 0.0213753604888916, 0.021215423583984375, 0.021157888412475585, 0.021331071853637695, 0.021202911376953126, 0.021169055938720704, 0.02122534370422363, 0.021091840744018556, 0.021190656661987304, 0.021189247131347656, 0.021069631576538086, 0.02112441635131836, 0.021148544311523437, 0.021065088272094728, 0.021146240234375, 0.021263711929321288, 0.021209344863891602, 0.021224992752075195, 0.021220224380493164, 0.022190080642700196, 0.022796287536621093, 0.02145280075073242, 0.02123936080932617, 0.021246400833129883, 0.021233024597167968, 0.02114419174194336, 0.02112499237060547, 0.02121945571899414, 0.021168127059936523, 0.021120576858520507, 0.021100128173828125, 0.02114143943786621, 0.021126047134399414, 0.021281951904296874, 0.021146463394165038, 0.02127052879333496, 0.02125347137451172, 0.02116035270690918, 0.02113727951049805, 0.021346111297607422, 0.02123958396911621, 0.021390111923217773, 0.02147532844543457, 0.021444608688354492, 0.021551103591918946, 0.021749568939208985, 0.021727392196655274, 0.02165558433532715, 0.02162073516845703, 0.021575679779052736, 0.0216494083404541, 0.021685407638549804, 0.021668256759643553, 0.02177222442626953, 0.021766143798828123, 0.02188083267211914, 0.021894336700439453, 0.021768896102905274, 0.021727359771728516, 0.021567487716674806, 0.021436416625976562, 0.021698816299438477, 0.021447999954223633, 0.021309120178222656, 0.021416160583496095, 0.021252639770507814, 0.02132080078125, 0.021240255355834962, 0.02120342445373535, 0.021347711563110352, 0.021299840927124024, 0.022634496688842775, 0.02186614418029785, 0.02173529624938965, 0.021683807373046874, 0.021563488006591795, 0.021449440002441405, 0.022058271408081056, 0.021977664947509766, 0.021433696746826172, 0.021557600021362304, 0.021426719665527345, 0.021342016220092772, 0.021618879318237305, 0.022124544143676757, 0.021352447509765626, 0.021294368743896484, 0.021433055877685545, 0.021340160369873046, 0.021581823348999024, 0.02143846321105957, 0.021356544494628905, 0.021389312744140625, 0.02126950454711914, 0.021857280731201172, 0.021792768478393554, 0.02189107131958008, 0.02156915283203125, 0.021430656433105467, 0.021548128128051756, 0.021414815902709963, 0.021403648376464843, 0.021360095977783204, 0.021219263076782225, 0.02120150375366211, 0.02127872085571289, 0.02124595260620117, 0.021215391159057618, 0.0213703670501709, 0.02143881607055664, 0.021614591598510743, 0.021563295364379884, 0.021590112686157226, 0.021568544387817384, 0.021582048416137697, 0.021701055526733397, 0.02228223991394043, 0.021885215759277345, 0.021937183380126953, 0.021858207702636717, 0.021904159545898437, 0.02169241523742676, 0.021776384353637695, 0.021816831588745117, 0.021541376113891602, 0.021525951385498048, 0.02146771240234375, 0.021495744705200194, 0.021441696166992187, 0.021754783630371095, 0.021652767181396484, 0.02156390380859375, 0.021403295516967773, 0.021350080490112305, 0.02134105682373047, 0.0212359676361084, 0.02200873565673828, 0.02217622375488281, 0.02245465660095215, 0.02165350341796875, 0.02149504089355469, 0.021742015838623046, 0.021493152618408205, 0.021566368103027343, 0.02164121627807617, 0.02167158317565918, 0.02160470390319824, 0.021462303161621094, 0.021352767944335938, 0.021293472290039063, 0.021399551391601563, 0.021391359329223633, 0.021448511123657227, 0.021310783386230468, 0.021241952896118164, 0.021412448883056642, 0.021446847915649415, 0.0225218563079834, 0.022253568649291993, 0.02166988754272461, 0.021649023056030274, 0.0215350399017334, 0.021579679489135743, 0.021483295440673827, 0.02140937614440918, 0.021327840805053712, 0.021265216827392578, 0.02232646369934082, 0.02142223930358887, 0.021602527618408203, 0.02128121566772461, 0.02140323257446289, 0.021135295867919922, 0.021246431350708007, 0.02457779121398926, 0.021712831497192383, 0.021319999694824218, 0.021298559188842773, 0.021289600372314452, 0.02165727996826172, 0.021391807556152345, 0.021428096771240236, 0.021446399688720703, 0.021307647705078123, 0.021364736557006835, 0.021186559677124024, 0.021198432922363283, 0.021387807846069334, 0.021534591674804687, 0.021391359329223633, 0.021272512435913087, 0.021285087585449218, 0.021284704208374024, 0.021374975204467773, 0.021301151275634766, 0.02129520034790039, 0.02131328010559082, 0.02129852867126465, 0.021381792068481446, 0.021346208572387695, 0.021415647506713868, 0.021497919082641603, 0.021879360198974608, 0.021769216537475586, 0.021852800369262695, 0.021827583312988282, 0.021688703536987305, 0.021755903244018555, 0.02170822334289551, 0.021596416473388672, 0.02169683265686035, 0.021665632247924806, 0.021700544357299803, 0.021724767684936523, 0.021641536712646483, 0.021533088684082033, 0.021551008224487304, 0.021612543106079102, 0.02156447982788086, 0.021730239868164063, 0.021696672439575196, 0.021716064453125, 0.021832096099853517, 0.02163337516784668, 0.02152448081970215, 0.02195039939880371, 0.02130335998535156, 0.02135180854797363, 0.021492351531982423, 0.021356544494628905, 0.021380607604980468, 0.021422592163085938, 0.02136412811279297, 0.02248966407775879, 0.023604896545410155, 0.02178700828552246, 0.021743616104125976, 0.02186444854736328, 0.021777856826782228, 0.021709375381469727, 0.021693599700927733, 0.021680992126464845, 0.022190176010131835, 0.021877792358398436, 0.021856000900268555, 0.021824928283691408, 0.02176272010803223, 0.021639167785644533, 0.021506048202514647, 0.021557247161865235, 0.02149990463256836, 0.021420032501220702, 0.02151219177246094, 0.02164454460144043, 0.021850879669189454, 0.0218723201751709, 0.02152467155456543, 0.02127401542663574, 0.02135932731628418, 0.02128486442565918, 0.02122256088256836, 0.02121404838562012, 0.021241472244262694, 0.021686943054199218, 0.02139423942565918, 0.02118320083618164, 0.02130143928527832, 0.021794975280761717, 0.02172297668457031, 0.021738719940185548, 0.02168707275390625, 0.021855424880981446, 0.021734207153320313, 0.021792160034179688, 0.021720672607421877, 0.021808128356933593, 0.021938175201416017, 0.021677503585815428, 0.021975616455078124, 0.02191980743408203, 0.022157247543334962, 0.02176188850402832, 0.02196905517578125, 0.021548927307128905, 0.02140096092224121, 0.021684991836547853, 0.02149580764770508, 0.02152038383483887, 0.021573408126831055, 0.022048608779907226, 0.021480831146240234, 0.021572288513183595, 0.021485696792602538, 0.021294687271118166, 0.02148384094238281, 0.021352256774902344, 0.021579616546630858, 0.021392000198364257, 0.021372127532958984, 0.02145359992980957, 0.02191084861755371, 0.0215765438079834, 0.02173731231689453, 0.02186240005493164, 0.021817344665527344, 0.021977855682373048, 0.02186419105529785, 0.02224358367919922, 0.02185420799255371, 0.021716991424560548, 0.021549055099487305, 0.021716064453125, 0.021767072677612305, 0.02170412826538086, 0.021631551742553712, 0.021732479095458983, 0.021588863372802733, 0.021769760131835937, 0.021447040557861327, 0.021463136672973632, 0.021544960021972655, 0.02185215950012207, 0.02200739288330078, 0.021901311874389647, 0.02184547233581543, 0.021840831756591798, 0.021743616104125976, 0.02182931137084961, 0.021810911178588868, 0.02179542350769043, 0.02190745544433594, 0.021810943603515626, 0.021985536575317384, 0.02192300796508789, 0.021944320678710938, 0.021974880218505858, 0.021902111053466795, 0.021991615295410157, 0.02202828788757324, 0.02204876708984375, 0.02209587287902832, 0.02208153533935547, 0.02211840057373047, 0.021937471389770508, 0.021940128326416015, 0.022806848526000977, 0.02202467155456543, 0.022091775894165038, 0.022063104629516602, 0.022018047332763673, 0.02199734306335449, 0.021797088623046874, 0.02188595199584961, 0.022258399963378906, 0.022010143280029298, 0.021906848907470702, 0.021981792449951174, 0.02187990379333496, 0.021883039474487304, 0.021747711181640626, 0.021752159118652345, 0.02187868881225586, 0.021740032196044923, 0.021776384353637695, 0.0216494083404541, 0.02165760040283203, 0.021542911529541017, 0.021503679275512694]",tokens/s,46.456556182655845,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Qwen/Qwen-14B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Qwen/Qwen-14B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Qwen/Qwen-7B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Qwen/Qwen-7B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 1013, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 839, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 467, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 397, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 199, in _flash_attention_forward attn_output_unpad = flash_attn_varlen_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 1124, in flash_attn_varlen_func return FlashAttnVarlenFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 620, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_varlen_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 90, in _flash_attn_varlen_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.varlen_fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Deci/DeciCoder-1b contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Deci/DeciCoder-1b. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Deci/DeciLM-7B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Deci/DeciLM-7B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Qwen/Qwen-72B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Qwen/Qwen-72B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1141, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 944, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 677, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 500, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Qwen/Qwen-14B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Qwen/Qwen-14B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Qwen/Qwen-7B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Qwen/Qwen-7B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4169.797632,4725.866496,0.0,4347.396096,4328.833024,s,1,10.168271484375,10.168271484375,0.0,10.168271484375,10.168271484375,10.168271484375,10.168271484375,[10.168271484375],,kWh,9.025456107503941e-05,9.946536858022936e-06,2.9206134476017498e-05,0.00012940723240907983,,MB,1367.089152,5212.40576,0.0,4804.575232,4748.27776,s,10,3.915197418212891,0.3915197418212891,0.004421960725253466,0.3930104522705078,0.3947453002929687,0.39542315673828127,0.39596544189453126,"[0.3796112060546875, 0.3929827880859375, 0.39303811645507813, 0.38973516845703127, 0.39199102783203127, 0.39459466552734374, 0.3935941467285156, 0.38942996215820314, 0.39411932373046876, 0.39610101318359375]",tokens/s,653.862302853817,kWh,1.1309097754005902e-05,1.2471398294166202e-06,7.5085636991536815e-06,2.0064801282576203e-05,tokens/kWh,12758661.119774176,MB,1372.336128,5321.457664,0.0,4913.627136,4878.100992,s,10,19.812693237304686,1.9812693237304686,0.011225055036968759,1.9823427734374999,1.9938609130859375,1.9943759521484377,1.9947879833984377,"[1.991484619140625, 1.9722833251953125, 1.964276123046875, 1.9937464599609376, 1.9711112060546876, 1.9695313720703125, 1.974825439453125, 1.989860107421875, 1.99068359375, 1.9948909912109376]",tokens/s,31.797797121988097,kWh,5.8298273269744395e-05,6.430324215388951e-06,3.8719584394445744e-05,0.0001034481818795791,tokens/kWh,609000.5532754206,,s,630,19.80954297065734,0.03144371900104341,0.0005867593209498358,0.0313822717666626,0.031838938522338864,0.03218983402252197,0.03419297065734864,"[0.033762046813964844, 0.03153267288208008, 0.03125686454772949, 0.031538591384887696, 0.031449663162231446, 0.03143084716796875, 0.03164499282836914, 0.03148371124267578, 0.03145123291015625, 0.031691551208496094, 0.031807424545288086, 0.03168671989440918, 0.03228876876831055, 0.03387596893310547, 0.03170918464660644, 0.0314466552734375, 0.03171494483947754, 0.0313798713684082, 0.03173417663574219, 0.03166201591491699, 0.031355903625488284, 0.03144150352478027, 0.031711328506469724, 0.03152025604248047, 0.031591007232666016, 0.03161315155029297, 0.03167225646972656, 0.03169696044921875, 0.03174399948120117, 0.03182796859741211, 0.031707263946533205, 0.03164508819580078, 0.03175267219543457, 0.0313384952545166, 0.03167436790466309, 0.031765951156616214, 0.03162719917297363, 0.03149596786499023, 0.031416671752929684, 0.03132671928405762, 0.031637535095214844, 0.03200960159301758, 0.031658592224121096, 0.0315098876953125, 0.031615167617797854, 0.03180108833312988, 0.031480192184448245, 0.03139769554138184, 0.031547679901123046, 0.03143843269348145, 0.0314783992767334, 0.03172294425964355, 0.03136774444580078, 0.03142643165588379, 0.03149017524719238, 0.03126681518554687, 0.031184223175048827, 0.031093408584594726, 0.031110240936279298, 0.030880512237548827, 0.030982303619384765, 0.031395263671875, 0.03121753692626953, 0.03423360061645508, 0.032199424743652345, 0.03136511993408203, 0.03141961669921875, 0.0315830078125, 0.03149619293212891, 0.031080448150634765, 0.031857696533203125, 0.031238880157470703, 0.031097087860107422, 0.03209830474853516, 0.031213567733764647, 0.031252479553222655, 0.030887136459350584, 0.03086809539794922, 0.03114409637451172, 0.03079987144470215, 0.03102720069885254, 0.030686656951904298, 0.030695968627929688, 0.031242271423339844, 0.03122790336608887, 0.031100927352905275, 0.03153919982910156, 0.0312805118560791, 0.03120102310180664, 0.03148643112182617, 0.031166879653930665, 0.031178272247314454, 0.03159491157531738, 0.03162656021118164, 0.031146751403808594, 0.031141599655151366, 0.03126908874511719, 0.03128121566772461, 0.032555007934570314, 0.031068159103393556, 0.03102720069885254, 0.030674623489379882, 0.030873472213745118, 0.03096620750427246, 0.031001951217651365, 0.031462047576904295, 0.031440895080566404, 0.0313384952545166, 0.03135487937927246, 0.03130086326599121, 0.031492864608764645, 0.03129887962341309, 0.03163542366027832, 0.03129955291748047, 0.031156160354614257, 0.03147014427185059, 0.031333919525146484, 0.031103744506835937, 0.03105561637878418, 0.030732511520385742, 0.03088345527648926, 0.03110540771484375, 0.031098880767822266, 0.030988288879394532, 0.03125391960144043, 0.03131843185424805, 0.03395647811889648, 0.03156598472595215, 0.03160396766662597, 0.03127577590942383, 0.03112940788269043, 0.03162335968017578, 0.031346176147460936, 0.03095964813232422, 0.031176288604736327, 0.03092473602294922, 0.0308023681640625, 0.03069593620300293, 0.033898494720458985, 0.03160470390319824, 0.031188928604125976, 0.030807647705078125, 0.031191551208496093, 0.03123344039916992, 0.03143503952026367, 0.03143430328369141, 0.031374080657958985, 0.0314202880859375, 0.03162118339538574, 0.031266048431396486, 0.03120207977294922, 0.03147145652770996, 0.03131411170959473, 0.031085887908935548, 0.03092323112487793, 0.030941408157348634, 0.031380607604980466, 0.03133529663085938, 0.031133312225341797, 0.0309169921875, 0.030830591201782227, 0.030744575500488282, 0.030658559799194338, 0.03105120086669922, 0.030701791763305664, 0.030746976852416993, 0.03095919990539551, 0.03132246398925781, 0.03105183982849121, 0.03080147171020508, 0.030652864456176758, 0.031959039688110355, 0.03069308853149414, 0.030578975677490235, 0.030580480575561522, 0.030607616424560547, 0.03068083190917969, 0.031090944290161134, 0.03092889595031738, 0.03119215965270996, 0.03136588859558105, 0.030842432022094725, 0.03073699188232422, 0.030701568603515625, 0.030693376541137695, 0.030830240249633788, 0.031175008773803713, 0.03136860847473145, 0.031228511810302735, 0.03370697784423828, 0.03141734313964844, 0.03135456085205078, 0.03136748886108399, 0.03124336051940918, 0.03082326316833496, 0.03085443115234375, 0.031116064071655274, 0.03114188766479492, 0.031117311477661135, 0.031065887451171875, 0.031004735946655274, 0.030936447143554688, 0.030752767562866212, 0.03070572853088379, 0.03071251106262207, 0.030824415206909178, 0.030691295623779296, 0.03109622383117676, 0.031332128524780276, 0.031385951995849606, 0.03260678482055664, 0.03509862518310547, 0.03164723205566406, 0.03148441505432129, 0.03128319931030273, 0.031734783172607424, 0.03146009635925293, 0.031379680633544925, 0.03143888092041015, 0.031455232620239255, 0.031485151290893555, 0.031589151382446286, 0.03143270492553711, 0.03125862312316895, 0.033635425567626956, 0.03182480049133301, 0.031780672073364255, 0.03176057624816894, 0.03196108818054199, 0.0318525447845459, 0.031834112167358396, 0.033685504913330076, 0.03216707229614258, 0.03191484832763672, 0.032161792755126956, 0.03198566436767578, 0.03184435272216797, 0.032012161254882814, 0.03172681617736817, 0.031576992034912106, 0.031531007766723636, 0.031537023544311524, 0.031475231170654296, 0.03108268737792969, 0.031095199584960938, 0.031073856353759765, 0.031526912689208986, 0.03425750350952148, 0.03189743995666504, 0.0315863037109375, 0.03146751976013184, 0.031211328506469727, 0.034103294372558594, 0.03159654426574707, 0.03129948806762695, 0.031247711181640624, 0.0314270076751709, 0.031494464874267575, 0.03139536094665527, 0.031375839233398435, 0.031262144088745114, 0.03098876762390137, 0.03104316711425781, 0.030996288299560547, 0.0311213436126709, 0.031124223709106447, 0.031236095428466795, 0.03101286315917969, 0.03085312080383301, 0.030785375595092774, 0.03090652847290039, 0.03096575927734375, 0.030711807250976563, 0.03091632080078125, 0.031137855529785156, 0.03121379280090332, 0.031170560836791993, 0.030650367736816408, 0.030506464004516603, 0.03062633514404297, 0.03080396842956543, 0.030980096817016602, 0.031172607421875, 0.031049440383911133, 0.03095961570739746, 0.030976287841796873, 0.030875648498535156, 0.030939136505126953, 0.030926368713378907, 0.03093948745727539, 0.030834527969360353, 0.03068115234375, 0.03104364776611328, 0.030785696029663086, 0.030727712631225586, 0.030978527069091797, 0.03086479949951172, 0.03086409568786621, 0.03165961647033692, 0.03186307144165039, 0.03154710388183594, 0.031617311477661135, 0.03178886413574219, 0.03158582305908203, 0.031599264144897464, 0.03180496025085449, 0.032250335693359375, 0.031727231979370114, 0.03188931274414063, 0.03193900871276856, 0.03196854400634765, 0.0325557746887207, 0.03170918464660644, 0.031817728042602536, 0.031838207244873046, 0.03379836654663086, 0.031479711532592776, 0.031326303482055666, 0.031223743438720704, 0.031350847244262695, 0.03161292839050293, 0.031567264556884765, 0.031584447860717776, 0.03150886344909668, 0.03165750312805176, 0.0313492488861084, 0.03140812873840332, 0.03135676765441894, 0.031633119583129886, 0.03137580871582031, 0.031152128219604492, 0.03104732894897461, 0.031123039245605468, 0.03165056037902832, 0.031678335189819336, 0.031092863082885742, 0.030976095199584962, 0.03073219108581543, 0.030707712173461913, 0.030643648147583007, 0.03135750389099121, 0.030621696472167968, 0.03114726448059082, 0.030713760375976562, 0.03126153564453125, 0.031254528045654296, 0.03117670440673828, 0.03135487937927246, 0.0313093433380127, 0.03131030464172363, 0.031307775497436525, 0.031210943222045897, 0.03140665626525879, 0.031509855270385745, 0.031191295623779296, 0.031038944244384765, 0.03116713523864746, 0.0316275520324707, 0.03096575927734375, 0.030957311630249024, 0.03087299156188965, 0.030589792251586916, 0.03084671974182129, 0.03089638328552246, 0.030841983795166016, 0.030845632553100587, 0.031102783203125, 0.030741888046264647, 0.031175296783447267, 0.03141465568542481, 0.031145984649658204, 0.03135638427734375, 0.031384063720703126, 0.031514656066894534, 0.03128934478759766, 0.0314040641784668, 0.03162953567504883, 0.03122150421142578, 0.03393308639526367, 0.031508575439453124, 0.0314619197845459, 0.031218912124633787, 0.031371648788452146, 0.031297504425048826, 0.03255353546142578, 0.031887359619140625, 0.03153302383422851, 0.031383583068847656, 0.03139705657958984, 0.03148064041137695, 0.031352352142333985, 0.031246816635131836, 0.03128319931030273, 0.03116166305541992, 0.03130179214477539, 0.03136073684692383, 0.031231903076171876, 0.031365472793579104, 0.03149676895141602, 0.03138150405883789, 0.03132566452026367, 0.031446720123291014, 0.03125766372680664, 0.03132374382019043, 0.031567840576171874, 0.031393983840942385, 0.031162399291992188, 0.032578655242919925, 0.03157488059997558, 0.03134016036987305, 0.031318464279174806, 0.031029247283935548, 0.03077731132507324, 0.030867488861083984, 0.03107244873046875, 0.030803775787353514, 0.031198911666870117, 0.031740224838256836, 0.031155744552612306, 0.0311441593170166, 0.031125759124755858, 0.03123606491088867, 0.031064096450805663, 0.031041536331176758, 0.031248607635498048, 0.03110028839111328, 0.031025568008422853, 0.03126860809326172, 0.03094553565979004, 0.030772960662841797, 0.03089779281616211, 0.030763423919677735, 0.03096601676940918, 0.030919872283935546, 0.03096793556213379, 0.030866111755371094, 0.0310435848236084, 0.03122585678100586, 0.03118489646911621, 0.03135897636413574, 0.03318368148803711, 0.03370880126953125, 0.03143270492553711, 0.031115264892578126, 0.03125043106079101, 0.030922752380371094, 0.030916608810424805, 0.03132806396484375, 0.031244319915771486, 0.031383712768554686, 0.03155353546142578, 0.03131711959838867, 0.03134144020080566, 0.031532608032226565, 0.03123040008544922, 0.031092735290527345, 0.031117311477661135, 0.03121971130371094, 0.03111689567565918, 0.03126438331604004, 0.03142511940002441, 0.032388801574707034, 0.03143935966491699, 0.031531007766723636, 0.031559680938720705, 0.03143475151062012, 0.031268863677978515, 0.031492095947265625, 0.03159859275817871, 0.0321781120300293, 0.03144915199279785, 0.03134464073181152, 0.03145465660095215, 0.03142025566101074, 0.03184902381896973, 0.0313753604888916, 0.03159670448303223, 0.031219743728637697, 0.0317869758605957, 0.03455740737915039, 0.03144758415222168, 0.03147081565856934, 0.03150297546386719, 0.031446880340576175, 0.03136953544616699, 0.03131740760803223, 0.031004831314086913, 0.03083513641357422, 0.03123200035095215, 0.03138768005371094, 0.03133616065979004, 0.031371519088745116, 0.03215359878540039, 0.03302809524536133, 0.03146873664855957, 0.03154003143310547, 0.03142051124572754, 0.031354784011840824, 0.03139104080200195, 0.03177542304992676, 0.03168870353698731, 0.03465798568725586, 0.03160505676269531, 0.031291391372680666, 0.03422959899902344, 0.03273731231689453, 0.0315645751953125, 0.03142451286315918, 0.0316231689453125, 0.03148595237731933, 0.03142646408081055, 0.03160073661804199, 0.03159654426574707, 0.03150233650207519, 0.03172761535644531, 0.031582143783569334, 0.03168671989440918, 0.03170918464660644, 0.03155763244628906, 0.031808639526367186, 0.03231356811523438, 0.03207030487060547, 0.031913631439208986, 0.03171977615356445, 0.031667583465576174, 0.03168115234375, 0.03149955177307129, 0.031439584732055666, 0.03141548728942871, 0.03139769554138184, 0.03134502410888672, 0.03132019233703613, 0.0313492488861084, 0.03141836738586426, 0.031664127349853514, 0.031477567672729495, 0.03129267120361328, 0.03142304039001465, 0.03130982398986817, 0.031402368545532224, 0.031322111129760744, 0.03129958343505859, 0.031051136016845702, 0.03130633544921875, 0.031200576782226562, 0.031242143630981444, 0.031363136291503904, 0.031494911193847654, 0.031409280776977536, 0.03170582389831543, 0.03212249755859375, 0.031676959991455075, 0.03188662338256836, 0.031677152633666994, 0.03156377601623535, 0.03211468887329102, 0.03171439933776855, 0.03147276878356933, 0.031399711608886716, 0.03147747230529785, 0.031232288360595703, 0.031336448669433595, 0.031383039474487305, 0.031231935501098634, 0.031145824432373046, 0.03132080078125, 0.03178313636779785, 0.03428972625732422, 0.031711679458618164, 0.03147107124328613, 0.03144908714294434, 0.031643999099731444, 0.031697088241577145, 0.03154889678955078, 0.031453151702880856, 0.031647872924804685, 0.03142086410522461, 0.03134383964538574, 0.03145603179931641, 0.03141993522644043, 0.031441375732421874, 0.03152252769470215, 0.03130406379699707, 0.03132131195068359, 0.031183135986328124, 0.03161123275756836, 0.03152207946777344, 0.03135087966918945, 0.03143894386291504, 0.03145692825317383, 0.03145619201660156, 0.03163545608520508, 0.03151785659790039, 0.03150729560852051, 0.031776607513427736, 0.03159212875366211, 0.03170761680603027, 0.03212287902832031, 0.03167014312744141, 0.031838336944580076, 0.03182796859741211, 0.032481281280517575, 0.0319815673828125, 0.03202816009521484, 0.03167897605895996, 0.03164159965515137, 0.0317640323638916, 0.03174006462097168, 0.031448768615722655, 0.031506975173950194, 0.032005889892578125, 0.031450815200805664, 0.03152960014343262, 0.03179929542541504, 0.031645696640014646, 0.031553119659423826, 0.0318666877746582, 0.03148041534423828, 0.031550592422485355, 0.03173219108581543, 0.031647872924804685, 0.03158201599121094, 0.03161750411987305, 0.03157401657104492, 0.03165798377990723, 0.03160883140563965, 0.031643648147583005, 0.031647743225097655, 0.031510528564453126, 0.03181158447265625]",tokens/s,31.802853853477593,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1303.523328,1083.113472,0.0,704.643072,675.01056,s,1,8.4019462890625,8.4019462890625,0.0,8.4019462890625,8.4019462890625,8.4019462890625,8.4019462890625,[8.4019462890625],,kWh,3.635637827502857e-05,4.003097544512459e-06,1.1330009063975632e-05,5.1689484883516665e-05,,MB,1330.29888,1403.977728,0.0,996.1472,942.608384,s,10,0.3297423706054688,0.03297423706054688,0.000629820502509376,0.032741056442260746,0.03361081008911133,0.03414358177185058,0.03456979911804199,"[0.033492416381835935, 0.032814849853515626, 0.032725887298583986, 0.032740161895751956, 0.0329180793762207, 0.03271689605712891, 0.034676353454589845, 0.032493854522705076, 0.032421920776367186, 0.03274195098876953]",tokens/s,7763.636791047994,kWh,1.0881640219019337e-06,1.200033213602986e-07,7.203083003272397e-07,1.928475643589472e-06,tokens/kWh,132747333.80791223,MB,1348.374528,1412.366336,0.0,1004.535808,942.610944,s,10,18.513116699218752,1.8513116699218752,0.023257271581990157,1.8506577758789065,1.8760862670898437,1.888021649169922,1.8975699548339844,"[1.853917236328125, 1.8137567138671875, 1.85634521484375, 1.8734339599609375, 1.859286376953125, 1.89995703125, 1.8473983154296876, 1.8208701171875, 1.8445382080078125, 1.843613525390625]",tokens/s,34.029926469732985,kWh,5.3085660723922634e-05,5.8550354700993705e-06,2.1319292664671445e-05,8.025998885869346e-05,tokens/kWh,784949.0249857676,,s,630,18.510737812042247,0.029382123511178155,0.0012229306279341063,0.02936871910095215,0.02985663948059082,0.030057206535339356,0.03157573390960693,"[0.028767648696899413, 0.028725536346435546, 0.028883424758911133, 0.028831584930419922, 0.028948480606079102, 0.028816831588745116, 0.028691007614135743, 0.028680192947387696, 0.02928803253173828, 0.028837631225585938, 0.02898192024230957, 0.029050880432128907, 0.02945961570739746, 0.02886128044128418, 0.02919628715515137, 0.02906460762023926, 0.02924959945678711, 0.02922550392150879, 0.02917087936401367, 0.02905939292907715, 0.028994047164916992, 0.02905625534057617, 0.0289881591796875, 0.029278207778930664, 0.02925347137451172, 0.029429920196533205, 0.029216768264770508, 0.029077503204345705, 0.029198047637939453, 0.029296287536621092, 0.02955084800720215, 0.02940083122253418, 0.02930748748779297, 0.02929199981689453, 0.02954092788696289, 0.0293621768951416, 0.02929248046875, 0.029343807220458984, 0.029493152618408205, 0.02933564758300781, 0.03016841506958008, 0.02949964714050293, 0.029784479141235352, 0.029855104446411134, 0.029760128021240236, 0.0304616641998291, 0.03053919982910156, 0.02989695930480957, 0.030007871627807617, 0.029971744537353514, 0.03020262336730957, 0.02999407958984375, 0.029700096130371095, 0.029897632598876952, 0.029853696823120116, 0.029636608123779298, 0.029706047058105468, 0.030012639999389648, 0.02989529609680176, 0.029684032440185547, 0.029855743408203125, 0.029736991882324218, 0.02998476791381836, 0.0294932804107666, 0.02938409614562988, 0.02949305534362793, 0.029430559158325195, 0.02933456039428711, 0.029295936584472656, 0.028919456481933593, 0.028618175506591795, 0.028500543594360352, 0.02854911994934082, 0.02851430320739746, 0.028528383255004883, 0.02874150466918945, 0.02864780807495117, 0.028675264358520507, 0.028736480712890623, 0.028751712799072265, 0.028764320373535157, 0.028676095962524413, 0.028485471725463868, 0.02874777603149414, 0.028733440399169922, 0.028626943588256838, 0.028631040573120117, 0.028691808700561525, 0.028729312896728514, 0.028613311767578125, 0.028515871047973634, 0.028633567810058595, 0.028516351699829103, 0.02855731201171875, 0.02933318328857422, 0.028815807342529295, 0.02877017593383789, 0.028647232055664062, 0.02871062469482422, 0.028643808364868163, 0.028817087173461913, 0.028846399307250976, 0.02893840026855469, 0.028620031356811522, 0.02867465591430664, 0.028647104263305665, 0.028639551162719726, 0.02853068733215332, 0.028893184661865235, 0.028903423309326173, 0.028620800018310546, 0.028692480087280273, 0.028575935363769532, 0.029648448944091795, 0.028958976745605467, 0.028868608474731446, 0.02872319984436035, 0.028755008697509767, 0.028952735900878906, 0.028740095138549804, 0.02864156723022461, 0.028635135650634767, 0.028790111541748046, 0.028741600036621094, 0.028614944458007812, 0.028635360717773437, 0.02870822334289551, 0.02881827163696289, 0.02895235252380371, 0.028876800537109375, 0.02876825523376465, 0.028794879913330077, 0.0287673282623291, 0.028591007232666017, 0.02879443168640137, 0.029009536743164064, 0.02947974395751953, 0.03056844711303711, 0.029394304275512696, 0.029907392501831054, 0.028948671340942384, 0.028737535476684572, 0.02886000061035156, 0.02877020835876465, 0.02928428840637207, 0.02937094306945801, 0.029122047424316407, 0.02913484764099121, 0.029334016799926758, 0.028948768615722657, 0.02905673599243164, 0.029222944259643554, 0.029455999374389648, 0.029151456832885742, 0.029460607528686525, 0.029566976547241212, 0.02950713539123535, 0.029747840881347656, 0.029542207717895508, 0.029491199493408202, 0.029587007522583007, 0.029663007736206056, 0.029576992034912108, 0.029620256423950195, 0.030034784317016602, 0.029689855575561523, 0.029612031936645508, 0.029621376037597656, 0.029500288009643556, 0.02953215980529785, 0.02953625679016113, 0.02993062400817871, 0.029444959640502928, 0.029476287841796876, 0.029286624908447266, 0.02931455993652344, 0.029468639373779297, 0.029610912322998048, 0.03031449508666992, 0.032233470916748046, 0.029894655227661132, 0.03143827247619629, 0.029747264862060547, 0.029739007949829102, 0.029495935440063476, 0.029877824783325194, 0.029491487503051757, 0.029552415847778322, 0.029671072006225586, 0.0295316162109375, 0.029436447143554687, 0.029471839904785156, 0.0295281925201416, 0.029670080184936522, 0.02957276725769043, 0.029469120025634767, 0.029782079696655275, 0.02949318313598633, 0.02953215980529785, 0.029423519134521483, 0.029779455184936524, 0.029502111434936522, 0.029550527572631834, 0.029566240310668946, 0.02948089599609375, 0.029600543975830076, 0.029515487670898437, 0.029607295989990234, 0.029658016204833985, 0.029624160766601564, 0.029802656173706053, 0.029566816329956055, 0.02952822494506836, 0.02955264091491699, 0.029706239700317383, 0.02972876739501953, 0.029515775680541992, 0.029586816787719728, 0.02978879928588867, 0.02951315116882324, 0.029571136474609374, 0.029573631286621094, 0.0295314884185791, 0.02958812713623047, 0.030703039169311525, 0.030691904067993166, 0.029748544692993165, 0.029891103744506837, 0.02964233589172363, 0.029957759857177736, 0.029565696716308595, 0.029704383850097656, 0.02980659294128418, 0.02989583969116211, 0.029924192428588868, 0.029490367889404297, 0.029613887786865235, 0.02984163284301758, 0.029651744842529297, 0.031116704940795898, 0.030810367584228514, 0.029875904083251952, 0.029304927825927734, 0.02931769561767578, 0.029249536514282228, 0.029031904220581054, 0.030568256378173828, 0.030089471817016603, 0.029699871063232422, 0.03127555274963379, 0.02993987274169922, 0.029437183380126953, 0.029253503799438477, 0.028942176818847656, 0.028889280319213867, 0.02884150314331055, 0.028747455596923828, 0.029227872848510743, 0.029655168533325196, 0.02932316780090332, 0.029196544647216795, 0.029333248138427734, 0.029509632110595704, 0.029558784484863283, 0.029396991729736328, 0.02934809684753418, 0.029265663146972657, 0.029198335647583007, 0.029289663314819334, 0.029224992752075196, 0.029117216110229494, 0.029392959594726563, 0.029624223709106445, 0.029722591400146485, 0.0296694393157959, 0.029619712829589844, 0.02954051208496094, 0.029628768920898437, 0.029711551666259765, 0.029548448562622072, 0.029526943206787108, 0.029616128921508788, 0.02953011131286621, 0.02975030326843262, 0.02929478454589844, 0.029422143936157226, 0.029458656311035156, 0.0294335994720459, 0.029391008377075194, 0.02940732765197754, 0.02954787254333496, 0.029463167190551757, 0.029582592010498048, 0.02954524803161621, 0.029535327911376953, 0.02947727966308594, 0.02966988754272461, 0.02964614486694336, 0.029774688720703126, 0.029720415115356447, 0.029634048461914062, 0.029938175201416017, 0.029701919555664064, 0.029575391769409178, 0.029814720153808594, 0.029603519439697266, 0.029646623611450196, 0.029679647445678713, 0.029553216934204103, 0.02962611198425293, 0.029829376220703124, 0.029612031936645508, 0.030300384521484376, 0.030142240524291992, 0.02979430389404297, 0.029786239624023436, 0.02983795166015625, 0.03049033546447754, 0.02978937530517578, 0.029664127349853516, 0.029669376373291017, 0.02976883125305176, 0.02986470413208008, 0.029694080352783203, 0.02977177619934082, 0.029727935791015625, 0.02966192054748535, 0.02958345603942871, 0.029663232803344725, 0.029689855575561523, 0.029708288192749024, 0.056118751525878904, 0.029751840591430663, 0.0297574405670166, 0.029654720306396484, 0.029544767379760743, 0.029511680603027345, 0.02939641571044922, 0.02932793617248535, 0.030744575500488282, 0.029437376022338868, 0.029295167922973632, 0.02937651252746582, 0.029884544372558594, 0.02979827117919922, 0.029634559631347656, 0.02945142364501953, 0.02968047904968262, 0.02952396774291992, 0.029628000259399413, 0.029942176818847657, 0.029693056106567382, 0.029690336227416993, 0.029780384063720702, 0.029652992248535157, 0.029876224517822264, 0.029741056442260744, 0.029666656494140624, 0.030075551986694336, 0.030234624862670898, 0.029716480255126954, 0.02976972770690918, 0.029558784484863283, 0.029620223999023438, 0.02985321617126465, 0.02968806457519531, 0.02979043197631836, 0.02952396774291992, 0.0294783992767334, 0.029592063903808592, 0.029949951171875, 0.029633695602416993, 0.029616735458374024, 0.029671680450439452, 0.02992937660217285, 0.030313983917236328, 0.029833824157714843, 0.029910911560058595, 0.02978160095214844, 0.02958812713623047, 0.029869855880737303, 0.02975152015686035, 0.029549760818481444, 0.02966815948486328, 0.02957107162475586, 0.029525503158569336, 0.02952038383483887, 0.029398975372314454, 0.029613407135009765, 0.029682304382324217, 0.029595743179321288, 0.02941472053527832, 0.0291843204498291, 0.029131135940551757, 0.029345792770385744, 0.028943775177001953, 0.029055583953857423, 0.029128639221191407, 0.02897926330566406, 0.029212671279907225, 0.029474815368652343, 0.029464128494262696, 0.02966147232055664, 0.02951587104797363, 0.029544511795043946, 0.029577215194702147, 0.029454336166381836, 0.029480960845947264, 0.029560768127441406, 0.02967558479309082, 0.029591615676879884, 0.029703840255737305, 0.029513055801391602, 0.029467584609985352, 0.02975129508972168, 0.02970419120788574, 0.029799583435058594, 0.029756256103515625, 0.02960972785949707, 0.02972431945800781, 0.029567583084106445, 0.0295133113861084, 0.02945680046081543, 0.029241344451904298, 0.029264896392822266, 0.02998147201538086, 0.02936649513244629, 0.02922064018249512, 0.02894460868835449, 0.028651519775390624, 0.028846015930175783, 0.028540512084960938, 0.028616640090942384, 0.028563999176025392, 0.028605567932128907, 0.028563583374023437, 0.02855193519592285, 0.028590080261230468, 0.028446720123291015, 0.02856550407409668, 0.028556991577148437, 0.02855526351928711, 0.028661760330200195, 0.028829599380493166, 0.029269567489624025, 0.02881795120239258, 0.028657920837402345, 0.028601472854614257, 0.028951168060302734, 0.02876150321960449, 0.02906502342224121, 0.028575775146484374, 0.029484832763671875, 0.02884444808959961, 0.02873353576660156, 0.02873289680480957, 0.02869964790344238, 0.028680192947387696, 0.028669952392578125, 0.028849920272827147, 0.028719423294067382, 0.028994943618774412, 0.028765920639038087, 0.028805984497070312, 0.028833791732788085, 0.029142368316650392, 0.028918432235717773, 0.028762048721313476, 0.028867712020874025, 0.03230160140991211, 0.029228799819946288, 0.0290199031829834, 0.028844959259033204, 0.028884288787841796, 0.028736064910888672, 0.028663936614990233, 0.03196086311340332, 0.029190208435058592, 0.029009151458740234, 0.028922624588012695, 0.02872470474243164, 0.028543615341186525, 0.028526527404785156, 0.02873334312438965, 0.02881558418273926, 0.02882262420654297, 0.028820383071899415, 0.028559104919433594, 0.028582143783569335, 0.02857369613647461, 0.02852454376220703, 0.029298688888549803, 0.028636703491210936, 0.028629472732543945, 0.02873910331726074, 0.028673824310302735, 0.02874361610412598, 0.02864204788208008, 0.028684288024902343, 0.028635135650634767, 0.028708864212036132, 0.02884003257751465, 0.02856857681274414, 0.028627552032470704, 0.028678144454956055, 0.028610559463500978, 0.02940835189819336, 0.028757984161376954, 0.029036735534667967, 0.028530912399291994, 0.02878220748901367, 0.029135103225708007, 0.031590656280517576, 0.030779808044433594, 0.028827648162841796, 0.028570911407470704, 0.028731359481811523, 0.028635904312133788, 0.02921062469482422, 0.028778495788574218, 0.02870195198059082, 0.028444896697998046, 0.0285283203125, 0.02848182487487793, 0.028691007614135743, 0.028572799682617188, 0.028676992416381837, 0.028554943084716795, 0.02855353546142578, 0.02896281623840332, 0.028825599670410155, 0.028962047576904296, 0.02915001678466797, 0.02904672050476074, 0.02896447944641113, 0.028825120925903322, 0.028754783630371095, 0.028811264038085937, 0.028677215576171877, 0.028666784286499023, 0.028645376205444335, 0.029060447692871094, 0.029295263290405275, 0.029495296478271486, 0.029284351348876952, 0.02927743911743164, 0.02935475158691406, 0.02941276741027832, 0.02903071975708008, 0.029052608489990233, 0.0288753604888916, 0.029146976470947265, 0.029575328826904296, 0.029904895782470704, 0.02969599914550781, 0.02955673599243164, 0.029597343444824217, 0.029276512145996095, 0.030193023681640625, 0.03394009780883789, 0.029671424865722655, 0.029439008712768555, 0.02939798355102539, 0.029756959915161134, 0.03422256088256836, 0.0295996150970459, 0.029621856689453125, 0.02929484748840332, 0.029556575775146483, 0.0290863037109375, 0.02901452827453613, 0.02889289665222168, 0.03153919982910156, 0.029198432922363283, 0.029204383850097656, 0.029108224868774416, 0.029732864379882814, 0.02956470489501953, 0.02956924819946289, 0.02998886489868164, 0.029845504760742186, 0.02941542434692383, 0.02939084815979004, 0.029468671798706055, 0.029347072601318358, 0.029451007843017577, 0.029298688888549803, 0.029403135299682616, 0.029337440490722656, 0.029405344009399415, 0.029337600708007814, 0.02931622314453125, 0.02918684768676758, 0.02903049659729004, 0.029120223999023438, 0.029248863220214843, 0.029399391174316405, 0.02929929542541504, 0.02931711959838867, 0.028997631072998048, 0.029310976028442383, 0.02926748847961426, 0.029182432174682617, 0.02920038414001465, 0.03021356773376465, 0.029272031784057618, 0.029350496292114257, 0.029351455688476562, 0.029309215545654296, 0.029218656539916992, 0.029290847778320313, 0.029265920639038087, 0.029263391494750976, 0.029177824020385743, 0.029099872589111328, 0.02893276786804199, 0.029087648391723633, 0.028843456268310547, 0.028854944229125976, 0.02911027145385742, 0.028751615524291993, 0.028799232482910157, 0.02866511917114258, 0.028626720428466798, 0.028600671768188476, 0.028682144165039062, 0.028983552932739257, 0.029071487426757813, 0.029044416427612303, 0.02918412780761719]",tokens/s,34.03429978842609,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTJForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3798, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ self.model = DeciCoderModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ self.self_attn = DeciCoderAttention(config=config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ self._init_rope() File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1709, in __getattr__ raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2187.812864,2479.816704,0.0,2101.346304,1978.345472,s,1,8.974201171875,8.974201171875,0.0,8.974201171875,8.974201171875,8.974201171875,8.974201171875,[8.974201171875],,kWh,5.411849286667044e-05,5.962155132541093e-06,1.7299180505997436e-05,7.737982850520897e-05,,MB,2176.872448,2775.515136,0.0,2367.684608,2241.29024,s,10,1.5665681762695316,0.15665681762695313,0.0006752935553594327,0.1566033477783203,0.1574509567260742,0.15768791122436523,0.15787747482299805,"[0.15792486572265624, 0.1566112060546875, 0.15604762268066405, 0.15642784118652345, 0.15707344055175781, 0.15613334655761718, 0.15659548950195312, 0.15544761657714845, 0.15739830017089843, 0.156908447265625]",tokens/s,1634.1452857137235,kWh,4.813034964958767e-06,5.307966585377469e-07,3.2062001970166566e-06,8.55003182051317e-06,tokens/kWh,29941409.0349707,MB,2183.655424,2859.401216,0.0,2451.570688,2334.829056,s,10,19.94945349121094,1.9949453491210936,0.014159525420776585,1.9951657104492186,2.0117086303710936,2.0154676574707033,2.0184748791503906,"[2.010873291015625, 2.0029676513671877, 1.993408447265625, 2.0192266845703126, 1.988684326171875, 1.99708642578125, 1.9908304443359375, 1.9644140625, 1.9850391845703126, 1.9969229736328125]",tokens/s,31.579812463411944,kWh,5.700232462295954e-05,6.287186092617085e-06,3.092157710517864e-05,9.421108782075526e-05,tokens/kWh,668711.0982081317,,s,630,19.946422578811628,0.03166098822033595,0.0006018373266020355,0.03161563205718994,0.03217015266418457,0.03240320053100586,0.033214518318176275,"[0.032889984130859376, 0.03231804656982422, 0.032000286102294925, 0.03188083267211914, 0.03180768013000488, 0.03148726463317871, 0.031736288070678714, 0.03147599983215332, 0.03157417678833008, 0.031692384719848636, 0.031965599060058594, 0.03184611129760742, 0.03229520034790039, 0.032833534240722655, 0.03225190353393555, 0.03208294296264649, 0.032471519470214844, 0.0320599365234375, 0.03210649490356445, 0.03217168045043945, 0.03204265594482422, 0.03194256019592285, 0.032318241119384764, 0.03223756790161133, 0.03223875045776367, 0.03191689682006836, 0.031968704223632814, 0.03215008163452148, 0.032366592407226565, 0.031762304306030276, 0.0318670711517334, 0.0321860466003418, 0.03219232177734375, 0.03172831916809082, 0.03186662483215332, 0.03156959915161133, 0.03164396858215332, 0.03164883232116699, 0.03288345718383789, 0.033003711700439455, 0.03191398429870605, 0.031528959274291994, 0.03140630340576172, 0.031825599670410154, 0.03306505584716797, 0.03144864082336426, 0.031617599487304686, 0.03117430305480957, 0.03154483222961426, 0.03125888061523437, 0.03157449531555176, 0.031302911758422855, 0.032250625610351566, 0.03278643035888672, 0.03172319984436035, 0.03151228713989258, 0.03201273727416992, 0.031408287048339846, 0.03126681518554687, 0.03168012809753418, 0.03135894393920898, 0.03103785514831543, 0.03141190338134765, 0.0320516471862793, 0.03142223930358887, 0.031312639236450196, 0.03241164779663086, 0.031774688720703125, 0.0315794563293457, 0.03139673614501953, 0.03135385513305664, 0.031564640045166015, 0.031143936157226562, 0.03139910316467285, 0.031166559219360353, 0.03119900894165039, 0.031224096298217773, 0.031252960205078124, 0.03114112091064453, 0.03128211212158203, 0.03189555168151856, 0.032024574279785153, 0.031817728042602536, 0.03196649551391602, 0.03175907135009766, 0.032231422424316404, 0.032233470916748046, 0.031903743743896484, 0.03184409523010254, 0.03180159950256348, 0.03166524887084961, 0.031528863906860355, 0.03175472068786621, 0.03184614372253418, 0.03161110305786133, 0.03158006477355957, 0.031179071426391602, 0.031271263122558596, 0.031145984649658204, 0.03160883140563965, 0.03139097595214844, 0.03183283233642578, 0.03181711959838867, 0.03220873641967773, 0.031924640655517575, 0.0318460807800293, 0.03189827156066895, 0.031948320388793944, 0.031877344131469726, 0.032013790130615234, 0.03183286476135254, 0.03205961608886719, 0.03216531372070312, 0.032037216186523436, 0.03172700881958008, 0.03183638381958008, 0.032077728271484376, 0.03259235382080078, 0.03216998291015625, 0.03221263885498047, 0.032050529479980466, 0.03239158248901367, 0.032449119567871096, 0.032185630798339845, 0.03230563354492187, 0.03240300750732422, 0.032710304260253904, 0.032088512420654296, 0.032403358459472655, 0.03194675254821777, 0.03172099113464356, 0.03144956779479981, 0.03147724723815918, 0.03111577606201172, 0.03134390449523926, 0.030922752380371094, 0.03104755210876465, 0.03083964729309082, 0.03139379119873047, 0.031079647064208984, 0.03140892791748047, 0.03102902412414551, 0.031275232315063475, 0.03133609580993652, 0.03153545570373535, 0.03114188766479492, 0.031251935958862304, 0.031160863876342773, 0.03173686408996582, 0.03124496078491211, 0.03134086418151855, 0.03130118370056152, 0.031734144210815426, 0.03164985656738281, 0.03160204887390137, 0.031506303787231446, 0.03137107276916504, 0.03164051246643067, 0.03151158332824707, 0.031689695358276364, 0.03194844818115234, 0.031983968734741214, 0.03188236808776856, 0.03173670387268066, 0.03195699119567871, 0.03188556861877442, 0.03212988662719726, 0.03174614334106445, 0.03205612945556641, 0.0316231689453125, 0.03198361587524414, 0.0319815673828125, 0.03210361480712891, 0.03160668754577637, 0.03170806312561035, 0.03172537612915039, 0.0322841911315918, 0.031632032394409176, 0.03168460845947266, 0.031541248321533204, 0.03170918464660644, 0.031415807723999024, 0.03193497657775879, 0.03174195289611816, 0.03186870384216309, 0.03175036811828613, 0.03184774398803711, 0.03177337646484375, 0.03182387161254883, 0.03248796844482422, 0.03185462379455566, 0.032111038208007814, 0.031784959793090824, 0.03181158447265625, 0.03153305625915527, 0.03158815956115723, 0.03166022491455078, 0.041218048095703126, 0.03204198455810547, 0.03200511932373047, 0.03180694389343262, 0.03221673583984375, 0.03204800033569336, 0.03209759902954101, 0.03538201522827149, 0.03205545425415039, 0.0320261116027832, 0.032129310607910154, 0.03189948844909668, 0.032084129333496095, 0.03196640014648437, 0.03245344161987305, 0.031673759460449216, 0.03173814392089844, 0.031669631958007816, 0.03206601715087891, 0.03177110481262207, 0.032072799682617184, 0.0323199691772461, 0.03193900871276856, 0.03174195289611816, 0.0318659839630127, 0.031637920379638675, 0.03195337677001953, 0.032137279510498044, 0.03220217514038086, 0.03194259262084961, 0.03203539276123047, 0.031889408111572266, 0.03210649490356445, 0.03163036727905273, 0.031515775680541994, 0.03142195129394531, 0.03302435302734375, 0.031244031906127928, 0.03127935981750488, 0.03134230422973633, 0.03149648094177246, 0.03164159965515137, 0.03170918464660644, 0.031641504287719724, 0.03166745567321777, 0.031555551528930664, 0.031656831741333005, 0.03160268783569336, 0.031473663330078124, 0.03130150413513184, 0.03137548828125, 0.03153670310974121, 0.03194207954406738, 0.03175699234008789, 0.03205955123901367, 0.03352163314819336, 0.03171737670898438, 0.031535327911376955, 0.03138742446899414, 0.031633535385131833, 0.0313809928894043, 0.031387807846069336, 0.03116054344177246, 0.03162300872802734, 0.031510528564453126, 0.03174963188171387, 0.03134927940368652, 0.03131347274780273, 0.03139027214050293, 0.03205036926269531, 0.03182675170898437, 0.03157366371154785, 0.031236448287963868, 0.03125651168823242, 0.03104956817626953, 0.0310798397064209, 0.03093587112426758, 0.031126880645751955, 0.03157600021362305, 0.031248735427856444, 0.031054208755493164, 0.031205375671386718, 0.030864671707153322, 0.03111599922180176, 0.03098975944519043, 0.03127292823791504, 0.03108105659484863, 0.03218851089477539, 0.030863264083862304, 0.03160016059875488, 0.031308256149291995, 0.03185849571228027, 0.0318670711517334, 0.03250790405273438, 0.03242393493652344, 0.03208201599121094, 0.03176972770690918, 0.03194345664978027, 0.03186252784729004, 0.031770719528198245, 0.03154496002197266, 0.03159708786010742, 0.03141430473327637, 0.03170265579223633, 0.03161737632751465, 0.03163459205627441, 0.031344831466674807, 0.0313350715637207, 0.031285247802734374, 0.03157401657104492, 0.03162851142883301, 0.03161167907714844, 0.03160204887390137, 0.032131744384765626, 0.03208188629150391, 0.03188281631469726, 0.03161065673828125, 0.03155167961120606, 0.03256447982788086, 0.032025344848632814, 0.03206758499145508, 0.03193014335632324, 0.03192860794067383, 0.0318033275604248, 0.0317270393371582, 0.03138319969177246, 0.03149203109741211, 0.031928447723388674, 0.031505279541015625, 0.03126473617553711, 0.031331647872924806, 0.03186963272094727, 0.03286220932006836, 0.031472896575927736, 0.03151747131347656, 0.030986207962036133, 0.03147792053222656, 0.03167625617980957, 0.031890911102294923, 0.03151107215881348, 0.031696895599365234, 0.03155958366394043, 0.03179718399047852, 0.031819328308105466, 0.031826528549194336, 0.03327142333984375, 0.033505374908447266, 0.03192655944824219, 0.031780864715576174, 0.0315861759185791, 0.03195097541809082, 0.03209820938110351, 0.03337424087524414, 0.031807519912719726, 0.03151055908203125, 0.03127840042114258, 0.031344383239746094, 0.031214111328125, 0.03138368034362793, 0.03136745643615722, 0.031514623641967776, 0.03165593528747559, 0.03165388870239258, 0.03134998321533203, 0.0313526725769043, 0.031090784072875976, 0.031222623825073244, 0.031223392486572264, 0.031449504852294925, 0.03183843231201172, 0.03134169578552246, 0.03188969612121582, 0.031234432220458984, 0.031064064025878906, 0.03130316734313965, 0.03128780746459961, 0.03123200035095215, 0.03140812873840332, 0.03164892768859863, 0.031806400299072266, 0.03188435173034668, 0.03270057678222656, 0.031624063491821286, 0.031741920471191405, 0.031459327697753905, 0.03191110420227051, 0.03135980796813965, 0.03135487937927246, 0.031090688705444337, 0.031180192947387695, 0.031179359436035156, 0.031287296295166016, 0.0310515193939209, 0.031217920303344728, 0.03138355255126953, 0.03224371337890625, 0.03166982460021973, 0.03242438507080078, 0.031666112899780274, 0.031952959060668945, 0.031853567123413085, 0.032008224487304685, 0.03176752090454102, 0.03181977653503418, 0.03160444831848144, 0.031781152725219725, 0.03140108871459961, 0.031607679367065426, 0.03149619293212891, 0.03159568023681641, 0.032191326141357425, 0.032048351287841795, 0.03130652809143066, 0.031254240036010746, 0.03125833511352539, 0.03118342399597168, 0.031369216918945314, 0.031627264022827145, 0.03176860809326172, 0.0317620792388916, 0.03175446319580078, 0.03200009536743164, 0.03168608093261719, 0.03179782485961914, 0.031821088790893556, 0.03191456031799316, 0.03219827270507813, 0.031989664077758787, 0.03145750427246094, 0.03142051124572754, 0.03125689506530762, 0.03151203155517578, 0.03145577621459961, 0.03125388717651367, 0.03108425521850586, 0.03117558479309082, 0.030971904754638672, 0.031270912170410156, 0.030935039520263673, 0.03140140724182129, 0.03165446472167969, 0.03205843353271484, 0.03181868743896484, 0.031460351943969726, 0.03174723243713379, 0.03139788818359375, 0.03138851165771484, 0.03146732711791992, 0.031549568176269534, 0.03133856010437012, 0.031479263305664064, 0.03116659164428711, 0.03124393653869629, 0.03169740867614746, 0.031146240234375, 0.031021055221557618, 0.031223487854003907, 0.031137952804565428, 0.031195295333862304, 0.031101951599121092, 0.0313246078491211, 0.031085248947143554, 0.03108233642578125, 0.031182880401611327, 0.03161497688293457, 0.03123788833618164, 0.03127097511291504, 0.03107244873046875, 0.031178144454956053, 0.03164143943786621, 0.03133097648620606, 0.031161792755126955, 0.03139667129516602, 0.031110591888427734, 0.031179168701171874, 0.030974079132080078, 0.03116428756713867, 0.03099238395690918, 0.03113926315307617, 0.03094175910949707, 0.030892032623291016, 0.030887935638427736, 0.030896127700805662, 0.030850496292114258, 0.030906944274902343, 0.03144636726379395, 0.031074975967407225, 0.031045631408691408, 0.031031295776367186, 0.030877151489257813, 0.031005216598510743, 0.030829599380493164, 0.031122400283813478, 0.03096575927734375, 0.03100054359436035, 0.030932064056396483, 0.03117900848388672, 0.03202732849121094, 0.031909280776977536, 0.030955135345458986, 0.03104867172241211, 0.030963712692260743, 0.031059648513793944, 0.030936607360839842, 0.031062816619873045, 0.030930944442749023, 0.03093708801269531, 0.0320662727355957, 0.031303680419921875, 0.03135078430175781, 0.031385408401489255, 0.0318383674621582, 0.03221059036254883, 0.03252467346191406, 0.031616287231445314, 0.031945215225219724, 0.031704288482666015, 0.03186108779907226, 0.03163759994506836, 0.03174822425842285, 0.03157036781311035, 0.03180127906799316, 0.03153696060180664, 0.03211430358886719, 0.03223820877075195, 0.03165184020996094, 0.03127033615112305, 0.03135487937927246, 0.031594112396240236, 0.0313578872680664, 0.031145984649658204, 0.03160883140563965, 0.03177471923828125, 0.03178662490844727, 0.03144921684265137, 0.0313940486907959, 0.031185056686401365, 0.031293279647827146, 0.031178752899169923, 0.03126067161560059, 0.03103539276123047, 0.03132620811462403, 0.031190303802490233, 0.03142495918273926, 0.031754079818725585, 0.03128713607788086, 0.031230560302734377, 0.03132825660705566, 0.0311778564453125, 0.031248672485351563, 0.03141894340515137, 0.03331689453125, 0.031940607070922854, 0.031715616226196286, 0.03112291145324707, 0.03163804817199707, 0.031210559844970703, 0.0312961597442627, 0.031116447448730468, 0.031277471542358395, 0.031098976135253906, 0.031155616760253906, 0.031037567138671875, 0.0313558406829834, 0.031100799560546875, 0.031264768600463864, 0.031071807861328123, 0.031144384384155274, 0.031129056930541993, 0.031562400817871095, 0.0324956169128418, 0.033075199127197266, 0.03180518341064453, 0.031428255081176755, 0.03145171165466309, 0.03133852767944336, 0.03139529609680176, 0.031187488555908204, 0.031514623641967776, 0.030951103210449218, 0.030934688568115234, 0.031108991622924805, 0.03125737571716308, 0.031024383544921874, 0.031035999298095703, 0.031266975402832034, 0.03188889694213867, 0.03237529754638672, 0.03182089614868164, 0.031609760284423825, 0.031866880416870115, 0.03184025573730469, 0.03196246337890625, 0.031812000274658206, 0.03192857551574707, 0.03195449638366699, 0.03197587203979492, 0.03157740783691406, 0.031401792526245115, 0.03140083122253418, 0.031664127349853514, 0.03185971260070801, 0.031908607482910155, 0.03165996742248535, 0.03171360015869141, 0.03141222381591797, 0.03149004745483398, 0.03135078430175781, 0.03132620811462403, 0.031393152236938476, 0.031576704025268555, 0.031991519927978516, 0.03199551963806152, 0.03201103973388672, 0.03211251068115235, 0.03190390396118164, 0.03200579071044922, 0.03183760070800781, 0.032088863372802735, 0.03197120094299316, 0.03193628883361816, 0.03182150459289551, 0.03177481651306152, 0.03142918395996094, 0.03153408050537109, 0.03170816040039062, 0.03158016014099121, 0.03140812873840332, 0.031290752410888675, 0.032989822387695315, 0.03220585632324219, 0.03144598388671875, 0.031547391891479495]",tokens/s,31.584611100600355,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,7394.59072,7808.679936,0.0,7430.209536,7414.23104,s,1,11.9593759765625,11.9593759765625,0.0,11.9593759765625,11.9593759765625,11.9593759765625,11.9593759765625,[11.9593759765625],,kWh,0.00013767597920414496,1.517940819217984e-05,4.538892519999338e-05,0.00019824431259631817,,MB,1693.106176,8574.140416,0.0,8166.309888,8044.111872,s,10,8.038221557617188,0.8038221557617188,0.005012772321141111,0.8047204284667968,0.8073851684570312,0.8079228393554687,0.8083529760742187,"[0.789686279296875, 0.8072656860351562, 0.8084605102539062, 0.806500732421875, 0.805220947265625, 0.8042199096679687, 0.8065791625976563, 0.8036324462890625, 0.8027498168945313, 0.8039060668945313]",tokens/s,318.4784074002153,kWh,2.3507717869551923e-05,2.5917034602081007e-06,1.5627469766923974e-05,4.1726891096684e-05,tokens/kWh,6135132.3636077,MB,1696.731136,8741.912576,0.0,8334.082048,8265.758208,s,10,33.337338378906246,3.3337338378906254,0.005705343794965879,3.3318209228515627,3.3361259033203123,3.3430854125976563,3.348653020019531,"[3.350044921875, 3.329379150390625, 3.331510498046875, 3.331429931640625, 3.33213134765625, 3.32924072265625, 3.331159912109375, 3.333744873046875, 3.334579345703125, 3.33411767578125]",tokens/s,18.89772941197441,kWh,9.74628008175307e-05,1.0751531806857375e-05,6.470653894467688e-05,0.00017292087156906493,tokens/kWh,364328.48983668047,,s,630,33.33370851516723,0.052910648436773386,0.0008905385000460291,0.05287142372131348,0.05350011253356934,0.05368690528869629,0.05574180805206299,"[0.05570675277709961, 0.05300624084472656, 0.05237977600097656, 0.05226287841796875, 0.05208563232421875, 0.052071582794189455, 0.052112064361572265, 0.052238494873046874, 0.05226816177368164, 0.052148769378662106, 0.05244483184814453, 0.05241468811035156, 0.05248764801025391, 0.05235331344604492, 0.05242134475708008, 0.0526429443359375, 0.05261321640014648, 0.052783905029296876, 0.05306777572631836, 0.05314870452880859, 0.053048065185546875, 0.052811870574951174, 0.06940009307861328, 0.051090015411376956, 0.052620704650878904, 0.05255014419555664, 0.05237155151367188, 0.052563968658447265, 0.052725215911865235, 0.05257664108276367, 0.05271075057983399, 0.05247673416137695, 0.052622657775878906, 0.052935489654541014, 0.05291737747192383, 0.053076320648193356, 0.05301683044433594, 0.053219486236572265, 0.05346713638305664, 0.05408358383178711, 0.05385420989990235, 0.05309417724609375, 0.05304054260253906, 0.052920639038085936, 0.05280752182006836, 0.0528267822265625, 0.052838401794433595, 0.0529323501586914, 0.052867328643798825, 0.05304489517211914, 0.052897216796875, 0.05303388977050781, 0.053034496307373044, 0.053230079650878906, 0.05322547149658203, 0.05331308746337891, 0.0533520622253418, 0.053750591278076174, 0.054042625427246097, 0.05389904022216797, 0.05370806503295898, 0.053513153076171875, 0.05337241744995117, 0.05586240005493164, 0.052767616271972656, 0.0519659538269043, 0.0521662712097168, 0.05222643280029297, 0.05204991912841797, 0.052055904388427734, 0.05190598297119141, 0.05198076629638672, 0.0521710090637207, 0.05203523254394531, 0.0522193603515625, 0.05218598556518555, 0.05230742263793945, 0.05228188705444336, 0.05264543914794922, 0.05249683380126953, 0.05298995208740234, 0.053110752105712894, 0.05305510330200195, 0.05286953735351563, 0.05267583847045899, 0.0530536003112793, 0.052284000396728515, 0.05268070220947266, 0.052886783599853514, 0.05301103973388672, 0.052374687194824215, 0.05219126510620117, 0.05222067260742187, 0.05233843231201172, 0.0524169921875, 0.052625343322753905, 0.05276652908325195, 0.052857086181640624, 0.05330099105834961, 0.05318988800048828, 0.05343948745727539, 0.05321932983398438, 0.053190654754638675, 0.05326383972167969, 0.05290399932861328, 0.05293686294555664, 0.052963649749755856, 0.05283020782470703, 0.05283225631713867, 0.05288140869140625, 0.05292828750610352, 0.052607200622558595, 0.05268889617919922, 0.052821151733398436, 0.05313827133178711, 0.05312464141845703, 0.05342051315307617, 0.05349526214599609, 0.053391902923583985, 0.053714942932128903, 0.0538419189453125, 0.053319263458251956, 0.05348956680297851, 0.053314048767089846, 0.05347942352294922, 0.053510143280029294, 0.05607964706420898, 0.052924575805664065, 0.05240275192260742, 0.05188531112670899, 0.052144447326660154, 0.05229350280761719, 0.052294208526611326, 0.05233391952514648, 0.052230335235595705, 0.05211795043945312, 0.05215439987182617, 0.05206220626831055, 0.0522342414855957, 0.05260854339599609, 0.052512351989746096, 0.05256691360473633, 0.05248614501953125, 0.052891647338867184, 0.0532213134765625, 0.05313119888305664, 0.05277094268798828, 0.05260678482055664, 0.05267500686645508, 0.05251456069946289, 0.05242819213867188, 0.052367969512939455, 0.05264540863037109, 0.052679134368896485, 0.05262460708618164, 0.052441886901855465, 0.05243904113769531, 0.05237724685668945, 0.052722015380859376, 0.052983806610107424, 0.05277891159057617, 0.052897857666015624, 0.05301046371459961, 0.05318041610717773, 0.053148735046386716, 0.05282051086425781, 0.05285859298706055, 0.05302880096435547, 0.053056255340576175, 0.0530247688293457, 0.05287936019897461, 0.05284864044189453, 0.053043201446533204, 0.053055038452148436, 0.0530579833984375, 0.05314748764038086, 0.0530351676940918, 0.05332787322998047, 0.053460990905761716, 0.053564510345458984, 0.053577953338623044, 0.053686977386474606, 0.053947841644287106, 0.05349353790283203, 0.053250846862792967, 0.053198848724365234, 0.053349632263183594, 0.0532795524597168, 0.05324755096435547, 0.055804641723632815, 0.0528240966796875, 0.05206016159057617, 0.052321502685546875, 0.05195446395874023, 0.05211699295043945, 0.052216320037841796, 0.05217279815673828, 0.05192652893066406, 0.05198499298095703, 0.052184993743896485, 0.052094974517822266, 0.052307968139648435, 0.05255372619628906, 0.05236940765380859, 0.052600414276123046, 0.05237187194824219, 0.052887168884277344, 0.0530431022644043, 0.053108257293701173, 0.052835262298583985, 0.052580352783203124, 0.052751678466796875, 0.05282476806640625, 0.05287116622924805, 0.052514816284179686, 0.05253683090209961, 0.05251327896118164, 0.05243084716796875, 0.05258006286621094, 0.05259843063354492, 0.052759166717529296, 0.052819393157958985, 0.052795967102050784, 0.05298720169067383, 0.05314012908935547, 0.0531126708984375, 0.053319873809814455, 0.05315283203125, 0.05319776153564453, 0.05315584182739258, 0.05301862335205078, 0.05288755035400391, 0.05288742446899414, 0.05283852767944336, 0.05268201446533203, 0.05263183975219726, 0.0531214714050293, 0.05318041610717773, 0.05320499038696289, 0.05279743957519531, 0.05287526321411133, 0.05338851165771484, 0.0534617919921875, 0.05349990463256836, 0.054063102722167966, 0.053593727111816404, 0.05357401657104492, 0.053612510681152345, 0.05346646499633789, 0.05330960083007812, 0.05354140853881836, 0.05309235382080078, 0.05615280151367188, 0.0526802864074707, 0.051922782897949216, 0.0520648307800293, 0.05206399917602539, 0.05209523010253906, 0.05183283233642578, 0.051888126373291016, 0.052077888488769535, 0.052107105255126955, 0.05201391983032227, 0.05225807952880859, 0.052642017364501956, 0.05287168121337891, 0.052563968658447265, 0.0526737289428711, 0.052679489135742184, 0.05301776123046875, 0.05361484909057617, 0.053140064239501954, 0.052664321899414064, 0.05267852783203125, 0.05264950561523438, 0.05247036743164062, 0.05247580718994141, 0.052590686798095705, 0.05272073745727539, 0.05265631866455078, 0.05255241775512695, 0.05258963012695313, 0.05273491287231445, 0.0524389762878418, 0.05276467132568359, 0.05294496154785156, 0.053116928100585936, 0.053268383026123044, 0.05334640121459961, 0.05309439849853516, 0.05339340972900391, 0.053542465209960935, 0.05313328170776367, 0.05301910400390625, 0.05299523162841797, 0.052936542510986326, 0.05303398513793945, 0.0529202880859375, 0.05273977661132812, 0.05288995361328125, 0.052772575378417966, 0.0528361930847168, 0.053019073486328124, 0.052977664947509766, 0.053319679260253904, 0.05357056045532226, 0.05336975860595703, 0.05342604827880859, 0.05370083236694336, 0.05387059020996094, 0.05354515075683594, 0.05321020889282226, 0.05300707244873047, 0.05319855880737305, 0.053274017333984375, 0.05554691314697266, 0.052928703308105465, 0.05213673782348633, 0.05229363250732422, 0.0518551025390625, 0.052289695739746095, 0.052119647979736325, 0.052197662353515625, 0.05229334259033203, 0.05223747253417969, 0.05240665435791016, 0.05190089416503906, 0.052332542419433595, 0.0524799690246582, 0.052346046447753904, 0.052685665130615233, 0.05266841506958008, 0.05293875122070312, 0.053249534606933595, 0.053254657745361325, 0.052738048553466796, 0.052570014953613284, 0.05261116790771484, 0.0525926399230957, 0.052752384185791014, 0.052451072692871095, 0.05238604736328125, 0.052631553649902345, 0.05257625579833984, 0.052205535888671876, 0.05254873657226562, 0.05235148620605469, 0.05272003173828125, 0.052935745239257814, 0.05292294311523438, 0.05327449417114258, 0.05349017715454102, 0.0531517448425293, 0.05309235382080078, 0.05312067031860351, 0.05279983901977539, 0.05280515289306641, 0.05282249450683594, 0.05285395050048828, 0.052757022857666015, 0.05300809478759766, 0.0528922233581543, 0.0530145263671875, 0.05300428771972656, 0.05286707305908203, 0.05282310485839844, 0.053111713409423826, 0.05306774520874023, 0.05339913558959961, 0.05344073486328125, 0.05360460662841797, 0.05355724716186523, 0.05335785675048828, 0.05335833740234375, 0.05339440155029297, 0.05335209655761719, 0.053063007354736326, 0.053302272796630856, 0.055842720031738284, 0.05286297607421875, 0.052152320861816405, 0.05200032043457031, 0.05201119995117188, 0.05198451232910156, 0.05191692733764648, 0.05190262222290039, 0.051902305603027346, 0.052226047515869144, 0.05212355041503906, 0.052043201446533204, 0.05224310302734375, 0.052714656829833985, 0.052684959411621095, 0.052626174926757814, 0.05258848190307617, 0.052886688232421875, 0.05314031982421875, 0.0530081901550293, 0.05279353713989258, 0.05261068725585937, 0.052547969818115235, 0.052491905212402344, 0.05258601760864258, 0.05242726516723633, 0.05222550582885742, 0.05232326507568359, 0.05238700866699219, 0.05235292816162109, 0.05251567840576172, 0.05264547348022461, 0.05311894226074219, 0.05307846450805664, 0.05310211181640625, 0.05332118225097656, 0.05352124786376953, 0.053276832580566404, 0.053411838531494144, 0.05325823974609375, 0.05309030532836914, 0.05323571014404297, 0.05307187271118164, 0.05300953674316406, 0.05308047866821289, 0.05309711837768555, 0.053020606994628905, 0.05284441757202148, 0.05281587219238281, 0.053065727233886716, 0.05302220916748047, 0.052972030639648435, 0.05313523101806641, 0.053284927368164064, 0.053292160034179685, 0.05327727890014648, 0.0534101448059082, 0.05354860687255859, 0.05346140670776367, 0.05395417785644531, 0.05346550369262695, 0.05345894241333008, 0.053372928619384766, 0.055756126403808594, 0.052789344787597656, 0.05200953674316406, 0.05200476837158203, 0.052856544494628906, 0.05212969589233398, 0.05211180877685547, 0.05196393585205078, 0.05208268737792969, 0.052170753479003906, 0.05224038314819336, 0.052111358642578126, 0.05206835174560547, 0.05242012786865234, 0.05262793731689453, 0.0526376953125, 0.05298995208740234, 0.05326438522338867, 0.05356243133544922, 0.05299296188354492, 0.052585662841796874, 0.05260787200927734, 0.05270719909667969, 0.052453441619873045, 0.05243699264526367, 0.05237145614624023, 0.05257612609863281, 0.05251289749145508, 0.05221907043457031, 0.05208147048950195, 0.05228915023803711, 0.05252124786376953, 0.05253539276123047, 0.05280115127563476, 0.05292070388793945, 0.0531104621887207, 0.05355929565429687, 0.05357526397705078, 0.053290752410888674, 0.05294179153442383, 0.05298332977294922, 0.053156318664550783, 0.053020126342773435, 0.05283484649658203, 0.052973281860351565, 0.05288729476928711, 0.05296796798706055, 0.05292828750610352, 0.05297948837280273, 0.053084606170654296, 0.05320908737182617, 0.053579776763916016, 0.05349158477783203, 0.05383536148071289, 0.05368681716918945, 0.053972991943359375, 0.054005184173583985, 0.05376671981811523, 0.05370787048339844, 0.05342044830322266, 0.053262847900390625, 0.053440513610839846, 0.05333401489257812, 0.05561801528930664, 0.0530470085144043, 0.05219971084594727, 0.05208883285522461, 0.05218624114990234, 0.05224668884277344, 0.052050113677978516, 0.05215900802612305, 0.052170753479003906, 0.05204377746582031, 0.05210726547241211, 0.05237324905395508, 0.052311294555664065, 0.052652416229248045, 0.05262195205688477, 0.05273977661132812, 0.052455551147460935, 0.052905471801757815, 0.053246654510498044, 0.052910079956054686, 0.052778305053710936, 0.05270150375366211, 0.05254182434082031, 0.0526146240234375, 0.0524881591796875, 0.052756671905517576, 0.05260121536254883, 0.052596736907958984, 0.05245132827758789, 0.052381214141845704, 0.05259516906738281, 0.052719615936279295, 0.05281299209594727, 0.05297235107421875, 0.052908031463623044, 0.05351119995117187, 0.05359622573852539, 0.053392288208007815, 0.05329235076904297, 0.05305158233642578, 0.053146110534667966, 0.05299750518798828, 0.053211776733398435, 0.052940032958984376, 0.05299481582641601, 0.05294879913330078, 0.05301631927490234, 0.053238014221191406, 0.052846782684326174, 0.0529112319946289, 0.053185279846191404, 0.05325836944580078, 0.053440513610839846, 0.053694465637207034, 0.05359775924682617, 0.05366620635986328, 0.053501983642578126, 0.05355059051513672, 0.05332223892211914, 0.05350572967529297, 0.053358913421630856, 0.05339516830444336, 0.05350409698486328, 0.055047870635986325, 0.052676929473876956, 0.05187353515625, 0.05205632019042969, 0.0520577278137207, 0.05216470336914063, 0.052181278228759766, 0.052278270721435545, 0.052429824829101565, 0.052242431640625, 0.05228543853759766, 0.05232572937011719, 0.05269750213623047, 0.052752544403076175, 0.052592094421386716, 0.052757118225097654, 0.052711071014404295, 0.052918624877929685, 0.05329305648803711, 0.05309030532836914, 0.05295513534545898, 0.05279948806762695, 0.05280153656005859, 0.05258444976806641, 0.05261270523071289, 0.05252751922607422, 0.05264384078979492, 0.052759872436523435, 0.05255033493041992, 0.052649120330810546, 0.052644001007080075, 0.052834945678710936, 0.052799552917480466, 0.05294633483886719, 0.05324451065063476, 0.05346051025390625, 0.05340127944946289, 0.05345974349975586, 0.05357139205932617, 0.05328844833374023, 0.05292921447753906, 0.05298175811767578, 0.0527968635559082, 0.05281433486938476, 0.052936225891113284, 0.05289833450317383, 0.052975616455078124, 0.0529692497253418, 0.053039169311523436, 0.053045406341552734, 0.053108478546142576, 0.05326399993896484, 0.05324643325805664, 0.05342838287353516, 0.05369241714477539, 0.05357328033447266, 0.05345315170288086, 0.05346294403076172, 0.05343148803710938, 0.053230175018310545, 0.05317631912231445, 0.05321897506713867, 0.053178878784179685]",tokens/s,18.899787274294503,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,8197.865472,11078.139904,0.0,10699.669504,10468.411392,s,1,13.45709765625,13.45709765625,0.0,13.45709765625,13.45709765625,13.45709765625,13.45709765625,[13.45709765625],,kWh,0.0001882311869374765,2.0753804868426186e-05,6.216949417997553e-05,0.00027115448598587824,,MB,3945.213952,11596.136448,0.0,11188.30592,10923.769856,s,10,7.943306579589843,0.7943306579589844,0.0047716555835403416,0.794206756591797,0.7988052307128907,0.800322964477539,0.8015371514892579,"[0.784219970703125, 0.7924779052734375, 0.7932594604492188, 0.7914694213867187, 0.8018406982421875, 0.795154052734375, 0.7906920776367188, 0.7984679565429688, 0.79746826171875, 0.7982567749023437]",tokens/s,322.28392223685097,kWh,2.3125247469863842e-05,2.550328321795675e-06,1.5457726041385905e-05,4.113330183304542e-05,tokens/kWh,6223667.651069438,MB,3945.213952,11598.2336,0.0,11190.403072,10923.772416,s,10,35.77825537109375,3.577825537109375,0.005760191824826535,3.5798658447265623,3.58402451171875,3.5842446777343753,3.584420810546875,"[3.56743115234375, 3.57274609375, 3.56941748046875, 3.576769287109375, 3.5807265625, 3.58446484375, 3.579005126953125, 3.582830322265625, 3.580888916015625, 3.5839755859375]",tokens/s,17.60846059891994,kWh,0.00010474067334097117,1.1553485363672723e-05,6.941428630061068e-05,0.00018570844500525457,tokens/kWh,339241.43836440734,,s,630,35.7751304016113,0.05678592127239893,0.0007543399283584048,0.056694927215576174,0.05728260116577148,0.057606292915344234,0.06109767765045166,"[0.06104678344726563, 0.0568721923828125, 0.05640176010131836, 0.05626144027709961, 0.05611859130859375, 0.05637529754638672, 0.05663616180419922, 0.05645075225830078, 0.05637263870239258, 0.055994464874267576, 0.056225761413574216, 0.056209632873535156, 0.056580894470214846, 0.05636288070678711, 0.05614960098266602, 0.05589443206787109, 0.056635391235351565, 0.05762870407104492, 0.057264095306396486, 0.05636710357666016, 0.05656576156616211, 0.055984127044677735, 0.05602886581420898, 0.056226112365722655, 0.05611635208129883, 0.055866432189941403, 0.05608995056152344, 0.05687753677368164, 0.05641372680664063, 0.056235488891601564, 0.056392318725585935, 0.05638387298583984, 0.056239776611328125, 0.056869216918945316, 0.05690572738647461, 0.05664153671264648, 0.057019489288330075, 0.05674662399291992, 0.05665203094482422, 0.05625654220581055, 0.05695283126831055, 0.05671315383911133, 0.05638931274414063, 0.0564854736328125, 0.05685702514648437, 0.05692851257324219, 0.05634467315673828, 0.056270305633544924, 0.056752670288085935, 0.05655551910400391, 0.05676614379882813, 0.05675980758666992, 0.05677753448486328, 0.05695078277587891, 0.05695484924316406, 0.056772640228271484, 0.05698771286010742, 0.0569117431640625, 0.05710665512084961, 0.056638721466064454, 0.056545631408691406, 0.05644927978515625, 0.05694668960571289, 0.061118465423583984, 0.05682390213012695, 0.056452320098876956, 0.05634838485717773, 0.0563988151550293, 0.05616643142700195, 0.056104896545410156, 0.05622784042358398, 0.05613078308105469, 0.05603615951538086, 0.056377342224121094, 0.05621062469482422, 0.05617958450317383, 0.05663532638549805, 0.056128990173339846, 0.05669507217407226, 0.057019775390625, 0.056853374481201174, 0.056842239379882815, 0.05651865768432617, 0.056268798828125, 0.05628326416015625, 0.05656768035888672, 0.05633187103271484, 0.05597564697265625, 0.05704518508911133, 0.05634905624389648, 0.05662665557861328, 0.05606671905517578, 0.05609187316894531, 0.05721548843383789, 0.05687283325195312, 0.05723587036132813, 0.0570511360168457, 0.056614753723144534, 0.0567353286743164, 0.05706787109375, 0.05688956832885742, 0.05684633636474609, 0.056371200561523435, 0.05677056121826172, 0.05670502471923828, 0.056721153259277346, 0.05662508773803711, 0.05642601776123047, 0.056516895294189455, 0.05644339370727539, 0.05677423858642578, 0.057073982238769534, 0.057460830688476565, 0.057112575531005856, 0.0568642578125, 0.05698934555053711, 0.056973697662353516, 0.05718255996704102, 0.05684348678588867, 0.05676668930053711, 0.05667295837402344, 0.056600574493408204, 0.05713670349121094, 0.05698809432983398, 0.0563507194519043, 0.056704513549804686, 0.06123471832275391, 0.056406463623046875, 0.05642419052124024, 0.05622332763671875, 0.05585580825805664, 0.05534080123901367, 0.05591475296020508, 0.05597814559936523, 0.05584880065917969, 0.05613363265991211, 0.0564901123046875, 0.056629119873046876, 0.05653094482421875, 0.056430591583251956, 0.056668159484863284, 0.05641625595092774, 0.05682902526855469, 0.05763782501220703, 0.057277793884277346, 0.0568408317565918, 0.05624422454833984, 0.05599027252197265, 0.056594432830810545, 0.0564381103515625, 0.056349342346191406, 0.05624422454833984, 0.05653299331665039, 0.05644902420043945, 0.056457183837890626, 0.056401950836181644, 0.05581545639038086, 0.05628339385986328, 0.05695945739746094, 0.057270271301269535, 0.0570880012512207, 0.05728179168701172, 0.056937217712402344, 0.05663536071777344, 0.05686483383178711, 0.056522239685058595, 0.05658998489379883, 0.05645779037475586, 0.056541439056396484, 0.05628623962402344, 0.05660976028442383, 0.056639072418212894, 0.05644236755371094, 0.05639251327514649, 0.056340576171875, 0.0569826545715332, 0.05729475021362305, 0.05692924880981445, 0.05711872100830078, 0.05713100814819336, 0.05705657577514649, 0.056982208251953125, 0.056909534454345705, 0.05672323226928711, 0.056666622161865236, 0.0566640625, 0.05710147094726563, 0.056546142578125, 0.05630271911621094, 0.06100582504272461, 0.05661036682128906, 0.05628768157958984, 0.05673779296875, 0.05624406433105469, 0.05603958511352539, 0.05575833511352539, 0.05679679870605469, 0.05667830276489258, 0.05613471984863281, 0.05644278335571289, 0.05633337783813477, 0.05675523376464844, 0.05673334503173828, 0.056715198516845707, 0.05645344161987305, 0.057624671936035154, 0.057440193176269534, 0.05730710220336914, 0.05688115310668945, 0.05669491195678711, 0.0566803207397461, 0.05634857559204102, 0.05636105728149414, 0.056712799072265625, 0.05644739151000976, 0.056225727081298825, 0.056428062438964845, 0.05630966567993164, 0.05642303848266601, 0.05644902420043945, 0.05600403213500976, 0.05680620956420898, 0.05697417449951172, 0.05722000122070312, 0.0572042236328125, 0.05759027099609375, 0.05688857650756836, 0.05688361740112305, 0.056439136505126955, 0.056610816955566405, 0.056444190979003904, 0.05655376052856445, 0.05657545471191406, 0.056676609039306644, 0.056799968719482424, 0.056530879974365233, 0.05680953598022461, 0.0569420166015625, 0.05674860763549805, 0.05662105560302735, 0.05690163040161133, 0.05698294448852539, 0.0571602897644043, 0.057434112548828124, 0.057304672241210934, 0.057018302917480466, 0.05678873443603515, 0.05637174224853515, 0.05689785766601563, 0.05678681564331055, 0.05676851272583008, 0.056643585205078124, 0.061511806488037106, 0.05690806579589844, 0.05637315368652344, 0.05581177520751953, 0.05602105712890625, 0.05600457763671875, 0.05583638381958008, 0.05635913467407227, 0.05643299102783203, 0.05620240020751953, 0.05673455810546875, 0.05678444671630859, 0.05643513488769531, 0.05648384094238281, 0.05654937744140625, 0.056948734283447267, 0.057040897369384766, 0.05761843109130859, 0.0573004150390625, 0.06046937561035156, 0.055427391052246096, 0.056199295043945316, 0.05657782363891602, 0.05653484725952149, 0.056378944396972656, 0.056253280639648434, 0.05630771255493164, 0.056156158447265625, 0.05681884765625, 0.05671539306640625, 0.05658108901977539, 0.05676134490966797, 0.05705804824829101, 0.057016063690185546, 0.05690521621704102, 0.05795926284790039, 0.05729462432861328, 0.0568034553527832, 0.05668044662475586, 0.05618460845947266, 0.05676012802124023, 0.05664790344238281, 0.05636140823364258, 0.05643648147583008, 0.0566569595336914, 0.057178848266601565, 0.056612350463867187, 0.05652880096435547, 0.057064254760742186, 0.05716793441772461, 0.057013694763183596, 0.057591457366943356, 0.057766334533691406, 0.056961441040039064, 0.05681478500366211, 0.05708803176879883, 0.057058078765869144, 0.05647974395751953, 0.056899585723876954, 0.056586238861083986, 0.05675740814208984, 0.05673625564575195, 0.056762432098388674, 0.061968479156494144, 0.05691775894165039, 0.05628553771972656, 0.05667216110229492, 0.05638143920898438, 0.0562973747253418, 0.05626211166381836, 0.05600934219360352, 0.05566009521484375, 0.05706172943115234, 0.05659247970581055, 0.056512161254882814, 0.05624457550048828, 0.056426048278808594, 0.0566071662902832, 0.056559616088867185, 0.057923583984375, 0.057030654907226565, 0.056578048706054686, 0.056915969848632814, 0.056608352661132816, 0.05645897674560547, 0.05648454284667969, 0.0563974723815918, 0.056538753509521485, 0.05657276916503906, 0.05652044677734375, 0.05669900894165039, 0.056610721588134766, 0.05668665695190429, 0.05687020874023437, 0.05690851211547852, 0.05689142227172852, 0.05677257537841797, 0.057021888732910156, 0.057382720947265625, 0.05727104187011719, 0.05697884750366211, 0.05680550384521484, 0.05643312072753906, 0.05671321487426758, 0.0567275505065918, 0.05687295913696289, 0.0566743049621582, 0.05671456146240234, 0.05666681671142578, 0.056784511566162106, 0.05662553787231445, 0.05688323211669922, 0.05715350341796875, 0.05705900955200195, 0.057166145324707034, 0.05758566284179688, 0.05748735809326172, 0.057451583862304687, 0.056947647094726564, 0.056587905883789064, 0.056691104888916016, 0.05690774536132812, 0.05718412780761719, 0.0569672966003418, 0.05698121643066406, 0.05945577621459961, 0.06121228790283203, 0.05647536087036133, 0.055841663360595706, 0.05600246429443359, 0.05631372833251953, 0.056010753631591796, 0.05613881683349609, 0.05623440170288086, 0.05650076675415039, 0.05663049697875976, 0.05665158462524414, 0.05648249435424805, 0.05704297637939453, 0.05653260803222656, 0.05639228820800781, 0.05634051132202148, 0.0567459831237793, 0.05739110565185547, 0.05751193618774414, 0.05670220947265625, 0.056228126525878906, 0.056389633178710936, 0.05665840148925781, 0.05666332626342773, 0.056339168548583986, 0.05604556655883789, 0.055875007629394534, 0.05612550354003906, 0.056530784606933594, 0.05790995025634765, 0.05705929565429688, 0.056667457580566405, 0.05789971160888672, 0.05653094482421875, 0.05789878463745117, 0.0575860481262207, 0.05715494537353515, 0.05646768188476563, 0.05639817428588867, 0.05661481475830078, 0.056576160430908205, 0.05664751815795899, 0.05665792083740234, 0.056357982635498044, 0.056955360412597654, 0.05654739379882812, 0.0569409294128418, 0.05670041656494141, 0.056617343902587894, 0.05679731369018555, 0.057148799896240235, 0.05721152114868164, 0.057178112030029295, 0.056971263885498044, 0.0571575698852539, 0.05744566345214844, 0.05716662216186524, 0.05695632171630859, 0.05683465576171875, 0.057059326171875, 0.0566514892578125, 0.05656598281860352, 0.05727648162841797, 0.06091088104248047, 0.056832736968994144, 0.05597990417480469, 0.05595904159545898, 0.05570624160766602, 0.05648793411254883, 0.05630771255493164, 0.05653299331665039, 0.05626857757568359, 0.056344799041748043, 0.05629132843017578, 0.05647324752807617, 0.05673731231689453, 0.05681840133666992, 0.056403743743896485, 0.06030684661865234, 0.05637936019897461, 0.05786697769165039, 0.05786563110351563, 0.05717007827758789, 0.05704771041870117, 0.05639382553100586, 0.056266273498535156, 0.05608047866821289, 0.05670310211181641, 0.05669494247436523, 0.05628313446044922, 0.05648121643066406, 0.056586017608642576, 0.05678982543945312, 0.05686687850952148, 0.05688636779785156, 0.05643142318725586, 0.05708595275878906, 0.0570511360168457, 0.05706547164916992, 0.057059326171875, 0.05731068801879883, 0.056686241149902346, 0.056722305297851563, 0.05636640167236328, 0.05653369522094726, 0.05657964706420898, 0.05673740768432617, 0.05680748748779297, 0.05657881546020508, 0.056899585723876954, 0.05698332977294922, 0.0567946891784668, 0.056885089874267575, 0.056742721557617185, 0.057048095703125, 0.05748630523681641, 0.05718425750732422, 0.05711183929443359, 0.05693008041381836, 0.056983871459960936, 0.05702105712890625, 0.056809249877929686, 0.05679740905761719, 0.056641216278076174, 0.05702892684936523, 0.05739724731445312, 0.06168371200561523, 0.05670912170410156, 0.05594879913330078, 0.055761409759521485, 0.056280830383300784, 0.055890174865722654, 0.05582640075683594, 0.056550590515136716, 0.056293663024902345, 0.056046142578125, 0.057369792938232425, 0.057072574615478516, 0.0571226577758789, 0.05672700881958008, 0.057131584167480466, 0.05670665740966797, 0.05688361740112305, 0.05783347320556641, 0.057761791229248044, 0.05732352066040039, 0.0566640625, 0.056202529907226566, 0.05634860610961914, 0.05648873519897461, 0.056469024658203124, 0.0563707504272461, 0.056112033843994144, 0.0563507194519043, 0.056303390502929686, 0.056387680053710934, 0.055979839324951174, 0.05595577621459961, 0.05664767837524414, 0.057320831298828125, 0.05727910232543945, 0.05697945785522461, 0.057112575531005856, 0.05705833435058594, 0.056715904235839845, 0.05672585678100586, 0.05713510513305664, 0.0566927375793457, 0.05650022506713867, 0.05670502471923828, 0.05694460678100586, 0.056833953857421876, 0.05698537445068359, 0.05672175979614258, 0.05672367858886719, 0.056839969635009764, 0.05722304153442383, 0.057079936981201174, 0.05722534561157227, 0.057433982849121094, 0.05732352066040039, 0.057427967071533206, 0.05719244766235351, 0.05709388732910156, 0.056669471740722656, 0.05651068878173828, 0.057086719512939456, 0.05701551818847656, 0.05685532760620117, 0.06136342239379883, 0.05695318222045898, 0.05648854446411133, 0.056069984436035156, 0.05587558364868164, 0.05600249481201172, 0.05608262252807617, 0.055869312286376954, 0.05610934448242187, 0.055835582733154296, 0.05591939163208008, 0.05662656021118164, 0.056871551513671875, 0.05648588943481445, 0.05651827239990234, 0.05707321548461914, 0.05778515243530274, 0.05780889511108398, 0.057587039947509765, 0.057114334106445314, 0.05664863967895508, 0.05661859130859375, 0.056586334228515625, 0.05643500900268555, 0.05655299377441406, 0.056560096740722654, 0.05659852981567383, 0.056406017303466796, 0.056774654388427735, 0.05695484924316406, 0.0563546257019043, 0.05677008056640625, 0.057082561492919924, 0.0573908462524414, 0.058119552612304684, 0.057770881652832035, 0.05704499053955078, 0.05688649749755859, 0.05655590438842773, 0.05680374526977539, 0.0568682861328125, 0.056583934783935544, 0.0566544303894043, 0.056807647705078124, 0.05668364715576172, 0.05710435104370117, 0.05686947250366211, 0.05674220657348633, 0.05661859130859375, 0.05717203140258789, 0.05745904159545898, 0.05728988647460938, 0.05798963165283203, 0.05721942520141601, 0.056764415740966793, 0.05706137466430664, 0.05724063873291016, 0.05680633544921875, 0.056799232482910154, 0.05685164642333984, 0.05673654556274414, 0.056977344512939454, 0.05699388885498047]",tokens/s,17.60999870378179,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,808.18176,537.853952,0.0,159.383552,141.514752,s,1,7.6847021484375,7.6847021484375,0.0,7.6847021484375,7.6847021484375,7.6847021484375,7.6847021484375,[7.6847021484375],,kWh,1.780135854581507e-05,1.955537785813888e-06,5.623060053999707e-06,2.5379956385628663e-05,,MB,1145.286656,607.059968,0.0,199.22944,184.525824,s,24,0.17903702497482296,0.0074598760406176245,7.120415389335122e-05,0.00743944001197815,0.007569103956222534,0.007601238560676574,0.007614982557296753,"[0.00751251220703125, 0.007367648124694824, 0.007426559925079346, 0.00760364818572998, 0.0073504958152771, 0.00743887996673584, 0.007481056213378906, 0.007424928188323974, 0.007618368148803711, 0.007587584018707275, 0.007427648067474365, 0.0075259838104248045, 0.007387135982513428, 0.007512735843658447, 0.007440000057220459, 0.007423744201660156, 0.007369279861450195, 0.007523935794830322, 0.007436384201049804, 0.007444096088409424, 0.007463903903961182, 0.0074160962104797366, 0.00744649600982666, 0.007407904148101807]",tokens/s,34316.924115914,kWh,2.2021106029556478e-07,2.428474853573679e-08,1.2590722860827754e-07,3.704030374395791e-07,tokens/kWh,691139040.7854289,MB,1156.395008,609.15712,0.0,201.326592,184.528384,s,24,10.067228973388671,0.4194678738911947,0.0022759284913262696,0.4189407043457031,0.42268928833007813,0.42317518920898434,0.42340203582763675,"[0.42137295532226565, 0.4157905578613281, 0.41947012329101563, 0.41589859008789065, 0.4232285461425781, 0.4189258422851562, 0.42165786743164063, 0.423453857421875, 0.4181371154785156, 0.41752493286132814, 0.4171626281738281, 0.4222610168457031, 0.4212044677734375, 0.42287283325195313, 0.41895556640625, 0.41828021240234375, 0.4200715026855469, 0.4179279479980469, 0.4219461364746094, 0.41966140747070313, 0.41865359497070315, 0.4182972106933594, 0.41855194091796877, 0.415922119140625]",tokens/s,150.19028612508595,kWh,1.2044924637447583e-05,1.3283434455706353e-06,4.416584831309737e-06,1.778985291432796e-05,tokens/kWh,3541344.6251295176,,s,1512,10.05633614206313,0.006651015966972977,0.00013737650396346672,0.006620895862579346,0.0067276641845703125,0.0068116720438003535,0.007398489761352548,"[0.006614304065704345, 0.0067135682106018064, 0.006707200050354004, 0.006791679859161377, 0.006664127826690674, 0.00666534423828125, 0.006680960178375244, 0.006682752132415771, 0.006635007858276368, 0.006681536197662354, 0.006719200134277343, 0.006696832180023193, 0.006676896095275879, 0.00708403205871582, 0.0066312642097473146, 0.0066845760345458985, 0.006676735877990722, 0.006629471778869629, 0.006638815879821777, 0.006664896011352539, 0.00664086389541626, 0.0066646718978881836, 0.0066624641418457035, 0.0066427521705627445, 0.0067265920639038084, 0.006782911777496338, 0.006838335990905762, 0.006679967880249023, 0.006670368194580078, 0.0067090878486633304, 0.006726655960083008, 0.006737631797790527, 0.006741727828979492, 0.006703328132629394, 0.006743775844573975, 0.006756703853607177, 0.006674655914306641, 0.00667625617980957, 0.0067051520347595215, 0.0066416640281677245, 0.006670335769653321, 0.006656000137329102, 0.006715199947357178, 0.00678876781463623, 0.006634016036987305, 0.006612991809844971, 0.006660096168518067, 0.006628896236419678, 0.006616991996765137, 0.006631840229034424, 0.006596159934997558, 0.006607295989990235, 0.006650015830993652, 0.0066112961769104, 0.006600351810455323, 0.0067108159065246585, 0.0067957439422607424, 0.006623167991638184, 0.0066592001914978025, 0.006662816047668457, 0.006553887844085693, 0.006590271949768066, 0.0065924801826477055, 0.006522719860076904, 0.006564000129699707, 0.006567935943603515, 0.006599967956542969, 0.006543935775756836, 0.006595744132995606, 0.006657023906707763, 0.006621471881866455, 0.00654691219329834, 0.006570335865020752, 0.0065469760894775395, 0.006513023853302002, 0.00654252815246582, 0.006560512065887451, 0.0065270400047302245, 0.006557568073272705, 0.006506432056427002, 0.006537407875061035, 0.00653926420211792, 0.0065413122177124024, 0.006574368000030518, 0.006569695949554443, 0.006624256134033203, 0.006556672096252441, 0.006533055782318115, 0.006533184051513672, 0.006551551818847656, 0.006553791999816895, 0.006554848194122315, 0.006531712055206299, 0.006518080234527588, 0.00659500789642334, 0.006510816097259522, 0.006547520160675049, 0.006627264022827148, 0.006546944141387939, 0.006510784149169922, 0.006689087867736816, 0.006649407863616943, 0.006562240123748779, 0.00659660816192627, 0.006594560146331787, 0.006621183872222901, 0.0065842242240905765, 0.006576320171356201, 0.006584256172180176, 0.006590432167053223, 0.006593728065490722, 0.00658739185333252, 0.006640480041503906, 0.0065771198272705075, 0.006610720157623291, 0.00668665599822998, 0.006608831882476807, 0.006586719989776611, 0.006831488132476807, 0.006709440231323242, 0.006783103942871094, 0.006623456001281738, 0.006610367774963379, 0.006637216091156006, 0.006624256134033203, 0.006814943790435791, 0.0065598077774047855, 0.0066026878356933594, 0.006608672142028808, 0.006762720108032226, 0.006608640193939209, 0.006654399871826172, 0.00662713623046875, 0.007237055778503418, 0.0067794561386108395, 0.006615039825439453, 0.006597792148590088, 0.006591519832611084, 0.006575488090515137, 0.006595071792602539, 0.006575647830963135, 0.006574495792388916, 0.006586368083953857, 0.006586080074310303, 0.006774208068847656, 0.0066179518699646, 0.006536767959594726, 0.0066497278213500976, 0.006600448131561279, 0.0066117758750915525, 0.00659881591796875, 0.0066118078231811525, 0.006621632099151611, 0.006595424175262451, 0.0066343040466308595, 0.006648736000061035, 0.0066212801933288575, 0.006612736225128174, 0.006621344089508057, 0.006754303932189941, 0.006635519981384277, 0.006635072231292724, 0.0066273918151855465, 0.006602911949157715, 0.006643680095672608, 0.006648287773132324, 0.006637216091156006, 0.00662886381149292, 0.006621823787689209, 0.006615039825439453, 0.006651616096496582, 0.006680863857269287, 0.006643008232116699, 0.006615583896636963, 0.006744287967681885, 0.006665184020996094, 0.006791679859161377, 0.006666719913482666, 0.006673408031463623, 0.006632448196411133, 0.006635263919830322, 0.006629087924957275, 0.00667907190322876, 0.0067226881980895995, 0.0066689600944519044, 0.006693088054656983, 0.006743711948394775, 0.006660799980163574, 0.0066776638031005855, 0.006540639877319336, 0.007158751964569092, 0.006627103805541992, 0.006575967788696289, 0.006596735954284668, 0.006600895881652832, 0.006565855979919434, 0.006589759826660156, 0.006623968124389649, 0.006581823825836181, 0.006560416221618653, 0.0065963840484619144, 0.00659660816192627, 0.006573440074920655, 0.006900352001190185, 0.006567520141601562, 0.006594687938690186, 0.006619423866271973, 0.006608895778656006, 0.006582272052764892, 0.006619135856628418, 0.006594592094421386, 0.006586336135864258, 0.006610559940338135, 0.006609280109405518, 0.00658841609954834, 0.006600160121917725, 0.006597152233123779, 0.006604800224304199, 0.006567039966583252, 0.006549824237823487, 0.006557248115539551, 0.00657260799407959, 0.006574175834655762, 0.006622591972351074, 0.006584735870361328, 0.00658844804763794, 0.0066219520568847655, 0.006602528095245361, 0.00656006383895874, 0.0066186881065368654, 0.006568064212799072, 0.006600800037384033, 0.006593823909759522, 0.006583136081695556, 0.0065781760215759275, 0.00656771183013916, 0.006536255836486817, 0.006560704231262207, 0.006528319835662842, 0.0065502080917358395, 0.006582079887390136, 0.006545599937438965, 0.00653926420211792, 0.006551424026489258, 0.006549056053161621, 0.00657260799407959, 0.0065474557876586915, 0.006551551818847656, 0.006559072017669677, 0.006549759864807129, 0.006533696174621582, 0.00662332820892334, 0.006468512058258056, 0.006569983959197998, 0.0066007041931152345, 0.006575263977050781, 0.006603616237640381, 0.00660211181640625, 0.006588575839996338, 0.006650335788726806, 0.0069324798583984375, 0.006631423950195312, 0.006909952163696289, 0.006774879932403564, 0.006651072025299072, 0.006802144050598144, 0.006714399814605713, 0.007000800132751465, 0.006676896095275879, 0.006831967830657959, 0.006677631855010986, 0.006697919845581055, 0.00675219202041626, 0.006668416023254395, 0.006627552032470703, 0.0066902079582214355, 0.006680831909179687, 0.006598495960235596, 0.006764480113983154, 0.006627647876739502, 0.006631328105926514, 0.006710559844970703, 0.00765718412399292, 0.006691359996795655, 0.006903327941894531, 0.00685971212387085, 0.006695968151092529, 0.006660831928253174, 0.006764800071716309, 0.006670400142669678, 0.00664569616317749, 0.006684000015258789, 0.0066258878707885745, 0.006780672073364258, 0.007104191780090332, 0.006660511970520019, 0.006642015933990478, 0.006622464179992676, 0.006645664215087891, 0.00663424015045166, 0.006627295970916748, 0.00674019193649292, 0.006684031963348389, 0.00669532823562622, 0.006703296184539795, 0.006645088195800781, 0.006656479835510254, 0.006897503852844238, 0.006700799942016601, 0.006646175861358642, 0.006684671878814697, 0.006632800102233887, 0.006611711978912353, 0.006588287830352783, 0.006610496044158936, 0.0066119680404663084, 0.006622208118438721, 0.006667744159698486, 0.006588543891906738, 0.006637440204620361, 0.006592895984649658, 0.006590047836303711, 0.006593088150024414, 0.006608799934387207, 0.006645535945892334, 0.006692768096923828, 0.006621600151062012, 0.006686560153961182, 0.006635744094848633, 0.006637792110443116, 0.006686431884765625, 0.006645760059356689, 0.006605088233947754, 0.006917856216430664, 0.006642911911010742, 0.006619232177734375, 0.006689472198486328, 0.006657375812530518, 0.006671072006225586, 0.006649792194366455, 0.006610208034515381, 0.006601439952850342, 0.006620863914489746, 0.006629183769226075, 0.006590976238250733, 0.006633471965789795, 0.006602399826049805, 0.006594687938690186, 0.006627552032470703, 0.006608895778656006, 0.006572288036346435, 0.006647552013397217, 0.006885087966918945, 0.0066276159286499025, 0.006625279903411865, 0.006660096168518067, 0.00658841609954834, 0.006643712043762207, 0.006708352088928222, 0.006582208156585693, 0.006611904144287109, 0.006589920043945313, 0.006605343818664551, 0.006623231887817383, 0.006610559940338135, 0.006605184078216553, 0.006611040115356445, 0.006627232074737549, 0.0067051520347595215, 0.006632832050323487, 0.006650496006011963, 0.006617311954498291, 0.006743840217590332, 0.006711264133453369, 0.006652991771697998, 0.006714560031890869, 0.00669052791595459, 0.006598144054412842, 0.006526976108551025, 0.006937632083892822, 0.006622176170349121, 0.006647071838378906, 0.006641407966613769, 0.006574687957763672, 0.006602784156799316, 0.006593952178955078, 0.006630112171173095, 0.006621407985687256, 0.006595935821533203, 0.006623104095458985, 0.006830880165100098, 0.0066007041931152345, 0.006660096168518067, 0.006619328022003173, 0.006595647811889648, 0.006652895927429199, 0.006641439914703369, 0.00661900806427002, 0.006659327983856201, 0.006628223896026611, 0.006645760059356689, 0.006665535926818848, 0.006634175777435303, 0.006629216194152832, 0.006684832096099854, 0.006692863941192627, 0.006680768013000488, 0.006692319869995117, 0.006651296138763428, 0.006697184085845948, 0.006695648193359375, 0.0066844158172607426, 0.006689023971557617, 0.006719359874725342, 0.006667744159698486, 0.006677152156829834, 0.006964863777160645, 0.006783616065979004, 0.006870783805847168, 0.006680768013000488, 0.006719295978546143, 0.0067235522270202635, 0.006684703826904297, 0.006696671962738037, 0.0066679039001464845, 0.006656832218170166, 0.006708767890930176, 0.006676608085632324, 0.0066070399284362796, 0.006639616012573242, 0.007557439804077149, 0.006667967796325684, 0.006647808074951172, 0.006633823871612549, 0.00659116792678833, 0.006674784183502197, 0.0066648321151733395, 0.006666240215301514, 0.006723743915557861, 0.006690976142883301, 0.006669727802276612, 0.0065797438621521, 0.006646240234375, 0.006647808074951172, 0.0066538558006286625, 0.00661900806427002, 0.0066500802040100095, 0.006686272144317627, 0.006638016223907471, 0.006633471965789795, 0.006717440128326416, 0.0067338237762451176, 0.006601984024047851, 0.006680768013000488, 0.006611519813537597, 0.00658406400680542, 0.007115007877349854, 0.006666240215301514, 0.006653952121734619, 0.0066183037757873535, 0.006623583793640137, 0.006624735832214355, 0.0066447358131408694, 0.006694911956787109, 0.006621183872222901, 0.00671724796295166, 0.006589888095855713, 0.006607615947723388, 0.006623136043548584, 0.0066269440650939945, 0.007408095836639404, 0.008238752365112305, 0.006781375885009765, 0.006733727931976318, 0.0067276802062988285, 0.006672383785247803, 0.006684351921081543, 0.006647264003753662, 0.006636096000671387, 0.006666528224945068, 0.006895296096801758, 0.00666860818862915, 0.0067051520347595215, 0.006676671981811523, 0.006612800121307373, 0.006663616180419922, 0.006658720016479492, 0.006610847949981689, 0.006612736225128174, 0.006594719886779785, 0.006803199768066406, 0.006658400058746338, 0.006677728176116943, 0.006660672187805175, 0.006670239925384521, 0.006709568023681641, 0.00666374397277832, 0.006714816093444824, 0.00671011209487915, 0.00698905611038208, 0.006664576053619385, 0.00668726396560669, 0.0066509442329406735, 0.006707680225372314, 0.006628287792205811, 0.006668320178985596, 0.006647583961486817, 0.006663839817047119, 0.006631968021392823, 0.006618944168090821, 0.00663372802734375, 0.006666175842285156, 0.006598656177520752, 0.006590464115142822, 0.006729504108428955, 0.006735775947570801, 0.006723264217376709, 0.006646399974822998, 0.006623551845550537, 0.006600575923919678, 0.006600607872009277, 0.00667193603515625, 0.006599135875701904, 0.006571904182434082, 0.006626399993896484, 0.0066139202117919925, 0.00656390380859375, 0.006621183872222901, 0.006598239898681641, 0.006578527927398682, 0.0066295361518859865, 0.006592351913452149, 0.006606847763061524, 0.006596447944641114, 0.006574111938476563, 0.0065651841163635255, 0.006622015953063965, 0.0066244797706604, 0.0065586881637573246, 0.006614848136901856, 0.00667625617980957, 0.006609119892120361, 0.006600096225738526, 0.006599264144897461, 0.006592031955718994, 0.006592991828918457, 0.006575456142425537, 0.006611231803894043, 0.006578495979309082, 0.0065598077774047855, 0.006594560146331787, 0.006612991809844971, 0.006575424194335938, 0.006607135772705078, 0.006581920146942139, 0.006877952098846436, 0.0068055038452148435, 0.006742335796356201, 0.006583487987518311, 0.006608863830566406, 0.006695680141448975, 0.006616864204406738, 0.0065946559906005855, 0.0065801281929016115, 0.006668288230895996, 0.006881279945373535, 0.006635519981384277, 0.006500671863555909, 0.00663315200805664, 0.006600224018096924, 0.00660425615310669, 0.006657023906707763, 0.006690815925598144, 0.006604032039642334, 0.00664243221282959, 0.006625279903411865, 0.0065732159614562986, 0.007563168048858642, 0.006654911994934082, 0.006645952224731445, 0.006911808013916015, 0.006748159885406494, 0.006660096168518067, 0.006622943878173828, 0.0065865921974182125, 0.0065905280113220215, 0.00658022403717041, 0.006602431774139404, 0.00657587194442749, 0.0066113600730896, 0.006590879917144775, 0.0065797119140625, 0.0065763840675354, 0.006596320152282715, 0.00658460807800293, 0.006586016178131104, 0.006597184181213379, 0.006563136100769043, 0.006602367877960205, 0.006590591907501221, 0.006568672180175782, 0.006563104152679443, 0.006568672180175782, 0.006612991809844971, 0.006596096038818359, 0.006574592113494873, 0.006563392162322998, 0.006605343818664551, 0.006539455890655518, 0.006555359840393066, 0.006536255836486817, 0.006560832023620605, 0.0066145601272583005, 0.006582623958587647, 0.006557248115539551, 0.006645247936248779, 0.006577055931091308, 0.006600575923919678, 0.006607135772705078, 0.006573567867279053, 0.006590400218963623, 0.006584832191467285, 0.006553120136260986, 0.00659497594833374, 0.006566080093383789, 0.006598688125610351, 0.006596672058105469, 0.006608607769012451, 0.006563136100769043, 0.006869696140289307, 0.006553599834442139, 0.0066109437942504885, 0.006612031936645508, 0.00711520004272461, 0.006574592113494873, 0.006612480163574219, 0.006578688144683838, 0.0065491518974304195, 0.006586719989776611, 0.00658841609954834, 0.006557568073272705, 0.006577600002288819, 0.0065560641288757325, 0.006591104030609131, 0.006571296215057373, 0.006582015991210937, 0.006707839965820313, 0.006612448215484619, 0.006619647979736328, 0.006620927810668946, 0.006592383861541748, 0.006580031871795654, 0.006632031917572022, 0.006636960029602051, 0.00661078405380249, 0.0066130561828613285, 0.006625984191894531, 0.006612256050109864, 0.006635392189025879, 0.006635712146759033, 0.006570655822753907, 0.0066375679969787596, 0.0066007041931152345, 0.006598015785217285, 0.00662553596496582, 0.006656383991241455, 0.006593696117401123, 0.006642271995544433, 0.006591775894165039, 0.006617760181427002, 0.006590591907501221, 0.006568128108978271, 0.006616896152496338, 0.00659884786605835, 0.006584320068359375, 0.0065781760215759275, 0.006574175834655762, 0.006575744152069091, 0.006631040096282959, 0.006613471984863281, 0.0065680317878723145, 0.006608672142028808, 0.006620800018310547, 0.0065790719985961916, 0.006601535797119141, 0.006587135791778565, 0.006634943962097168, 0.006587200164794922, 0.006598624229431152, 0.00663318395614624, 0.006723904132843018, 0.006658016204833984, 0.006695168018341064, 0.006567647933959961, 0.006608799934387207, 0.006635424137115479, 0.006624800205230713, 0.006593472003936768, 0.006665408134460449, 0.006660736083984375, 0.006604991912841797, 0.006643712043762207, 0.006620319843292236, 0.0065911998748779295, 0.006609024047851563, 0.0065987520217895505, 0.0066538558006286625, 0.006662144184112549, 0.006616352081298828, 0.007473887920379639, 0.0077844481468200685, 0.007501823902130127, 0.006674208164215088, 0.006666816234588623, 0.006903039932250977, 0.0068100161552429195, 0.006664415836334228, 0.006655168056488037, 0.006697504043579101, 0.006655519962310791, 0.006628896236419678, 0.006669312000274658, 0.006677504062652588, 0.006606912136077881, 0.00664467191696167, 0.007150847911834717, 0.006986368179321289, 0.006652031898498535, 0.0067153282165527345, 0.006674496173858643, 0.0066184959411621095, 0.006643807888031006, 0.0066260480880737304, 0.0065896639823913574, 0.006596447944641114, 0.006624127864837646, 0.006606080055236816, 0.006586495876312256, 0.006575839996337891, 0.006646527767181397, 0.006618175983428955, 0.006560704231262207, 0.00663750410079956, 0.006600031852722168, 0.006580959796905517, 0.006604800224304199, 0.006574368000030518, 0.006588287830352783, 0.006620192050933838, 0.006623936176300049, 0.006619264125823974, 0.006655168056488037, 0.006656864166259766, 0.006642911911010742, 0.006650623798370362, 0.00662332820892334, 0.006541056156158447, 0.006664703845977784, 0.006573344230651855, 0.0066377601623535155, 0.006588031768798828, 0.006594719886779785, 0.006611711978912353, 0.006612991809844971, 0.006594560146331787, 0.006647808074951172, 0.006700128078460693, 0.006593023777008057, 0.006679999828338623, 0.006628032207489014, 0.0066072320938110355, 0.006618624210357666, 0.006611648082733154, 0.006594592094421386, 0.006620863914489746, 0.006680575847625733, 0.0067645440101623535, 0.006639616012573242, 0.006617311954498291, 0.006631199836730957, 0.006610144138336182, 0.007117824077606201, 0.006667295932769776, 0.006633952140808106, 0.00664192008972168, 0.006625311851501465, 0.006656000137329102, 0.0066416640281677245, 0.006606847763061524, 0.006660096168518067, 0.0066085438728332516, 0.006617440223693848, 0.006675551891326905, 0.0066730880737304685, 0.0066665921211242676, 0.0067564802169799805, 0.00679091215133667, 0.006783008098602295, 0.00682377576828003, 0.006762176036834717, 0.006699264049530029, 0.006740096092224121, 0.006727231979370117, 0.0068059201240539555, 0.006739840030670166, 0.006724095821380615, 0.0067066879272460935, 0.0067053442001342774, 0.006733439922332764, 0.006721951961517334, 0.006684256076812744, 0.006729472160339356, 0.006733856201171875, 0.0067242240905761716, 0.0067358717918396, 0.006668479919433594, 0.006690624237060547, 0.006715392112731934, 0.006686399936676026, 0.006573503971099854, 0.006692512035369873, 0.006663167953491211, 0.0066715202331542964, 0.0067226881980895995, 0.00664089584350586, 0.0066656961441040035, 0.006665215969085693, 0.006647039890289306, 0.006638336181640625, 0.006679967880249023, 0.006648416042327881, 0.006684288024902344, 0.006730112075805664, 0.006735167980194092, 0.006699647903442383, 0.006764832019805908, 0.00671721601486206, 0.006694911956787109, 0.006813695907592773, 0.006788991928100586, 0.006764416217803955, 0.006728288173675537, 0.006737567901611328, 0.006727712154388428, 0.006673696041107177, 0.0067058558464050295, 0.00666428804397583, 0.006635424137115479, 0.006670335769653321, 0.007223199844360351, 0.006723680019378662, 0.007021632194519043, 0.006750495910644532, 0.006701536178588867, 0.00670739221572876, 0.006809599876403808, 0.006658048152923584, 0.0066109437942504885, 0.006633471965789795, 0.006668288230895996, 0.006612703800201416, 0.006745376110076904, 0.006613279819488525, 0.006922976016998291, 0.006795263767242431, 0.006642943859100342, 0.006636223793029785, 0.006662208080291748, 0.006597856044769287, 0.006572256088256836, 0.0066483840942382815, 0.006693088054656983, 0.006544511795043945, 0.0075615358352661135, 0.006627488136291504, 0.006577824115753174, 0.006608479976654052, 0.006609856128692627, 0.006586368083953857, 0.006635615825653076, 0.006622303962707519, 0.006561952114105225, 0.0065548157691955565, 0.0066415038108825685, 0.00671830415725708, 0.006623231887817383, 0.006585472106933594, 0.006568831920623779, 0.006598656177520752, 0.006623424053192138, 0.00702784013748169, 0.006654272079467774, 0.006630080223083496, 0.006603616237640381, 0.006650720119476318, 0.006610367774963379, 0.006574656009674072, 0.006637216091156006, 0.006608895778656006, 0.006817952156066894, 0.00666864013671875, 0.0066022400856018066, 0.00662886381149292, 0.006613311767578125, 0.006593056201934814, 0.006592351913452149, 0.0066152639389038086, 0.006705088138580322, 0.006645760059356689, 0.006633024215698242, 0.006609248161315918, 0.006655424118041992, 0.006627999782562256, 0.006601856231689453, 0.0067916159629821775, 0.006666079998016357, 0.006590112209320069, 0.006609856128692627, 0.0066007041931152345, 0.006561791896820069, 0.0066415038108825685, 0.006612927913665772, 0.006574560165405274, 0.006723328113555908, 0.006669792175292968, 0.006608416080474853, 0.006637792110443116, 0.006576831817626953, 0.006989920139312744, 0.006642975807189941, 0.0066546878814697265, 0.006587776184082031, 0.00658406400680542, 0.006581151962280273, 0.00657747220993042, 0.006617760181427002, 0.00659827184677124, 0.006592959880828858, 0.006614175796508789, 0.006594336032867432, 0.006562816143035889, 0.00658841609954834, 0.006942560195922851, 0.00677459192276001, 0.006623199939727783, 0.0065147199630737306, 0.006588384151458741, 0.006638815879821777, 0.006611711978912353, 0.0065924482345581055, 0.006608287811279297, 0.006623263835906982, 0.006611616134643555, 0.006776832103729248, 0.006998015880584717, 0.0073108158111572265, 0.007446144104003906, 0.00668559980392456, 0.006649600028991699, 0.006608511924743652, 0.006639359951019287, 0.006622335910797119, 0.00659660816192627, 0.006636735916137696, 0.006631872177124023, 0.006592832088470459, 0.00656163215637207, 0.0065576000213623045, 0.006585919857025146, 0.006570271968841553, 0.0065593280792236325, 0.006578815937042236, 0.0065651521682739255, 0.006554304122924804, 0.006577663898468018, 0.006613344192504883, 0.006582240104675293, 0.006584544181823731, 0.006619135856628418, 0.006590144157409668, 0.006625599861145019, 0.006591904163360596, 0.006597216129302978, 0.006575168132781983, 0.006550015926361084, 0.00662992000579834, 0.006584127902984619, 0.006578271865844727, 0.006625120162963867, 0.00659881591796875, 0.006649856090545654, 0.006606847763061524, 0.006595935821533203, 0.006592832088470459, 0.0066154561042785645, 0.006574016094207763, 0.006584320068359375, 0.006608895778656006, 0.0065775041580200195, 0.00659113597869873, 0.006608416080474853, 0.006562272071838379, 0.006653696060180664, 0.006630752086639404, 0.006594880104064942, 0.006611551761627197, 0.006582272052764892, 0.006557439804077149, 0.0064637441635131835, 0.006572896003723144, 0.0065504322052001955, 0.006788896083831787, 0.006713056087493897, 0.006718080043792724, 0.006646687984466553, 0.00661187219619751, 0.006676544189453125, 0.007654911994934082, 0.006598176002502442, 0.006611936092376709, 0.0068074560165405276, 0.0072145919799804685, 0.00663420820236206, 0.0066184959411621095, 0.006685056209564209, 0.006641791820526123, 0.0066102719306945805, 0.006682271957397461, 0.00667955207824707, 0.006637375831604004, 0.006662335872650147, 0.006639520168304444, 0.006624576091766358, 0.006673183917999268, 0.006619135856628418, 0.006627520084381103, 0.006714879989624023, 0.006639359951019287, 0.006652480125427246, 0.006657663822174072, 0.0066072320938110355, 0.006623487949371338, 0.00679091215133667, 0.006635359764099121, 0.006614496231079101, 0.00663212776184082, 0.006631423950195312, 0.0066375679969787596, 0.0066375679969787596, 0.006649856090545654, 0.006606880187988281, 0.006629024028778076, 0.006668223857879639, 0.006611328125, 0.006582176208496094, 0.00656924819946289, 0.00663804817199707, 0.006613344192504883, 0.006577856063842773, 0.006638207912445068, 0.006661824226379395, 0.006594336032867432, 0.006663839817047119, 0.006601632118225098, 0.006588064193725586, 0.006631775856018067, 0.0066202559471130375, 0.006560319900512695, 0.006635680198669434, 0.006606688022613525, 0.006542623996734619, 0.006697055816650391, 0.006638815879821777, 0.006601535797119141, 0.006631328105926514, 0.006856607913970947, 0.006852255821228027, 0.006642208099365234, 0.006586368083953857, 0.006567967891693115, 0.006608863830566406, 0.006607935905456543, 0.006577280044555664, 0.00678278398513794, 0.006620543956756592, 0.006566495895385742, 0.006581600189208984, 0.006634175777435303, 0.006578464031219483, 0.006608607769012451, 0.006586368083953857, 0.006608128070831299, 0.006596896171569824, 0.006551136016845703, 0.006621535778045654, 0.006607135772705078, 0.006572159767150879, 0.006620287895202637, 0.0065710082054138185, 0.0065855679512023925, 0.0066546878814697265, 0.006588479995727539, 0.006573952198028564, 0.006610847949981689, 0.00659219217300415, 0.006664735794067383, 0.006598656177520752, 0.006602719783782959, 0.0066089282035827635, 0.006587456226348877, 0.006562751770019531, 0.0065821118354797365, 0.006565824031829834, 0.006587935924530029, 0.006600959777832031, 0.006587039947509765, 0.006550528049468994, 0.0065862398147583006, 0.00655452823638916, 0.0065532798767089845, 0.006560256004333496, 0.006592319965362549, 0.007219200134277344, 0.006729568004608154, 0.006602079868316651, 0.006546239852905273, 0.0066109437942504885, 0.006643712043762207, 0.006667295932769776, 0.006671328067779541, 0.006715072154998779, 0.006660128116607666, 0.006650144100189209, 0.006633471965789795, 0.0065426878929138186, 0.006706079959869385, 0.006658207893371582, 0.006727519989013672, 0.006676479816436768, 0.006659135818481445, 0.006663455963134766, 0.00667190408706665, 0.006637119770050049, 0.0066334400177001955, 0.0078096318244934085, 0.0067840638160705565, 0.006841343879699707, 0.006657375812530518, 0.00667849588394165, 0.006984320163726807, 0.006617087841033936, 0.006655007839202881, 0.006640448093414307, 0.006621344089508057, 0.006626848220825195, 0.006571936130523682, 0.006605375766754151, 0.006653600215911866, 0.006646111965179443, 0.006616096019744873, 0.006656767845153809, 0.0069749121665954586, 0.006791967868804932, 0.006685855865478515, 0.006687935829162598, 0.006633120059967041, 0.006697279930114746, 0.006661664009094239, 0.006645919799804688, 0.0075797438621521, 0.006694784164428711, 0.0066499199867248535, 0.006594528198242187, 0.006649087905883789, 0.0066926078796386715, 0.006639711856842041, 0.0066893439292907714, 0.006629727840423584, 0.00659660816192627, 0.006581344127655029, 0.00661516809463501, 0.006613215923309326, 0.006596896171569824, 0.00655347204208374, 0.006610847949981689, 0.006699423789978027, 0.006592576026916504, 0.006666272163391113, 0.006619135856628418, 0.006608895778656006, 0.0066416640281677245, 0.006593920230865478, 0.00659827184677124, 0.006611839771270752, 0.006620448112487793, 0.006613632202148437, 0.006622432231903076, 0.006513696193695068, 0.00659884786605835, 0.006631360054016113, 0.0066507840156555175, 0.006612927913665772, 0.006587808132171631, 0.0065625920295715335, 0.006608704090118408, 0.006580448150634766, 0.0065699520111083985, 0.0066249918937683104, 0.006596672058105469, 0.0065782079696655275, 0.006606880187988281, 0.006567903995513916, 0.006576128005981445, 0.006592512130737305, 0.006592319965362549, 0.0065742721557617186, 0.006639616012573242, 0.006651904106140137, 0.006739967823028564, 0.0066744318008422855, 0.00662937593460083, 0.006602752208709717, 0.006627327919006347, 0.006602848052978515, 0.007052351951599121, 0.006633952140808106, 0.00659935998916626, 0.006631103992462159, 0.006623424053192138, 0.0066005120277404785, 0.00658784008026123, 0.006638144016265869, 0.0066085438728332516, 0.006584159851074219, 0.00662284803390503, 0.007574399948120117, 0.006637184143066406, 0.006988160133361817, 0.006635615825653076, 0.006664095878601074, 0.006707136154174805, 0.0067702078819274905, 0.006590432167053223, 0.00659935998916626, 0.006645631790161133, 0.006627327919006347, 0.006656320095062256, 0.006663584232330323, 0.0066583361625671385, 0.006694911956787109, 0.006660096168518067, 0.0066390719413757324, 0.0066418561935424805, 0.006651552200317383, 0.006619840145111084, 0.006666368007659912, 0.006731647968292237, 0.0066228160858154295, 0.006625311851501465, 0.0066596798896789555, 0.006598400115966797, 0.006672671794891357, 0.0066416640281677245, 0.006604512214660644, 0.006640096187591552, 0.006634655952453613, 0.006597280025482178, 0.006668447971343994, 0.006680416107177734, 0.006616256237030029, 0.006631872177124023, 0.00659222412109375, 0.006654592037200928, 0.006687967777252197, 0.0066427202224731445, 0.006907392024993897, 0.006701663970947265, 0.006667967796325684, 0.006617087841033936, 0.0066490240097045895, 0.006691648006439209, 0.006633471965789795, 0.00667251205444336, 0.00666815996170044, 0.006649280071258545, 0.006681248188018799, 0.0066622719764709475, 0.006628352165222168, 0.0066260800361633305, 0.006618144035339356, 0.006632415771484375, 0.006660223960876465, 0.006641536235809326, 0.006643136024475097, 0.006664256095886231, 0.006647424221038818, 0.006634367942810059, 0.006653952121734619, 0.0066304001808166506, 0.006615039825439453, 0.006588799953460693, 0.006611839771270752, 0.0065616960525512694, 0.00661897611618042, 0.006608575820922852, 0.006609216213226319, 0.006597792148590088, 0.006589280128479004, 0.006625279903411865, 0.006612991809844971, 0.0065814399719238284, 0.006634592056274414, 0.006629087924957275, 0.006606751918792724, 0.006621344089508057, 0.006608287811279297, 0.006650400161743164, 0.006658048152923584, 0.006610015869140625, 0.006631968021392823, 0.00662332820892334, 0.006576032161712647, 0.006691199779510498, 0.006516863822937011, 0.006606751918792724, 0.006596447944641114, 0.006588704109191894, 0.006573023796081543, 0.006597536087036133, 0.006620351791381836, 0.006576992034912109, 0.006604032039642334, 0.006597311973571778, 0.006553823947906494, 0.006619232177734375, 0.006583775997161865, 0.006598911762237549, 0.006610144138336182, 0.006564640045166016, 0.006595808029174805, 0.006596960067749023, 0.007074240207672119, 0.006799359798431396, 0.0076447358131408695, 0.006618847846984863, 0.006633823871612549, 0.007188864231109619, 0.006773952007293701, 0.006630080223083496, 0.0065762557983398435, 0.006620192050933838, 0.006611936092376709, 0.006559296131134033, 0.00660422420501709, 0.006572480201721191, 0.0065582718849182126, 0.006581535816192627, 0.006599071979522705, 0.006559264183044433, 0.006640416145324707, 0.006592031955718994, 0.006613247871398926, 0.006612576007843017, 0.006576896190643311, 0.006585343837738037, 0.006612063884735108, 0.0065922880172729495, 0.006692224025726319, 0.006633823871612549, 0.006599167823791504, 0.006643487930297852, 0.006565951824188233, 0.006574016094207763, 0.00659660816192627, 0.0065797119140625, 0.006566400051116943, 0.006602303981781006, 0.00656771183013916, 0.006574816226959228, 0.0065550079345703125, 0.006586944103240967, 0.00656163215637207, 0.006559904098510742, 0.0065413122177124024, 0.006651999950408935, 0.0065673599243164064, 0.006504767894744873, 0.006613376140594482, 0.00658022403717041, 0.006579872131347656, 0.006591104030609131, 0.006586080074310303, 0.006576128005981445, 0.006572192192077637, 0.006565728187561035, 0.006563839912414551, 0.006566239833831787, 0.00656876802444458, 0.0065730881690979, 0.006563839912414551, 0.006559904098510742, 0.006590112209320069, 0.006555520057678223, 0.0065697598457336425, 0.006609248161315918, 0.0065474557876586915, 0.00659660816192627, 0.007102719783782959, 0.007218815803527832, 0.007493760108947754, 0.007320767879486084, 0.00659881591796875, 0.0066013760566711426, 0.006582143783569336, 0.006600831985473633, 0.006616511821746826, 0.006584256172180176, 0.0066312642097473146, 0.006613183975219727, 0.006600895881652832, 0.006582687854766846, 0.0065487041473388675, 0.006565887928009034, 0.006605631828308105, 0.006574048042297363, 0.006584320068359375, 0.0066436161994934085, 0.00661897611618042, 0.006577727794647217, 0.006634175777435303, 0.006559743881225586, 0.006614143848419189, 0.006564191818237305, 0.006550111770629883, 0.006571839809417725, 0.006586368083953857, 0.006570112228393554, 0.006586495876312256, 0.006586143970489502, 0.0065342397689819335, 0.006808640003204346, 0.0065493440628051755, 0.0068854718208312985, 0.00658022403717041, 0.00658841609954834, 0.0065413122177124024, 0.006717696189880371, 0.00659552001953125, 0.00658457612991333, 0.00657203197479248, 0.006559135913848877, 0.006609504222869873, 0.0067049598693847655, 0.006594751834869385, 0.006606527805328369, 0.006570303916931152, 0.006557695865631104, 0.006594560146331787, 0.0065615038871765135, 0.0065642881393432615, 0.006588255882263184, 0.006615039825439453, 0.006574079990386963, 0.006612512111663818, 0.006594528198242187, 0.006583040237426758, 0.0066128640174865725, 0.006607744216918945, 0.006583104133605957, 0.006568128108978271, 0.006576128005981445, 0.00661078405380249, 0.006581984043121338, 0.006559936046600342, 0.006584447860717773, 0.006625408172607422, 0.00657369613647461, 0.0066211199760437016, 0.006576831817626953, 0.006603775978088379, 0.006607583999633789, 0.006588511943817139, 0.006586175918579101, 0.006637695789337158, 0.006565887928009034, 0.006555712223052979, 0.006603007793426514, 0.006573152065277099, 0.0065829439163208004, 0.00659449577331543, 0.0065660161972045895, 0.00659443187713623, 0.006563744068145752, 0.006584415912628174, 0.006631423950195312, 0.006601024150848389, 0.006586048126220703, 0.006621183872222901, 0.006582208156585693, 0.006587808132171631, 0.006611616134643555, 0.006608895778656006, 0.006606272220611572, 0.006596543788909912, 0.006650176048278809, 0.006594880104064942, 0.006602399826049805, 0.006606207847595215, 0.006622176170349121, 0.006600831985473633, 0.006590176105499268, 0.006624576091766358]",tokens/s,150.3529693757632,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: DeciLMForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3128.594432,4369.350656,0.0,3990.880256,3908.719616,s,1,10.4420439453125,10.4420439453125,0.0,10.4420439453125,10.4420439453125,10.4420439453125,10.4420439453125,[10.4420439453125],,kWh,9.952271615411518e-05,1.0970786173134378e-05,3.213197014997893e-05,0.00014262547247722847,,MB,3179.76576,4725.866496,0.0,4318.035968,4274.555904,s,10,2.15499136352539,0.215499136352539,0.0018707195778059705,0.2153262405395508,0.21774974975585937,0.2177758834838867,0.2177967904663086,"[0.21505363464355468, 0.21228501892089843, 0.2140268096923828, 0.21306515502929688, 0.21755984497070313, 0.21780201721191406, 0.21680245971679687, 0.21774394226074217, 0.21520895385742186, 0.2154435272216797]",tokens/s,1187.9397956435653,kWh,6.569238207130856e-06,7.244694912347458e-07,4.345781254399489e-06,1.1639488952765092e-05,tokens/kWh,21994092.78524933,MB,3181.510656,4727.963648,0.0,4320.13312,4274.558464,s,10,31.113200927734376,3.1113200927734375,0.011244953616570474,3.113736328125,3.122223779296875,3.1230088623046877,3.1236369287109373,"[3.1237939453125, 3.09076123046875, 3.120275634765625, 3.11411083984375, 3.11336181640625, 3.12204931640625, 3.111911376953125, 3.089753173828125, 3.112337158203125, 3.114846435546875]",tokens/s,20.248639844652455,kWh,9.059633147994427e-05,9.992893618104263e-06,4.7238454457402324e-05,0.00014782767955545087,tokens/kWh,426171.879241116,,s,630,31.109904785156225,0.04938080124627977,0.000835842498500185,0.0493242073059082,0.049889156723022465,0.05036906566619873,0.05271712337493898,"[0.050149375915527344, 0.04960255813598633, 0.049223678588867184, 0.049248096466064456, 0.04950236892700195, 0.04911513519287109, 0.049127422332763675, 0.0488158073425293, 0.04885440063476563, 0.048605567932128904, 0.04854950332641601, 0.048665534973144534, 0.04873011016845703, 0.04868505477905274, 0.048955230712890624, 0.04934672164916992, 0.0490332145690918, 0.049210784912109375, 0.049258655548095706, 0.0493633918762207, 0.04905136108398438, 0.04917686462402344, 0.04921343994140625, 0.04940367889404297, 0.049215038299560546, 0.04973814392089844, 0.04935091018676758, 0.049517696380615234, 0.049541919708251954, 0.04939952087402344, 0.049651615142822264, 0.05923273468017578, 0.04958560180664062, 0.04955100631713867, 0.04976732635498047, 0.049180255889892575, 0.049305313110351565, 0.049527103424072266, 0.051222305297851566, 0.04941817474365234, 0.049658622741699215, 0.049356704711914064, 0.04970819091796875, 0.04909347152709961, 0.048977344512939454, 0.04958812713623047, 0.049162910461425784, 0.04922745513916016, 0.04953878402709961, 0.04921750259399414, 0.049543807983398434, 0.0494919662475586, 0.049498111724853515, 0.04976796722412109, 0.04963375854492187, 0.0495022087097168, 0.04985651016235351, 0.049616897583007816, 0.04959151840209961, 0.04933817672729492, 0.049281089782714844, 0.05339161682128906, 0.04937599945068359, 0.05022768020629883, 0.05065283203125, 0.04936495971679687, 0.04943500900268555, 0.04931955337524414, 0.05036684799194336, 0.04907827377319336, 0.0490618896484375, 0.04904550552368164, 0.049036960601806644, 0.04872943878173828, 0.049646080017089846, 0.04898867034912109, 0.04922329711914063, 0.04921148681640625, 0.04924358367919922, 0.048995166778564456, 0.04921548843383789, 0.04952678298950195, 0.04953260803222656, 0.04918048095703125, 0.049299392700195316, 0.04897439956665039, 0.04863385772705078, 0.048349246978759766, 0.04802348709106445, 0.05030646514892578, 0.048337505340576174, 0.04769792175292969, 0.047695358276367186, 0.04773529434204102, 0.04779008102416992, 0.047806015014648436, 0.04799327850341797, 0.04776291275024414, 0.047788097381591794, 0.04769760131835937, 0.04806057739257812, 0.048637664794921875, 0.04900764846801758, 0.0489114875793457, 0.04876569747924805, 0.04917814254760742, 0.049754592895507814, 0.048790687561035155, 0.048982433319091793, 0.048955169677734375, 0.04910147094726563, 0.04911513519287109, 0.04884889602661133, 0.04893062210083008, 0.04897811126708984, 0.04975820922851563, 0.04922531127929688, 0.04927324676513672, 0.04926822280883789, 0.04948633575439453, 0.05095337677001953, 0.04936908721923828, 0.04988809585571289, 0.050270206451416014, 0.05026611328125, 0.04969065475463867, 0.05055897521972656, 0.04971142578125, 0.04948550415039062, 0.049285118103027346, 0.048964897155761716, 0.04893360137939453, 0.04931961441040039, 0.049000991821289065, 0.04892649459838867, 0.0489411849975586, 0.04884048080444336, 0.0492237777709961, 0.05177897644042969, 0.05037836837768555, 0.049218399047851566, 0.04993584060668945, 0.04935504150390625, 0.04946736145019531, 0.04968694305419922, 0.04941209411621094, 0.04929740905761719, 0.05240217590332031, 0.049468639373779294, 0.049781536102294924, 0.049604606628417966, 0.050033889770507815, 0.049854782104492186, 0.049400287628173827, 0.04923350524902344, 0.05019894409179688, 0.04957990264892578, 0.04941222381591797, 0.04980099105834961, 0.04969903945922852, 0.049514495849609375, 0.049391616821289064, 0.049532928466796876, 0.049648990631103514, 0.04930342483520508, 0.0493985595703125, 0.049202430725097654, 0.04937564849853516, 0.04940630340576172, 0.04948582458496094, 0.04926639938354492, 0.04946255874633789, 0.049007614135742186, 0.049229248046875, 0.04903129577636719, 0.04914432144165039, 0.04912691116333008, 0.04929548645019531, 0.049279296875, 0.04929945755004883, 0.04935475158691406, 0.04937331390380859, 0.0496044807434082, 0.04962508773803711, 0.04948604965209961, 0.049409568786621096, 0.04950451278686523, 0.04951859283447266, 0.049454559326171876, 0.0501096305847168, 0.04978176116943359, 0.04959033584594726, 0.0495230712890625, 0.04948144149780274, 0.04957443237304687, 0.049635326385498044, 0.04928291320800781, 0.049447296142578125, 0.049078048706054686, 0.049040992736816405, 0.04887091064453125, 0.04872608184814453, 0.04888246536254883, 0.04887065505981445, 0.04868588638305664, 0.04913151931762695, 0.049627391815185544, 0.04941107177734375, 0.04949679946899414, 0.04949151992797852, 0.04958214569091797, 0.04959686279296875, 0.049813472747802734, 0.04971110534667969, 0.04948787307739258, 0.05071392059326172, 0.0496126708984375, 0.04928185653686523, 0.04931584167480469, 0.0491635856628418, 0.04920595169067383, 0.05293875122070312, 0.04927283096313476, 0.04934617614746094, 0.049819072723388674, 0.049590496063232424, 0.04945993423461914, 0.04967628860473633, 0.05015347290039063, 0.05013708877563477, 0.04957548904418945, 0.04981190490722656, 0.049833984375, 0.049641311645507814, 0.049589599609375, 0.049722175598144534, 0.04976579284667969, 0.04926115036010742, 0.049848384857177734, 0.04929119873046875, 0.04913529586791992, 0.0486803207397461, 0.0488988151550293, 0.04873350524902344, 0.04910147094726563, 0.04879894256591797, 0.04860208129882813, 0.048641216278076174, 0.048738815307617187, 0.048580894470214846, 0.04883164978027344, 0.04911222457885742, 0.05549923324584961, 0.04984969711303711, 0.04945417785644531, 0.04946432113647461, 0.04919385528564453, 0.04902912139892578, 0.04897177505493164, 0.048602977752685544, 0.04851087951660156, 0.048020927429199216, 0.04816979217529297, 0.04881817626953125, 0.04976566314697266, 0.04858544158935547, 0.04849059295654297, 0.048694271087646485, 0.049328094482421876, 0.04877376174926758, 0.04895571136474609, 0.04914790344238281, 0.04882406234741211, 0.04889622497558594, 0.04931545639038086, 0.04988150405883789, 0.04891856002807617, 0.04871571350097656, 0.050221088409423825, 0.05005923080444336, 0.049063167572021484, 0.04925724792480469, 0.04925439834594727, 0.04931379318237305, 0.04923712158203125, 0.049132415771484375, 0.04917862319946289, 0.0493383674621582, 0.049205310821533205, 0.049237728118896484, 0.049473567962646486, 0.049311935424804686, 0.04948323059082031, 0.05007535934448242, 0.04911299133300781, 0.04906665420532227, 0.04938294219970703, 0.049068416595458984, 0.049285472869873045, 0.04951244735717773, 0.04940595245361328, 0.0493889274597168, 0.04963391876220703, 0.04950425720214844, 0.05011369705200195, 0.049813472747802734, 0.049547264099121094, 0.04952963256835938, 0.04965795135498047, 0.051337215423583986, 0.049657310485839844, 0.04940035247802734, 0.04998758316040039, 0.04980710220336914, 0.05008127975463867, 0.05046102523803711, 0.04986006546020508, 0.04949593734741211, 0.04975289535522461, 0.04933849716186523, 0.048993377685546874, 0.04902979278564453, 0.04944121551513672, 0.04934371185302734, 0.0492367057800293, 0.04909203338623047, 0.04938172912597656, 0.05195705413818359, 0.05027824020385742, 0.04982470321655273, 0.049847518920898434, 0.049476318359375, 0.04949612808227539, 0.049704959869384766, 0.04916239929199219, 0.04971491241455078, 0.05121036911010742, 0.04974943923950195, 0.04988166427612305, 0.04925030517578125, 0.050479103088378906, 0.049409278869628905, 0.04923843383789062, 0.052410720825195316, 0.04976617431640625, 0.049283294677734374, 0.049530879974365234, 0.04989091110229492, 0.049821216583251955, 0.049914752960205075, 0.04929945755004883, 0.04884380722045899, 0.04886601638793946, 0.048675201416015626, 0.04886105728149414, 0.04917814254760742, 0.0495211181640625, 0.04943667221069336, 0.049211040496826175, 0.0492973747253418, 0.04940224075317383, 0.04926790237426758, 0.04959519958496094, 0.049374496459960934, 0.050574302673339844, 0.049357982635498045, 0.04965232086181641, 0.04955750274658203, 0.05030044937133789, 0.04917510223388672, 0.04931916809082031, 0.049205921173095704, 0.049085502624511716, 0.04854880142211914, 0.04873420715332031, 0.0486195182800293, 0.04886495971679688, 0.04915555191040039, 0.053026302337646485, 0.04979321670532227, 0.04954608154296875, 0.04900191879272461, 0.04900409698486328, 0.048597023010253905, 0.048649185180664065, 0.048475807189941406, 0.048425086975097655, 0.04867583847045898, 0.048919551849365236, 0.048508926391601564, 0.04850400161743164, 0.04892300796508789, 0.0485994873046875, 0.04864409637451172, 0.04883635330200195, 0.04919132614135742, 0.04960345458984375, 0.049320865631103515, 0.04920038223266601, 0.04923270416259766, 0.04952585601806641, 0.04944579315185547, 0.049498111724853515, 0.04945724868774414, 0.04982979202270508, 0.04972771072387695, 0.04965923309326172, 0.04942278289794922, 0.04943769454956055, 0.049238113403320315, 0.04915702438354492, 0.05146787261962891, 0.04941865539550781, 0.04913971328735352, 0.04959436798095703, 0.04928224182128906, 0.049201984405517575, 0.04984201431274414, 0.04949225616455078, 0.04941945648193359, 0.049299968719482425, 0.049479873657226565, 0.04988896179199219, 0.05025823974609375, 0.049631233215332034, 0.049579105377197265, 0.04975094223022461, 0.049649185180664065, 0.049352638244628905, 0.049222175598144534, 0.04977983856201172, 0.049716094970703124, 0.049432575225830076, 0.049487648010253904, 0.049532672882080075, 0.049495838165283204, 0.04955775833129883, 0.049140159606933596, 0.04909619140625, 0.04916502380371094, 0.04909161758422852, 0.05037088012695313, 0.04959187316894531, 0.04932438278198242, 0.04922550582885742, 0.048699169158935546, 0.04857263946533203, 0.04835036849975586, 0.0480401611328125, 0.048071296691894534, 0.04807398223876953, 0.0478864631652832, 0.0477968635559082, 0.04871526336669922, 0.047880702972412106, 0.04779008102416992, 0.047702014923095705, 0.04786182403564453, 0.04869286346435547, 0.0482852783203125, 0.048425697326660154, 0.0485560302734375, 0.04868710327148437, 0.048639553070068356, 0.04844588851928711, 0.04914803314208984, 0.04901583862304688, 0.04925753784179687, 0.05124995040893555, 0.04903014373779297, 0.04905779266357422, 0.04910899353027344, 0.05103529739379883, 0.04935356903076172, 0.04958617782592773, 0.049037311553955076, 0.04937318420410156, 0.0490618896484375, 0.049631233215332034, 0.04930303955078125, 0.04959283065795898, 0.04948563385009765, 0.04963248062133789, 0.0487147216796875, 0.04865987014770508, 0.04873583984375, 0.048953857421875, 0.04889263916015625, 0.049090049743652345, 0.04889014434814453, 0.04899225616455078, 0.04871894454956055, 0.04906390380859375, 0.04911391830444336, 0.049727615356445314, 0.049127422332763675, 0.04939785766601563, 0.04964291381835938, 0.05087862396240234, 0.04977699279785156, 0.04964966583251953, 0.04962918472290039, 0.04942233657836914, 0.04969472122192383, 0.05077436828613281, 0.049870849609375, 0.049876991271972655, 0.050147232055664064, 0.04977849578857422, 0.04960192108154297, 0.04962406539916992, 0.04972329711914063, 0.049510017395019534, 0.04932140731811523, 0.049636096954345704, 0.04966947174072266, 0.050131103515625, 0.05200966262817383, 0.049486846923828126, 0.04921446228027344, 0.0528422737121582, 0.04955158233642578, 0.04936816024780273, 0.048890785217285154, 0.04919270324707031, 0.051128574371337894, 0.0493007698059082, 0.04904364776611328, 0.04918966293334961, 0.04946051025390625, 0.04997964859008789, 0.049418560028076174, 0.049232864379882814, 0.04956649780273437, 0.04939484786987305, 0.04959135818481445, 0.049694656372070316, 0.049544448852539065, 0.04909267044067383, 0.048872127532958984, 0.04862144088745117, 0.04847171020507812, 0.048196063995361325, 0.04825619125366211, 0.04829062271118164, 0.04843244934082031, 0.048290496826171876, 0.048649982452392576, 0.04923980712890625, 0.049360958099365235, 0.04916592025756836, 0.049218399047851566, 0.049285118103027346, 0.04967555236816406, 0.049375232696533204, 0.04958899307250977, 0.04930044937133789, 0.049124351501464845, 0.0498524169921875, 0.04929299163818359, 0.04894534301757812, 0.04886950302124023, 0.04893491363525391, 0.048639999389648435, 0.04849158477783203, 0.048758975982666014, 0.048978561401367186, 0.05046700668334961, 0.049582080841064455, 0.050132991790771485, 0.04927897644042969, 0.04920304107666015, 0.049299488067626955, 0.04902076721191406, 0.048994464874267576, 0.04874431991577149, 0.04864131164550781, 0.048710208892822265, 0.04857427215576172, 0.04873622512817383, 0.04929753494262695, 0.04912761688232422, 0.0489323844909668, 0.04909545516967773, 0.049091934204101566, 0.04916252899169922, 0.04927321624755859, 0.049192512512207034, 0.049358463287353514, 0.04933868789672852, 0.04921395111083984, 0.04903436660766602, 0.049895519256591796, 0.04951724624633789, 0.04949020767211914, 0.04952431869506836, 0.049418464660644534, 0.04953724670410156, 0.04866230392456055, 0.048234657287597654, 0.04877807998657226, 0.0520711669921875, 0.04949967956542969, 0.0491445426940918, 0.049324031829833984, 0.04936703872680664, 0.049153217315673826, 0.04949484634399414, 0.049170433044433595, 0.04914729690551758, 0.04967484664916992, 0.049213409423828125, 0.04951033782958984, 0.04935507202148438, 0.04930310440063477, 0.05007177734375, 0.04949139022827148, 0.04973625564575195, 0.049500160217285157, 0.049622337341308595, 0.04934064102172851, 0.04949244689941406, 0.04994867324829102, 0.049410049438476565, 0.04942598342895508, 0.049460670471191404, 0.04969526290893555, 0.0501049919128418, 0.05339321517944336, 0.04979859161376953]",tokens/s,20.250785219394096,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1007.427584,867.106816,0.0,488.636416,482.553856,s,1,8.09123828125,8.09123828125,0.0,8.09123828125,8.09123828125,8.09123828125,8.09123828125,[8.09123828125],,kWh,2.8708490387528704e-05,3.1514326365997525e-06,9.503063158000646e-06,4.1362986182129104e-05,,MB,1227.771904,1024.393216,0.0,616.562688,582.974464,s,10,0.27607552146911624,0.027607552146911622,0.0003845068961008729,0.027467807769775393,0.027722643280029297,0.028233929824829102,0.028642959060668948,"[0.028745216369628908, 0.02743939208984375, 0.027506624221801758, 0.027577375411987303, 0.02748700714111328, 0.027428768157958985, 0.02740598487854004, 0.027609024047851562, 0.027427520751953125, 0.0274486083984375]",tokens/s,9272.82500953776,kWh,9.256555171676362e-07,1.0208324456250316e-07,6.121418399240817e-07,1.639880601654221e-06,tokens/kWh,156108926.30948946,MB,1238.60992,1036.976128,0.0,629.1456,597.192192,s,10,14.34401013183594,1.4344010131835936,0.006175984855072775,1.4349803466796875,1.4417718017578125,1.4442687744140625,1.4462663525390627,"[1.4467657470703126, 1.4368922119140626, 1.4412169189453126, 1.42808984375, 1.4272528076171875, 1.4339974365234376, 1.4267244873046876, 1.4302384033203126, 1.4359632568359375, 1.4368690185546875]",tokens/s,43.92077209996813,kWh,4.1994459412828755e-05,4.631168553798432e-06,1.6676621991075785e-05,6.330224995770296e-05,tokens/kWh,995225.288865642,,s,630,14.339031734466571,0.022760367832486594,0.00038567347761655764,0.02265856075286865,0.02312019500732422,0.02335652503967285,0.024450917987823496,"[0.022715776443481446, 0.023112159729003906, 0.023201248168945313, 0.023357791900634764, 0.02347452735900879, 0.02612224006652832, 0.023566335678100587, 0.024295103073120116, 0.023484607696533204, 0.023119775772094727, 0.02305638313293457, 0.02300441551208496, 0.022880672454833984, 0.022670976638793944, 0.022641599655151366, 0.02269593620300293, 0.02285158348083496, 0.02285955238342285, 0.02268592071533203, 0.022556640625, 0.022546464920043946, 0.02249318313598633, 0.02254572868347168, 0.022439680099487304, 0.022442943572998048, 0.022618112564086915, 0.022562816619873048, 0.022451200485229493, 0.022602752685546876, 0.023031455993652344, 0.0229881591796875, 0.02298569679260254, 0.023617536544799804, 0.0226693115234375, 0.022666751861572267, 0.02257766342163086, 0.02289459228515625, 0.02451456069946289, 0.025574464797973633, 0.022708927154541016, 0.02253171157836914, 0.02252992057800293, 0.02262835121154785, 0.02354047966003418, 0.02261347198486328, 0.02253468894958496, 0.022486272811889647, 0.02249798393249512, 0.022713888168334962, 0.02264694404602051, 0.022564767837524414, 0.022679807662963868, 0.022587615966796874, 0.022747135162353514, 0.02319673538208008, 0.02265715217590332, 0.022672191619873047, 0.022564863204956053, 0.02296623992919922, 0.022745119094848634, 0.022891872406005858, 0.023302656173706054, 0.022687871932983397, 0.022186784744262694, 0.022656991958618165, 0.022682655334472657, 0.02268057632446289, 0.022566688537597655, 0.022718687057495118, 0.022597503662109376, 0.022585023880004884, 0.02248748779296875, 0.02292889595031738, 0.022671871185302735, 0.023203840255737306, 0.02313862419128418, 0.022859264373779296, 0.02263868713378906, 0.02262819290161133, 0.02263065528869629, 0.022681600570678712, 0.02273689651489258, 0.022609920501708985, 0.022647872924804687, 0.022739423751831054, 0.022563295364379884, 0.022587392807006838, 0.022595359802246095, 0.02279347229003906, 0.02253718376159668, 0.022945024490356444, 0.0227458553314209, 0.022798336029052735, 0.022796287536621093, 0.022747135162353514, 0.022971904754638672, 0.023081151962280274, 0.023021888732910157, 0.023173120498657225, 0.022910560607910156, 0.022757375717163086, 0.022775583267211914, 0.02274982452392578, 0.022691743850708008, 0.02265407943725586, 0.022703071594238282, 0.022693504333496095, 0.022673791885375976, 0.02279337692260742, 0.0226680965423584, 0.022643999099731447, 0.022769632339477538, 0.022760032653808594, 0.02264816093444824, 0.022608736038208007, 0.02267091178894043, 0.022730880737304688, 0.022752864837646485, 0.023999200820922852, 0.02401679992675781, 0.02295408058166504, 0.02285740852355957, 0.0230830078125, 0.022657344818115235, 0.022800384521484376, 0.023675104141235352, 0.02270083236694336, 0.023161983489990233, 0.023257984161376952, 0.022996383666992186, 0.0228002872467041, 0.02261062431335449, 0.022812671661376953, 0.022730752944946288, 0.022552127838134765, 0.022411712646484373, 0.024016416549682618, 0.0236712646484375, 0.0228351993560791, 0.022676959991455078, 0.02260799980163574, 0.022657440185546874, 0.02265907287597656, 0.022443071365356445, 0.022537151336669923, 0.022683488845825196, 0.02338204765319824, 0.022952064514160157, 0.022958080291748048, 0.022990848541259764, 0.0229171199798584, 0.022726655960083008, 0.022671104431152344, 0.02299929618835449, 0.023198944091796875, 0.02324518394470215, 0.023062944412231445, 0.02301139259338379, 0.0230032958984375, 0.02328553581237793, 0.0232857608795166, 0.02313644790649414, 0.02306390380859375, 0.02316454315185547, 0.023103872299194337, 0.02290060806274414, 0.022923936843872072, 0.0228123836517334, 0.02290505599975586, 0.02305638313293457, 0.022863872528076173, 0.022927616119384767, 0.02277507209777832, 0.02261244773864746, 0.022660192489624024, 0.022585407257080078, 0.022672224044799804, 0.022487039566040038, 0.02250873565673828, 0.022542783737182617, 0.022679935455322264, 0.022633888244628905, 0.02251807975769043, 0.022659263610839843, 0.02290287971496582, 0.023052288055419923, 0.022841344833374022, 0.022690912246704102, 0.02255264091491699, 0.022247135162353517, 0.02254841613769531, 0.023210943222045897, 0.02265497589111328, 0.02250752067565918, 0.02244918441772461, 0.0226760311126709, 0.022454687118530273, 0.022693248748779298, 0.02238528060913086, 0.022478847503662108, 0.022378496170043945, 0.02313804817199707, 0.022506879806518554, 0.02250636863708496, 0.02230633544921875, 0.022454368591308595, 0.02241779136657715, 0.02249728012084961, 0.022452192306518556, 0.022509599685668947, 0.022519807815551757, 0.022540191650390624, 0.022525152206420897, 0.022610815048217773, 0.02263859176635742, 0.022541568756103515, 0.02256358337402344, 0.022673408508300782, 0.022460416793823244, 0.02253004837036133, 0.022634496688842775, 0.02267136001586914, 0.022558719635009765, 0.023029727935791014, 0.023095327377319334, 0.02256671905517578, 0.02260323143005371, 0.022698720932006835, 0.022611488342285158, 0.022585376739501953, 0.022538463592529298, 0.022820159912109374, 0.022811168670654296, 0.023069055557250976, 0.022800384521484376, 0.022576255798339842, 0.02272092819213867, 0.02308348846435547, 0.023076543807983397, 0.023341375350952147, 0.022842399597167967, 0.023016000747680666, 0.023354976654052735, 0.02305449676513672, 0.022786720275878906, 0.022550304412841796, 0.022523391723632814, 0.022525951385498046, 0.022500064849853514, 0.022411264419555665, 0.022458368301391602, 0.022619136810302733, 0.022205184936523438, 0.02251568031311035, 0.022622335433959962, 0.022674560546875, 0.02252275276184082, 0.02252992057800293, 0.022497247695922852, 0.02255196762084961, 0.02251228713989258, 0.02253398323059082, 0.022561023712158203, 0.022384639739990234, 0.022525888442993164, 0.02247999954223633, 0.022490047454833986, 0.022579200744628908, 0.022511615753173828, 0.022518880844116212, 0.022495424270629883, 0.022481632232666016, 0.022627967834472656, 0.02257481575012207, 0.02259424018859863, 0.02277987289428711, 0.0227061767578125, 0.02247590446472168, 0.022614912033081056, 0.02280243110656738, 0.022556671142578123, 0.022576223373413085, 0.022547359466552733, 0.022640640258789063, 0.022833152770996092, 0.02265088081359863, 0.02264816093444824, 0.022575136184692382, 0.02283075141906738, 0.02285977554321289, 0.02270102310180664, 0.022634496688842775, 0.02279631996154785, 0.022681568145751955, 0.022548479080200197, 0.022867040634155275, 0.022604703903198242, 0.02274508857727051, 0.02273689651489258, 0.022681280136108397, 0.02275564765930176, 0.022720256805419923, 0.022733055114746093, 0.02271171188354492, 0.022772544860839843, 0.022764991760253907, 0.02291337585449219, 0.022601728439331056, 0.022720352172851562, 0.022722719192504882, 0.022951360702514648, 0.02277452850341797, 0.022869632720947265, 0.02269977569580078, 0.022966720581054687, 0.022420320510864258, 0.022601247787475586, 0.022651359558105467, 0.023166559219360353, 0.02293391990661621, 0.022816768646240236, 0.02271027183532715, 0.022534143447875975, 0.02265088081359863, 0.022511615753173828, 0.02267955207824707, 0.023102527618408204, 0.022862144470214844, 0.022765504837036134, 0.022673120498657228, 0.022658016204833983, 0.023590911865234376, 0.022726655960083008, 0.022544384002685547, 0.022589216232299803, 0.02269388771057129, 0.022714591979980468, 0.022597408294677733, 0.02259312057495117, 0.022571264266967775, 0.02267900848388672, 0.02264566421508789, 0.02266444778442383, 0.022624895095825194, 0.022583423614501955, 0.02263859176635742, 0.022592735290527344, 0.022708703994750976, 0.022874431610107424, 0.02263859176635742, 0.02262406349182129, 0.022536384582519532, 0.022613407135009766, 0.022768224716186523, 0.022724607467651366, 0.02257891273498535, 0.022655263900756836, 0.02259702491760254, 0.02272220802307129, 0.022696895599365233, 0.02273276710510254, 0.022601247787475586, 0.022603328704833985, 0.02274604797363281, 0.022654815673828124, 0.023660703659057616, 0.025016319274902343, 0.022773759841918945, 0.02270537567138672, 0.022602527618408204, 0.022808576583862306, 0.02279609680175781, 0.022894784927368163, 0.022680767059326173, 0.022702688217163085, 0.022847360610961914, 0.02253561592102051, 0.02264361572265625, 0.02231167984008789, 0.022517759323120116, 0.022624128341674803, 0.022509855270385744, 0.022627199172973633, 0.02256355285644531, 0.022693439483642577, 0.022663135528564454, 0.02260259246826172, 0.022480768203735353, 0.022665216445922853, 0.022527776718139648, 0.022604000091552733, 0.022552576065063477, 0.022788095474243163, 0.022509567260742186, 0.022744640350341797, 0.02253664016723633, 0.022632192611694336, 0.02247091293334961, 0.022665216445922853, 0.022652032852172852, 0.022573856353759764, 0.022587488174438477, 0.02254643249511719, 0.022529439926147463, 0.02256752014160156, 0.023143583297729493, 0.022706464767456056, 0.022573631286621095, 0.022455455780029297, 0.022658048629760744, 0.022519647598266603, 0.022550207138061523, 0.02265497589111328, 0.02252217674255371, 0.022648832321166993, 0.022580255508422853, 0.022686687469482422, 0.022563871383666993, 0.022875104904174805, 0.022519168853759767, 0.02256732749938965, 0.02263372802734375, 0.0225515193939209, 0.022507360458374023, 0.022468767166137694, 0.02244918441772461, 0.022559711456298828, 0.02240057563781738, 0.022591936111450196, 0.022510623931884764, 0.022598047256469727, 0.022556447982788087, 0.02247555160522461, 0.02248294448852539, 0.022582880020141603, 0.02257315254211426, 0.0234616641998291, 0.02460316848754883, 0.02282284736633301, 0.022814783096313476, 0.022632448196411133, 0.022362720489501952, 0.022566240310668947, 0.022702432632446288, 0.02259779167175293, 0.02267145538330078, 0.022560127258300783, 0.022575807571411134, 0.022517183303833007, 0.02251628875732422, 0.02251571273803711, 0.022552576065063477, 0.022503423690795898, 0.022681600570678712, 0.022561920166015624, 0.022489248275756837, 0.022577728271484375, 0.022519968032836915, 0.022697984695434572, 0.0226376953125, 0.022613920211791993, 0.022827999114990234, 0.023748607635498048, 0.022558176040649414, 0.022574623107910155, 0.022635679244995117, 0.02258927917480469, 0.022638240814208985, 0.022595935821533204, 0.022531232833862304, 0.025098495483398438, 0.02370368003845215, 0.02265555191040039, 0.022599584579467775, 0.023855104446411132, 0.02251468849182129, 0.022518783569335937, 0.02249728012084961, 0.022834400177001953, 0.022519647598266603, 0.022885311126708986, 0.022566879272460937, 0.022468511581420898, 0.022540416717529297, 0.02256208038330078, 0.02256764793395996, 0.022646751403808594, 0.02292937660217285, 0.022545536041259764, 0.022805440902709962, 0.02251571273803711, 0.022519712448120118, 0.023029855728149414, 0.02270412826538086, 0.02246246337890625, 0.022432992935180664, 0.022457120895385742, 0.022574880599975585, 0.022614240646362305, 0.02253824043273926, 0.022591360092163087, 0.02248716735839844, 0.022689279556274415, 0.022703903198242188, 0.022299264907836912, 0.022494911193847656, 0.0225532169342041, 0.02258243179321289, 0.02250428771972656, 0.022544448852539062, 0.022534080505371094, 0.022564863204956053, 0.022627487182617187, 0.022737728118896485, 0.02265910339355469, 0.022674943923950194, 0.022577600479125978, 0.02281667137145996, 0.022612127304077148, 0.02271753692626953, 0.022607967376708983, 0.022966495513916017, 0.022790016174316405, 0.02256355285644531, 0.022580671310424804, 0.02261459159851074, 0.02264473533630371, 0.0228306884765625, 0.022599456787109375, 0.022725248336791994, 0.022959423065185548, 0.02285212707519531, 0.02281484794616699, 0.022793855667114258, 0.023158367156982423, 0.023151039123535156, 0.022847808837890626, 0.022698047637939454, 0.022533920288085936, 0.022860000610351563, 0.022712160110473632, 0.022759136199951173, 0.022655424118041993, 0.022617088317871094, 0.02270096015930176, 0.024053855895996092, 0.022687744140625, 0.0231646728515625, 0.02378691291809082, 0.02313916778564453, 0.023580671310424805, 0.02327756881713867, 0.023011327743530274, 0.023003135681152344, 0.02267900848388672, 0.022501632690429686, 0.02258358383178711, 0.022763519287109374, 0.022541919708251954, 0.022651296615600586, 0.022665216445922853, 0.022906879425048828, 0.022565887451171874, 0.022780927658081054, 0.0226712646484375, 0.022857824325561524, 0.023045631408691408, 0.023267328262329103, 0.025502784729003906, 0.023565248489379884, 0.023369888305664062, 0.02298784065246582, 0.022897439956665037, 0.022775392532348632, 0.022647199630737306, 0.02268569564819336, 0.022588735580444337, 0.023145151138305665, 0.022662336349487305, 0.022797119140625, 0.022619680404663087, 0.02268822479248047, 0.022749183654785156, 0.022666240692138673, 0.022598560333251954, 0.02258070373535156, 0.022796384811401366, 0.023132608413696288, 0.02312396812438965, 0.022953184127807617, 0.022756223678588868, 0.023138080596923828, 0.022739360809326172, 0.022638399124145506, 0.022509056091308592, 0.02270787239074707, 0.022408031463623048, 0.022406496047973633, 0.022540735244750976, 0.022442207336425782, 0.022571008682250978, 0.023860416412353515, 0.022936384201049806, 0.023011327743530274, 0.022929311752319336, 0.022605920791625978, 0.022517759323120116, 0.02255411148071289, 0.022850048065185546, 0.022381792068481444, 0.02248137664794922, 0.022339487075805665, 0.02253593635559082, 0.02250579261779785, 0.022415712356567384, 0.022404191970825195, 0.022564895629882814, 0.022582271575927734, 0.022523391723632814, 0.022540672302246094, 0.022493152618408202, 0.022538272857666016, 0.022552255630493165, 0.022464479446411133, 0.022776159286499023, 0.022968320846557616, 0.02319919967651367, 0.02296272087097168, 0.02310758399963379, 0.023037343978881835]",tokens/s,43.93602104148196,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1545.527296,1546.584064,0.0,1168.113664,1154.613248,s,1,8.5124970703125,8.5124970703125,0.0,8.5124970703125,8.5124970703125,8.5124970703125,8.5124970703125,[8.5124970703125],,kWh,4.064591162913681e-05,4.4762609561822e-06,1.3347788456019583e-05,5.846996104133859e-05,,MB,1650.245632,1777.270784,0.0,1369.440256,1323.44832,s,10,0.7711464309692383,0.07711464309692383,0.0010966688767949236,0.07673311996459961,0.07833795166015625,0.07897385406494141,0.07948257598876954,"[0.07960975646972657, 0.07690499114990235, 0.07630387115478515, 0.07656124877929688, 0.0779483871459961, 0.07723251342773438, 0.07819664001464843, 0.07608220672607421, 0.0761595230102539, 0.07614729309082031]",tokens/s,3319.732669685559,kWh,2.497256980805001e-06,2.7540118011858825e-07,1.6537216458632824e-06,4.426379806786871e-06,tokens/kWh,57835073.16915752,MB,1656.844288,1798.242304,0.0,1390.411776,1377.26208,s,10,15.17874658203125,1.517874658203125,0.008304287893454717,1.5187279052734375,1.5292023071289063,1.5295946838378907,1.5299085852050782,"[1.500471923828125, 1.511712646484375, 1.529987060546875, 1.516548583984375, 1.5182763671875, 1.519807861328125, 1.5225753173828125, 1.511072265625, 1.5291151123046876, 1.519179443359375]",tokens/s,41.50540340042308,kWh,4.4015917780858925e-05,4.8546132297960675e-06,2.1319602288336758e-05,7.019013329899176e-05,tokens/kWh,897562.0509457695,,s,630,15.17503857803344,0.024087362822275315,0.0004091950367639072,0.024088511466979982,0.024378726768493654,0.02466751842498779,0.025500861911773684,"[0.02437936019897461, 0.023972448348999024, 0.023807743072509765, 0.023576576232910155, 0.023678432464599608, 0.023594079971313478, 0.02373017692565918, 0.023701248168945314, 0.023993471145629882, 0.023794559478759764, 0.02387104034423828, 0.023673280715942383, 0.024012256622314453, 0.023677471160888672, 0.023459968566894532, 0.023271295547485353, 0.023171072006225587, 0.023131904602050782, 0.0233221435546875, 0.023519264221191407, 0.02350150489807129, 0.02366873550415039, 0.02379385566711426, 0.023779136657714844, 0.02360495948791504, 0.023713632583618163, 0.023409088134765624, 0.023418880462646483, 0.023478111267089843, 0.023436864852905272, 0.023394048690795897, 0.023476512908935546, 0.023571104049682618, 0.02390185546875, 0.024166656494140626, 0.023975231170654296, 0.023651008605957032, 0.023588064193725587, 0.02362057685852051, 0.02380780792236328, 0.024567808151245117, 0.024211456298828125, 0.02431795120239258, 0.024098880767822267, 0.024003711700439453, 0.024068927764892577, 0.02411097526550293, 0.024008832931518554, 0.024016895294189454, 0.02410495948791504, 0.024246271133422852, 0.024086528778076172, 0.02418182373046875, 0.02414419174194336, 0.02423075294494629, 0.024092447280883788, 0.024194400787353517, 0.024008832931518554, 0.02395529556274414, 0.023833280563354493, 0.023826656341552736, 0.023807775497436522, 0.02379136085510254, 0.02450127983093262, 0.023922975540161134, 0.02467500877380371, 0.02388787269592285, 0.02380384063720703, 0.023787776947021486, 0.023928640365600586, 0.024093727111816406, 0.024026079177856444, 0.02394028854370117, 0.024025920867919923, 0.02408857536315918, 0.02396134376525879, 0.024058176040649415, 0.025439647674560546, 0.024270751953125, 0.023966239929199218, 0.02395350456237793, 0.024376800537109375, 0.024062496185302734, 0.024071392059326173, 0.023978784561157228, 0.023937023162841797, 0.02371788787841797, 0.02389740753173828, 0.024273151397705077, 0.024228288650512696, 0.024637439727783202, 0.024116800308227538, 0.024220096588134767, 0.02408768081665039, 0.02391334342956543, 0.02405743980407715, 0.023980159759521485, 0.02400284767150879, 0.023922016143798828, 0.0240994873046875, 0.023992191314697264, 0.023870719909667968, 0.024066944122314454, 0.02400886344909668, 0.023969215393066408, 0.023800159454345705, 0.023639488220214843, 0.023583423614501952, 0.023668672561645506, 0.02391244888305664, 0.02393071937561035, 0.02365827178955078, 0.023666847229003907, 0.023613664627075197, 0.024123231887817384, 0.02472326469421387, 0.023953344345092772, 0.023730592727661134, 0.02372825622558594, 0.023877504348754883, 0.02366454315185547, 0.023640159606933595, 0.023654495239257813, 0.02391155242919922, 0.02352617645263672, 0.023592832565307618, 0.024359743118286134, 0.024160255432128908, 0.02406768035888672, 0.02438800048828125, 0.024246271133422852, 0.024163871765136718, 0.024242656707763672, 0.024176639556884767, 0.024131584167480468, 0.024176607131958006, 0.024121376037597658, 0.02495699119567871, 0.024242111206054687, 0.024242271423339845, 0.02427280044555664, 0.024328191757202147, 0.0241889591217041, 0.024252384185791016, 0.024429664611816407, 0.024268960952758788, 0.02427907180786133, 0.02436751937866211, 0.0242936954498291, 0.024228031158447266, 0.024309215545654298, 0.024185184478759766, 0.02429132843017578, 0.02474185562133789, 0.02475391960144043, 0.024153791427612304, 0.024226400375366212, 0.02410041618347168, 0.02412384033203125, 0.024010751724243166, 0.024054847717285156, 0.024081344604492187, 0.023986175537109376, 0.023957311630249025, 0.02391059112548828, 0.02365545654296875, 0.024316896438598634, 0.025343360900878905, 0.025043487548828125, 0.024055904388427734, 0.024252416610717774, 0.02393907165527344, 0.02415001678466797, 0.023953407287597657, 0.02395136070251465, 0.023917631149291994, 0.02395849609375, 0.023862367630004884, 0.024007776260375976, 0.023930431365966797, 0.02397536087036133, 0.023990175247192384, 0.024030080795288088, 0.024069568634033204, 0.024582496643066408, 0.02712704086303711, 0.02558470344543457, 0.024188671112060547, 0.02418294334411621, 0.024673696517944335, 0.024170751571655272, 0.02562713623046875, 0.024176544189453125, 0.02420524787902832, 0.024092832565307618, 0.02418649673461914, 0.024152448654174805, 0.023966815948486327, 0.02401353645324707, 0.024149248123168945, 0.024132543563842774, 0.024024511337280275, 0.02394767951965332, 0.023879615783691407, 0.023731903076171876, 0.023832895278930663, 0.023824607849121094, 0.02382374382019043, 0.023801759719848634, 0.02409766387939453, 0.02383856010437012, 0.023736320495605468, 0.023572032928466796, 0.023383743286132814, 0.023509759902954102, 0.02342092704772949, 0.023650304794311523, 0.02354956817626953, 0.023607519149780272, 0.02384707260131836, 0.02376495933532715, 0.024141599655151367, 0.02406387138366699, 0.024033855438232422, 0.024305471420288084, 0.024108608245849608, 0.02413609504699707, 0.024121376037597658, 0.024161983489990234, 0.024288576126098634, 0.02433967971801758, 0.02418262481689453, 0.024206975936889648, 0.024233375549316406, 0.024211904525756837, 0.024222272872924805, 0.024344480514526368, 0.024336383819580077, 0.0243056640625, 0.024225088119506837, 0.024264991760253905, 0.024332223892211916, 0.024271327972412108, 0.024199167251586915, 0.024156160354614258, 0.024262655258178712, 0.024088191986083984, 0.024072576522827148, 0.023984128952026368, 0.02393622398376465, 0.024133920669555664, 0.024082944869995116, 0.02464543914794922, 0.02407606315612793, 0.024086751937866212, 0.024125631332397462, 0.024178144454956054, 0.024011104583740235, 0.023903520584106445, 0.0240032958984375, 0.023894208908081055, 0.02380985641479492, 0.023814144134521483, 0.0238983039855957, 0.02390415954589844, 0.02405571174621582, 0.024071615219116212, 0.024039680480957032, 0.02411552047729492, 0.024031007766723633, 0.02454550361633301, 0.025049087524414062, 0.02415964889526367, 0.024158815383911132, 0.024141727447509767, 0.024043615341186524, 0.023995807647705078, 0.02444553565979004, 0.02415420722961426, 0.02451241683959961, 0.024352767944335937, 0.024180543899536132, 0.024037567138671875, 0.024211616516113282, 0.024256223678588866, 0.024266880035400392, 0.024215551376342775, 0.02430975914001465, 0.024406112670898438, 0.024133344650268555, 0.024111328125, 0.024120447158813476, 0.02457676887512207, 0.024182880401611328, 0.02421366310119629, 0.02419046401977539, 0.024178592681884766, 0.024291776657104493, 0.02434230422973633, 0.024350303649902344, 0.02434480094909668, 0.024109312057495117, 0.0240231990814209, 0.024041471481323243, 0.024094720840454102, 0.023973184585571287, 0.02374928092956543, 0.023680992126464843, 0.024297536849975584, 0.023281280517578124, 0.023413055419921874, 0.023461919784545898, 0.02351241683959961, 0.02389219284057617, 0.02336591911315918, 0.024352224349975585, 0.024123872756958008, 0.024131040573120117, 0.023820703506469726, 0.023779520034790037, 0.023820127487182617, 0.024006591796875, 0.024205408096313476, 0.024202720642089844, 0.023976608276367186, 0.023904287338256835, 0.023658592224121092, 0.02369113540649414, 0.02392678451538086, 0.024000192642211916, 0.02394688034057617, 0.024320703506469726, 0.023963647842407225, 0.023733375549316406, 0.023901248931884767, 0.023914304733276368, 0.024053760528564453, 0.02391244888305664, 0.023955520629882814, 0.023927967071533204, 0.024025888442993165, 0.02399807929992676, 0.023912832260131835, 0.023953407287597657, 0.0243786563873291, 0.024249055862426757, 0.024205312728881836, 0.02465996742248535, 0.024195072174072265, 0.024284799575805663, 0.024426399230957033, 0.024235519409179687, 0.024265632629394532, 0.024172607421875, 0.024182783126831055, 0.02410700798034668, 0.024204896926879882, 0.02420368003845215, 0.024188512802124022, 0.024514976501464843, 0.024215295791625978, 0.024199424743652345, 0.024125375747680665, 0.02424835205078125, 0.024082527160644532, 0.024076223373413086, 0.02411315155029297, 0.024195072174072265, 0.02416640090942383, 0.02411315155029297, 0.024307968139648438, 0.024267776489257813, 0.02432080078125, 0.024238016128540037, 0.02412950325012207, 0.02425369644165039, 0.024361440658569336, 0.02434079933166504, 0.024786815643310547, 0.02611782455444336, 0.025200735092163085, 0.02437379264831543, 0.02435478401184082, 0.02418262481689453, 0.024242176055908202, 0.025047040939331053, 0.0242192325592041, 0.024297664642333985, 0.02423740768432617, 0.024193119049072266, 0.024201824188232423, 0.024145984649658205, 0.024189056396484374, 0.024174591064453126, 0.024176223754882813, 0.02422006416320801, 0.02425651168823242, 0.02454732894897461, 0.024207359313964845, 0.02424025535583496, 0.02429939270019531, 0.02433433532714844, 0.02412860870361328, 0.024162879943847658, 0.024125791549682616, 0.024258495330810547, 0.02417465591430664, 0.024231807708740234, 0.024237375259399414, 0.024224576950073243, 0.02415001678466797, 0.02400377655029297, 0.024060735702514647, 0.024006431579589843, 0.024006559371948243, 0.023999935150146486, 0.023795711517333985, 0.023837568283081055, 0.02384492874145508, 0.023586879730224608, 0.02353753662109375, 0.023535360336303712, 0.023472639083862306, 0.023443008422851564, 0.023470272064208986, 0.023346431732177736, 0.023503711700439453, 0.023560096740722656, 0.02437459182739258, 0.024555999755859374, 0.02450351905822754, 0.0245032958984375, 0.02405171203613281, 0.024139776229858398, 0.024138816833496092, 0.02406822395324707, 0.024126272201538086, 0.02411929512023926, 0.024207359313964845, 0.024084320068359377, 0.024305503845214845, 0.024704639434814452, 0.024228384017944336, 0.0243507194519043, 0.02428108787536621, 0.024401920318603516, 0.024278976440429687, 0.02440096092224121, 0.024318815231323242, 0.02438159942626953, 0.024391679763793944, 0.02433807945251465, 0.02430191993713379, 0.024373279571533204, 0.02414588737487793, 0.024207359313964845, 0.025393152236938478, 0.024238336563110353, 0.02380303955078125, 0.023583135604858398, 0.023893375396728516, 0.02386207962036133, 0.023749919891357423, 0.023941856384277344, 0.023820480346679686, 0.023523136138916014, 0.023402496337890624, 0.023322080612182616, 0.023273439407348634, 0.023329343795776367, 0.023592960357666014, 0.023605024337768555, 0.024938720703125, 0.023723264694213868, 0.02356710433959961, 0.023899520874023437, 0.023808063507080077, 0.02392531204223633, 0.02349056053161621, 0.02355187225341797, 0.023472095489501955, 0.023436511993408203, 0.023344064712524416, 0.023465728759765624, 0.023649631500244142, 0.023675264358520506, 0.023783967971801757, 0.024000511169433594, 0.02392460823059082, 0.023875104904174806, 0.025107040405273437, 0.023931968688964845, 0.02396883201599121, 0.02390412712097168, 0.024176607131958006, 0.02415363121032715, 0.02396726417541504, 0.023992767333984377, 0.024110687255859374, 0.024029760360717772, 0.024022464752197267, 0.023944128036499025, 0.02412748718261719, 0.024258687973022462, 0.024623104095458984, 0.02415555191040039, 0.025356224060058594, 0.024563776016235352, 0.024957536697387695, 0.024266752243041992, 0.024252031326293947, 0.024188608169555665, 0.02425107192993164, 0.024180736541748047, 0.02429120063781738, 0.02444300842285156, 0.024160480499267577, 0.02402895927429199, 0.0242872314453125, 0.02403296089172363, 0.024048063278198244, 0.023918464660644533, 0.02369536018371582, 0.02364825630187988, 0.02384486389160156, 0.023961599349975587, 0.023967744827270508, 0.02413132858276367, 0.02396771240234375, 0.023744800567626952, 0.023764991760253908, 0.023938688278198242, 0.024058143615722657, 0.02416854476928711, 0.024066047668457033, 0.02403443145751953, 0.024102783203125, 0.024017919540405275, 0.023957504272460937, 0.024164159774780272, 0.024223936080932616, 0.024253856658935546, 0.02418662452697754, 0.024176544189453125, 0.02424928092956543, 0.024848512649536133, 0.02424323272705078, 0.024302080154418947, 0.02416828727722168, 0.02404812812805176, 0.024203264236450195, 0.024151647567749023, 0.024191488265991212, 0.024165311813354493, 0.02417056083679199, 0.02404035186767578, 0.02413884735107422, 0.024492704391479492, 0.024123647689819335, 0.024172544479370117, 0.02415001678466797, 0.024235551834106445, 0.027676288604736327, 0.025770368576049803, 0.024306272506713866, 0.024530208587646485, 0.024254623413085936, 0.024817472457885743, 0.024364927291870116, 0.02446953582763672, 0.02440220832824707, 0.025464895248413087, 0.024067968368530274, 0.02406985664367676, 0.024103263854980468, 0.023832576751708984, 0.024138912200927735, 0.02516873550415039, 0.02427622413635254, 0.024103679656982423, 0.02383078384399414, 0.02396953582763672, 0.02398745536804199, 0.0238654727935791, 0.02381452751159668, 0.023568639755249022, 0.02420047950744629, 0.025280256271362305, 0.024034048080444338, 0.0248853759765625, 0.025515552520751952, 0.02379225540161133, 0.023761856079101563, 0.02365951919555664, 0.02346780776977539, 0.02368943977355957, 0.023474271774291993, 0.023490272521972656, 0.023414079666137694, 0.023658784866333007, 0.024224096298217774, 0.024170400619506836, 0.0241298885345459, 0.024070144653320313, 0.024008703231811524, 0.024016895294189454, 0.024153472900390625, 0.024132192611694334, 0.024057376861572267, 0.024082944869995116, 0.02396988868713379, 0.02408844757080078, 0.02396342468261719, 0.024019168853759765, 0.02382236862182617, 0.023940160751342775, 0.024053983688354492, 0.024126176834106446, 0.024201215744018553, 0.024000511169433594, 0.023973888397216796, 0.02429132843017578, 0.02409062385559082, 0.02409267234802246, 0.02434662437438965, 0.024172544479370117, 0.023940095901489256, 0.024015775680541994, 0.024057855606079103, 0.023969120025634765]",tokens/s,41.51554520012578,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,6726.176768,7525.564416,0.0,7147.094016,7138.9184,s,1,11.4454169921875,11.4454169921875,0.0,11.4454169921875,11.4454169921875,11.4454169921875,11.4454169921875,[11.4454169921875],,kWh,0.00013004618234999722,1.4337578162869243e-05,4.364336824799753e-05,0.000188027128760864,,MB,1571.659776,8236.498944,0.0,7828.668416,7715.649536,s,10,7.225968872070313,0.7225968872070312,0.004365442437216515,0.7243254699707031,0.7271126220703125,0.7274005187988282,0.7276308361816406,"[0.71417431640625, 0.7217698364257813, 0.7167904052734375, 0.7242161254882813, 0.7270486450195313, 0.7276884155273438, 0.7265798950195312, 0.724434814453125, 0.7244411010742188, 0.7188253173828125]",tokens/s,354.2777508902463,kWh,2.1206718300000676e-05,2.3387232408683474e-06,1.4104039061001359e-05,3.764948060187038e-05,tokens/kWh,6799562.594424802,MB,1579.339776,8383.299584,0.0,7975.469056,7906.518528,s,10,30.671578369140633,3.0671578369140624,0.005155731992710853,3.0669029541015624,3.0737444580078126,3.075187292480469,3.076341560058594,"[3.063729248046875, 3.05644482421875, 3.0667685546875, 3.066685791015625, 3.06440234375, 3.06835498046875, 3.067037353515625, 3.068101318359375, 3.076630126953125, 3.073423828125]",tokens/s,20.540188457789228,kWh,8.943397612999658e-05,9.865027055089127e-06,5.941510308759398e-05,0.00015871410627267971,tokens/kWh,396940.142748008,,s,630,30.667361335754393,0.048678351326594274,0.0005182079638856422,0.04866662406921387,0.04917172470092774,0.04932351245880127,0.05109099552154541,"[0.051337215423583986, 0.04890745544433594, 0.0479754867553711, 0.04796390533447266, 0.04803337478637695, 0.04811407852172851, 0.04793462371826172, 0.047780288696289065, 0.048269695281982425, 0.04780563354492187, 0.048038753509521484, 0.047972255706787106, 0.04797353744506836, 0.048315486907958984, 0.048239776611328125, 0.04808492660522461, 0.04820659255981445, 0.04831379318237305, 0.04803334426879883, 0.04864284896850586, 0.04904982376098633, 0.04881734466552735, 0.04834137725830078, 0.04871939086914062, 0.04845865631103516, 0.048361377716064455, 0.04835932922363281, 0.04830003356933594, 0.04814828872680664, 0.04830579376220703, 0.04815539169311524, 0.04835123062133789, 0.04865993499755859, 0.048832126617431644, 0.04874924850463867, 0.048369056701660154, 0.04870332717895508, 0.04861231994628906, 0.0488524169921875, 0.04904198455810547, 0.04909260940551758, 0.04888723373413086, 0.05062508773803711, 0.04871987152099609, 0.04917855834960937, 0.04908044815063477, 0.04886111831665039, 0.04890192031860351, 0.04880815887451172, 0.048549121856689456, 0.04845030212402344, 0.04878540802001953, 0.04871148681640625, 0.048750526428222654, 0.04869350433349609, 0.04862963104248047, 0.04882979202270508, 0.04900742340087891, 0.04926393508911133, 0.049119583129882814, 0.048978271484375, 0.04898812866210937, 0.049258495330810545, 0.050716670989990234, 0.04861244964599609, 0.047969184875488284, 0.047816703796386716, 0.04761932754516601, 0.04799465560913086, 0.04792623901367187, 0.04797439956665039, 0.04782284927368164, 0.04793958282470703, 0.047834270477294924, 0.04819235229492187, 0.048043617248535155, 0.04801507186889648, 0.04802630233764649, 0.04819353485107422, 0.04816502380371094, 0.04828553771972656, 0.04819529724121094, 0.048646430969238284, 0.04853964614868164, 0.04842496109008789, 0.04851900863647461, 0.04810947036743164, 0.047901153564453125, 0.04805814361572266, 0.04834099197387695, 0.048414913177490235, 0.04822809600830078, 0.04825676727294922, 0.04831059265136719, 0.04826259231567383, 0.04821049499511719, 0.04847555160522461, 0.048619968414306644, 0.04860956954956055, 0.048510848999023436, 0.048623615264892575, 0.04864604949951172, 0.04878550338745117, 0.048936958312988284, 0.048788833618164065, 0.048876190185546876, 0.04879359817504883, 0.04890009689331055, 0.04869734573364258, 0.04889136123657226, 0.048579265594482425, 0.04861244964599609, 0.04873484802246094, 0.04895347213745117, 0.04878745651245117, 0.048895999908447264, 0.0485494384765625, 0.04864863967895508, 0.049067489624023436, 0.048812576293945316, 0.04926873779296875, 0.04898787307739258, 0.04899238586425781, 0.04885638427734375, 0.04934441757202149, 0.04918163299560547, 0.051087520599365235, 0.04877091217041016, 0.04802716827392578, 0.047852001190185546, 0.047874431610107425, 0.04789388656616211, 0.04806300735473633, 0.04790691375732422, 0.04790857696533203, 0.04794806289672852, 0.04842291259765625, 0.048447296142578124, 0.04839657592773437, 0.04839759826660156, 0.04820851135253906, 0.04842214584350586, 0.048571136474609374, 0.04841267013549805, 0.048484447479248044, 0.04891366577148438, 0.04909872055053711, 0.04882092666625976, 0.0485618896484375, 0.04847030258178711, 0.048345088958740234, 0.048418113708496094, 0.04834998321533203, 0.04833065414428711, 0.04802105712890625, 0.048185150146484376, 0.04825356674194336, 0.04840614318847656, 0.04863209533691406, 0.04876502227783203, 0.048617473602294924, 0.0486932487487793, 0.04850483322143555, 0.04866457748413086, 0.04898569488525391, 0.049041824340820314, 0.04927078247070313, 0.049057151794433596, 0.04896422576904297, 0.04908031845092774, 0.04910220718383789, 0.04905023956298828, 0.048707103729248045, 0.04873401641845703, 0.048712352752685546, 0.04892614364624023, 0.048783935546875, 0.04888595199584961, 0.04882153701782226, 0.048690849304199216, 0.04896764755249024, 0.049171360015869144, 0.048996353149414064, 0.04913071823120117, 0.04927372741699219, 0.04922278213500977, 0.0489664306640625, 0.04928006362915039, 0.049327041625976564, 0.05166169738769531, 0.04873011016845703, 0.04817251205444336, 0.04821456146240234, 0.04800511932373047, 0.048293888092041014, 0.047983905792236325, 0.048022239685058594, 0.048080894470214845, 0.04817657470703125, 0.04842879867553711, 0.048306880950927736, 0.04818732833862305, 0.048265407562255856, 0.048715774536132815, 0.048306270599365236, 0.04833238220214844, 0.04828192138671875, 0.04856012725830078, 0.04902822494506836, 0.04876172637939453, 0.04864156723022461, 0.04847625732421875, 0.04850307083129883, 0.048631584167480466, 0.04875433731079101, 0.048339839935302734, 0.0484035530090332, 0.04832447814941406, 0.048593505859375, 0.0486907844543457, 0.048489086151123045, 0.048395999908447264, 0.048621631622314455, 0.048462047576904296, 0.048413726806640626, 0.048589790344238284, 0.04860313415527344, 0.04868627166748047, 0.04889190292358398, 0.04910976028442383, 0.04903430557250977, 0.04896819305419922, 0.049180831909179684, 0.04898646545410156, 0.048983776092529296, 0.0486607666015625, 0.04867891311645508, 0.04869529724121094, 0.048715774536132815, 0.048645633697509766, 0.04857062530517578, 0.04860879898071289, 0.048863967895507815, 0.04887142562866211, 0.04866867065429688, 0.04902707290649414, 0.04908982467651367, 0.049072319030761716, 0.04938108825683594, 0.04915488052368164, 0.049121280670166016, 0.04919311904907227, 0.05109241485595703, 0.04869721603393555, 0.04802764892578125, 0.048070655822753904, 0.04806576156616211, 0.04838604736328125, 0.04822844696044922, 0.04841337585449219, 0.04816915130615235, 0.04813804626464844, 0.04844748687744141, 0.04836457443237305, 0.04822073745727539, 0.048267166137695314, 0.04842342376708984, 0.04849382400512695, 0.048315135955810544, 0.048137569427490236, 0.04844543838500977, 0.048966304779052734, 0.04897792053222656, 0.04883865737915039, 0.04874649429321289, 0.04839219284057617, 0.04818124771118164, 0.04829740905761719, 0.04845209503173828, 0.0483326416015625, 0.04827936172485352, 0.04829430389404297, 0.04845676803588867, 0.04834400177001953, 0.04841062545776367, 0.048353279113769534, 0.04862566375732422, 0.04882339096069336, 0.04864441680908203, 0.04885974502563477, 0.0490967025756836, 0.049036865234375, 0.049050048828125, 0.04918201446533203, 0.04918137741088867, 0.04894425582885742, 0.04888051223754883, 0.04877878570556641, 0.048525791168212894, 0.04853952026367187, 0.048650367736816406, 0.04861692810058594, 0.048700000762939455, 0.04850067138671875, 0.04852912139892578, 0.04878950500488281, 0.0486640625, 0.048788257598876957, 0.048822048187255856, 0.04870576095581055, 0.04908425521850586, 0.0491333122253418, 0.04921782302856445, 0.04912140655517578, 0.04888345718383789, 0.05115887832641602, 0.0489697265625, 0.048191104888916016, 0.0480321273803711, 0.047865856170654295, 0.04808041763305664, 0.04803606414794922, 0.0481446418762207, 0.048205825805664064, 0.047917057037353515, 0.04813107299804688, 0.04844044876098633, 0.04828339385986328, 0.04810732650756836, 0.04827987289428711, 0.04821171188354492, 0.04808291244506836, 0.048451583862304685, 0.048551551818847655, 0.04882908630371094, 0.04904550552368164, 0.04890560150146484, 0.04854556655883789, 0.04856422424316406, 0.04831318283081055, 0.04836268615722656, 0.04827190399169922, 0.048259361267089844, 0.048601089477539064, 0.04828492736816406, 0.048226688385009764, 0.048621952056884764, 0.048347137451171876, 0.048830463409423826, 0.04877721786499024, 0.048773120880126954, 0.048748542785644534, 0.048656383514404294, 0.04906777572631836, 0.04908841705322266, 0.04927667236328125, 0.048976158142089846, 0.0488135986328125, 0.04905859375, 0.0488611831665039, 0.04900044631958008, 0.04878287887573242, 0.048852832794189456, 0.04865087890625, 0.04931174468994141, 0.0489345588684082, 0.04872022247314453, 0.04881123352050781, 0.04872476959228516, 0.048920574188232424, 0.0490885124206543, 0.04903055953979492, 0.0490379524230957, 0.049557472229003904, 0.049495742797851565, 0.04920556640625, 0.04908854293823242, 0.049436641693115235, 0.05170995330810547, 0.04914585494995117, 0.04816691207885742, 0.04819257736206055, 0.04774803161621094, 0.04834406280517578, 0.04811983871459961, 0.0480467529296875, 0.04791843032836914, 0.04809417724609375, 0.04802348709106445, 0.048019519805908205, 0.0482979850769043, 0.04832441711425781, 0.048093376159667967, 0.04824883270263672, 0.0481341438293457, 0.04832255935668945, 0.048336894989013675, 0.048795326232910156, 0.04898643112182617, 0.04898988723754883, 0.048775489807128904, 0.04847411346435547, 0.048345088958740234, 0.048387134552001956, 0.048140384674072265, 0.04812275314331055, 0.04810134506225586, 0.048387935638427734, 0.04823011016845703, 0.04837599945068359, 0.04852556610107422, 0.048643840789794925, 0.04840041732788086, 0.04840665435791015, 0.04871177673339844, 0.048633281707763674, 0.048925342559814455, 0.04911299133300781, 0.049102783203125, 0.04939795303344727, 0.04926451110839844, 0.048971294403076175, 0.04914742279052734, 0.049080894470214846, 0.04887948989868164, 0.048615966796875, 0.048573982238769534, 0.04876537704467773, 0.04873023986816406, 0.0487784309387207, 0.049079200744628904, 0.04896681594848633, 0.04886592102050781, 0.04859699249267578, 0.0488590087890625, 0.0489463996887207, 0.04892079925537109, 0.049498817443847654, 0.04955104064941406, 0.04975033569335938, 0.049458335876464844, 0.051140960693359376, 0.04872947311401367, 0.04797817611694336, 0.04787222290039062, 0.04774918365478516, 0.04783967971801758, 0.04783536148071289, 0.048130046844482424, 0.04829183959960937, 0.04821958541870117, 0.048263744354248045, 0.04830729675292969, 0.048347198486328125, 0.04826780700683594, 0.04844166564941406, 0.04839014434814453, 0.048411808013916015, 0.048536350250244144, 0.048356960296630856, 0.048691551208496095, 0.04892409515380859, 0.04893727874755859, 0.0487407341003418, 0.04873622512817383, 0.04838790512084961, 0.04823823928833008, 0.048075328826904296, 0.04831436920166016, 0.04874230575561524, 0.048739681243896486, 0.04853631973266601, 0.04841600036621094, 0.04865081787109375, 0.048736446380615236, 0.048662559509277344, 0.04869526290893555, 0.04870348739624023, 0.048653438568115236, 0.048949535369873044, 0.04881782531738281, 0.04903523254394531, 0.04885295867919922, 0.04942335891723633, 0.049190814971923826, 0.04900048065185547, 0.04908448028564453, 0.04870963287353516, 0.04873625564575195, 0.048874561309814456, 0.0488969612121582, 0.048775169372558595, 0.04883014297485352, 0.04880374526977539, 0.04849926376342773, 0.0488642578125, 0.04913443374633789, 0.04904912185668946, 0.04916374588012695, 0.04917500686645508, 0.049557727813720705, 0.04928316879272461, 0.04916604614257813, 0.04910335922241211, 0.05163417434692383, 0.049086463928222655, 0.04826057434082031, 0.04834560012817383, 0.04815187072753906, 0.048159358978271484, 0.04841891098022461, 0.04825696182250976, 0.04817887878417969, 0.047958145141601564, 0.04802585601806641, 0.04824883270263672, 0.04841471862792969, 0.048449535369873044, 0.04828979110717774, 0.04844342422485352, 0.04837577438354492, 0.048645694732666014, 0.048814208984375, 0.049041728973388675, 0.049430335998535156, 0.049503711700439455, 0.04905344009399414, 0.04879049682617188, 0.04870943832397461, 0.048801822662353514, 0.048360862731933595, 0.048489215850830075, 0.0485294075012207, 0.04857833480834961, 0.04853168106079102, 0.04859494400024414, 0.04878716659545899, 0.04882460784912109, 0.04888576126098633, 0.04879359817504883, 0.049061313629150394, 0.04906246566772461, 0.04927897644042969, 0.049209342956542966, 0.04924415969848633, 0.0493873291015625, 0.04915577697753906, 0.04910540771484375, 0.048963584899902345, 0.04887673568725586, 0.048931648254394534, 0.04866217422485351, 0.04908886337280274, 0.04897766494750976, 0.0486761589050293, 0.0489944953918457, 0.048715808868408206, 0.04884515380859375, 0.048724063873291014, 0.048820606231689455, 0.04909660720825195, 0.049225727081298826, 0.04928409576416016, 0.04932675170898437, 0.049434112548828124, 0.04906409454345703, 0.04916678237915039, 0.05058281707763672, 0.04844412612915039, 0.04788019180297851, 0.04783513641357422, 0.04787353515625, 0.04844819259643555, 0.04811161422729492, 0.04827654266357422, 0.048287841796875, 0.04855878448486328, 0.048553951263427736, 0.04829523086547852, 0.04845228958129883, 0.048476158142089845, 0.04845568084716797, 0.048508384704589846, 0.04835382461547852, 0.04833280181884766, 0.04857241439819336, 0.049391616821289064, 0.04919039916992188, 0.04842502212524414, 0.04864031982421875, 0.04833292770385742, 0.0486379508972168, 0.048684894561767576, 0.04843244934082031, 0.048780128479003905, 0.04887756729125976, 0.04882227325439453, 0.0486868782043457, 0.048774784088134765, 0.04886383819580078, 0.04867862319946289, 0.048809310913085935, 0.04889484786987305, 0.04877257537841797, 0.04877331161499023, 0.04908867263793945, 0.04931955337524414, 0.0491383056640625, 0.049188705444335935, 0.049000350952148435, 0.049030784606933595, 0.048853633880615234, 0.04892047882080078, 0.04897980880737305, 0.048981727600097655, 0.0489700813293457, 0.048946369171142576, 0.04880486297607422, 0.0490332145690918, 0.04877715301513672, 0.04898556900024414, 0.04901744079589844, 0.04884038543701172, 0.048922943115234374, 0.0490590705871582, 0.049330944061279296, 0.049225631713867186, 0.049332351684570314, 0.049290592193603516, 0.049166976928710936]",tokens/s,20.543012915346488,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,11125.399552,12225.282048,0.0,11846.811648,11814.785024,s,1,14.380548828125,14.380548828125,0.0,14.380548828125,14.380548828125,14.380548828125,14.380548828125,[14.380548828125],,kWh,0.0002051947944083281,2.2627169896284204e-05,6.826366572199039e-05,0.0002960856300266027,,MB,2025.725952,13158.514688,0.0,12750.68416,12632.68864,s,10,13.033351928710939,1.303335192871094,0.0030176019811829413,1.30350146484375,1.3065957397460939,1.307156756591797,1.3076055700683595,"[1.2995831298828124, 1.2985087890625, 1.3007293701171876, 1.302571044921875, 1.301576904296875, 1.304431884765625, 1.3059696044921876, 1.3077177734375, 1.3057923583984374, 1.3064710693359376]",tokens/s,196.4191570980771,kWh,3.8096743162915723e-05,4.20160009565377e-06,2.54181314456009e-05,6.77164747041704e-05,tokens/kWh,3780468.506642948,MB,2029.989888,13431.144448,0.0,13023.31392,12936.694272,s,10,48.8005390625,4.88005390625,0.005832230603029691,4.881221435546875,4.8853884765625,4.8875189453125,4.8892233203125,"[4.876849609375, 4.868734375, 4.87260205078125, 4.8784443359375, 4.88436376953125, 4.8896494140625, 4.8816982421875, 4.88074462890625, 4.88253759765625, 4.8849150390625]",tokens/s,12.909693460417397,kWh,0.00014267798116250106,1.5738480656990593e-05,9.457529788239882e-05,0.0002529917597018905,tokens/kWh,249019.96837460322,,s,630,48.79621814727783,0.07745431451948863,0.000700290987406876,0.07746585845947265,0.07827880630493164,0.07849911270141602,0.079734945602417,"[0.07999199676513671, 0.07639737701416016, 0.0753826904296875, 0.07479312133789062, 0.07575341033935547, 0.0764651870727539, 0.07665968322753906, 0.0768532485961914, 0.07630847930908204, 0.07698841857910156, 0.07695516967773437, 0.07716275024414063, 0.07774384307861328, 0.07737599945068359, 0.07691878509521484, 0.07713520050048828, 0.07645046234130859, 0.07700032043457031, 0.07659142303466797, 0.07694099426269531, 0.07653619384765625, 0.0762734375, 0.07755375671386719, 0.076770751953125, 0.07753183746337891, 0.07749779510498046, 0.07766278076171874, 0.07761100769042968, 0.07717241668701172, 0.07700220489501954, 0.07674348449707032, 0.07761087799072265, 0.07719747161865234, 0.07699481964111328, 0.0770184326171875, 0.07723455810546875, 0.07734054565429688, 0.07797164916992187, 0.0777541732788086, 0.07811862182617188, 0.07836924743652343, 0.07762032318115235, 0.07716547393798828, 0.07764582061767578, 0.07754752349853515, 0.07748198699951171, 0.0777359390258789, 0.07761920166015625, 0.07794656372070312, 0.07780178833007813, 0.07827865600585937, 0.07825408172607422, 0.07833920288085937, 0.07853555297851562, 0.07837286376953125, 0.07831324768066407, 0.07835465240478516, 0.0784336929321289, 0.07835298919677734, 0.07825408172607422, 0.07778902435302734, 0.07820899200439453, 0.0784975357055664, 0.07987193298339844, 0.07667190551757813, 0.07625631713867187, 0.07632972717285157, 0.0766343002319336, 0.07690774536132812, 0.07665510559082031, 0.07655043029785157, 0.07668736267089844, 0.07524352264404296, 0.07507465362548828, 0.07630726623535156, 0.07828070068359375, 0.07675299072265625, 0.07656175994873046, 0.07668745422363281, 0.07698694610595704, 0.07668313598632813, 0.07702950286865234, 0.07690444946289063, 0.07701280212402344, 0.07711353302001953, 0.07697612762451173, 0.07727308654785156, 0.07753727722167969, 0.07758233642578125, 0.07675494384765626, 0.07746150207519531, 0.07737526702880859, 0.077283203125, 0.0769682846069336, 0.07701248168945313, 0.07717938995361329, 0.07757004547119141, 0.07732838439941406, 0.07774329376220702, 0.07753401947021485, 0.07750617980957031, 0.07739568328857421, 0.07751337432861329, 0.07776025390625, 0.077275390625, 0.07776268768310547, 0.07724227142333985, 0.07694902038574218, 0.0773155517578125, 0.0774432601928711, 0.07782275390625, 0.07796694183349609, 0.07782841491699219, 0.07773776245117188, 0.07749049377441407, 0.07822745513916016, 0.07783958435058594, 0.07782479858398438, 0.07766630554199219, 0.07747980499267579, 0.07783414459228516, 0.07776483154296875, 0.0777534408569336, 0.07757059478759766, 0.07849132537841796, 0.078110595703125, 0.07980032348632812, 0.07654768371582031, 0.07603446197509765, 0.07628185272216798, 0.07658003234863281, 0.07645081329345703, 0.07648976135253906, 0.0760607681274414, 0.07629593658447266, 0.07701805114746094, 0.07666687774658203, 0.07742259216308593, 0.0774052505493164, 0.07685584259033203, 0.07745782470703125, 0.07658905792236329, 0.07712735748291015, 0.07725001525878906, 0.0773139877319336, 0.07686441802978515, 0.07624076843261719, 0.07762528228759766, 0.0772589111328125, 0.07702326202392579, 0.07736729431152344, 0.07758016204833984, 0.07744278717041016, 0.07758854675292969, 0.07741871643066406, 0.07684722900390625, 0.0766541748046875, 0.07709458923339843, 0.07702191925048828, 0.07731199645996094, 0.07689353942871094, 0.07730838775634766, 0.07757008361816406, 0.07766851043701171, 0.0775167999267578, 0.07765113830566406, 0.07742047882080078, 0.07821340942382812, 0.07727513885498047, 0.07757839965820312, 0.07765776062011719, 0.0775145263671875, 0.07760908508300782, 0.07752921295166015, 0.07760288238525391, 0.07750646209716797, 0.0776971206665039, 0.07751136016845703, 0.07765606689453125, 0.07830937957763671, 0.07824384307861328, 0.07803903961181641, 0.07749836730957031, 0.07750857543945312, 0.07770278167724609, 0.07808975982666015, 0.07785513305664063, 0.07825456237792969, 0.07838105773925781, 0.08045798492431641, 0.07623375701904297, 0.07686857604980468, 0.07663744354248046, 0.07620275115966797, 0.07719843292236328, 0.07636243438720704, 0.07667107391357422, 0.07659942626953126, 0.07678912353515625, 0.077056640625, 0.07693708801269532, 0.07681433868408204, 0.07693507385253906, 0.07678793334960937, 0.07703961944580077, 0.07701612854003906, 0.07697299194335938, 0.07710304260253906, 0.07665251159667968, 0.076912353515625, 0.07689810943603516, 0.07646265411376953, 0.07650220489501953, 0.07807469177246094, 0.07724851226806641, 0.07734067535400391, 0.07729151916503907, 0.07719926452636719, 0.07692502593994141, 0.07712694549560548, 0.07704214477539062, 0.07756006622314453, 0.07746489715576171, 0.07730451202392578, 0.07748812866210937, 0.07787725067138672, 0.07751475524902343, 0.07837411499023438, 0.07815452575683594, 0.0776253433227539, 0.07746681976318359, 0.07766838073730468, 0.07727372741699219, 0.0777811508178711, 0.07739507293701171, 0.0775582733154297, 0.07752880096435547, 0.07803766632080078, 0.07813484954833984, 0.07805734252929687, 0.07792406463623047, 0.0781566390991211, 0.07818150329589844, 0.0773333740234375, 0.07787462615966798, 0.07836492919921875, 0.07816019439697265, 0.07784857940673828, 0.07799967956542969, 0.07864979553222656, 0.0785264663696289, 0.07839129638671875, 0.07941907501220703, 0.07615885162353515, 0.07597049713134765, 0.076746337890625, 0.0766207046508789, 0.07650828552246093, 0.07702153778076172, 0.07669200134277344, 0.07639449310302734, 0.07783360290527344, 0.07718876647949219, 0.07700959777832031, 0.07788572692871094, 0.07724441528320312, 0.07688745880126953, 0.07706463623046875, 0.07700415802001953, 0.076847900390625, 0.07689215850830078, 0.07704985809326172, 0.07701708984375, 0.07694131469726563, 0.0773199691772461, 0.077070556640625, 0.07760662078857422, 0.07824559783935547, 0.07758086395263672, 0.07715824127197266, 0.07735052490234375, 0.07673420715332031, 0.07698307037353516, 0.07718860626220703, 0.07749427032470703, 0.07783475494384766, 0.07745327758789063, 0.0772874526977539, 0.07737958526611328, 0.07795027160644531, 0.07798985290527344, 0.07757091522216797, 0.07777471923828125, 0.0775167999267578, 0.07736115264892578, 0.07806928253173828, 0.07760867309570313, 0.07763785552978515, 0.07787904357910157, 0.07806031799316407, 0.07811446380615235, 0.07806601715087891, 0.07823715209960938, 0.07812969970703125, 0.07876812744140625, 0.0780206069946289, 0.07763148498535156, 0.07852236938476563, 0.0781979217529297, 0.07832457733154297, 0.07829708862304688, 0.07835852813720703, 0.07830290985107421, 0.07818067169189453, 0.07819798278808594, 0.07949008178710938, 0.07630947113037109, 0.07654399871826172, 0.0767279052734375, 0.07679571533203125, 0.07683452606201172, 0.07671817779541015, 0.07648336029052734, 0.077061279296875, 0.07672918701171876, 0.07650646209716797, 0.07709343719482421, 0.07784397125244141, 0.07778755187988282, 0.07696380615234374, 0.07658518218994141, 0.07724646759033203, 0.07698767852783203, 0.07731196594238281, 0.07696665954589844, 0.07759257507324219, 0.07744921875, 0.07677267456054687, 0.07754201507568359, 0.07744723510742188, 0.07741379547119141, 0.07773040008544922, 0.07749427032470703, 0.0771904296875, 0.07752777862548828, 0.07761212921142578, 0.07724739074707031, 0.07736524963378906, 0.07733964538574219, 0.07755852508544922, 0.07733478546142578, 0.07791206359863281, 0.07779235076904296, 0.07842243194580079, 0.07779993438720703, 0.07784595489501953, 0.07818070220947265, 0.07769865417480469, 0.07761984252929688, 0.0779447021484375, 0.07817842864990235, 0.07819058990478515, 0.0782171859741211, 0.07852655792236328, 0.07856278228759765, 0.07823407745361328, 0.07835823822021484, 0.0785223388671875, 0.07849353790283203, 0.07790435028076172, 0.07783968353271484, 0.0778779525756836, 0.07808159637451172, 0.07837535858154297, 0.07807708740234375, 0.0783532485961914, 0.07782505798339844, 0.07865647888183594, 0.07957488250732422, 0.07621440124511719, 0.07670076751708985, 0.07663410949707031, 0.07657545471191406, 0.07680223846435547, 0.07727922821044922, 0.07618150329589844, 0.07667021179199218, 0.07735491180419922, 0.07680857849121094, 0.07676509094238282, 0.07759315490722657, 0.07722975921630859, 0.07697644805908203, 0.07703961944580077, 0.07710924530029296, 0.07649638366699218, 0.0771690902709961, 0.07691004943847657, 0.07650128173828125, 0.07735337829589843, 0.07755152130126954, 0.07696534729003907, 0.07768256378173828, 0.07798441314697266, 0.07759574127197266, 0.0774395523071289, 0.07758182525634766, 0.0771654052734375, 0.07714406585693359, 0.077501953125, 0.07731158447265625, 0.0775052490234375, 0.0774487075805664, 0.0777808609008789, 0.07801439666748047, 0.07813555145263672, 0.07817833709716797, 0.07813308715820312, 0.07761564636230468, 0.07741667175292968, 0.07760691070556641, 0.07781692504882813, 0.07778934478759765, 0.07775465393066407, 0.0776475830078125, 0.07755648040771485, 0.07742864227294922, 0.07769430541992188, 0.07852256011962891, 0.07822188568115235, 0.07859110260009766, 0.0778944320678711, 0.07767782592773438, 0.07771631622314454, 0.07738572692871094, 0.07762739562988281, 0.07833372497558594, 0.07864956665039062, 0.0778156509399414, 0.07772994995117187, 0.07781075286865234, 0.08009136199951172, 0.07658531188964844, 0.07652806091308594, 0.07612006378173829, 0.07688972473144531, 0.07701129913330078, 0.07674781036376953, 0.07650816345214843, 0.07465984344482422, 0.07679587554931641, 0.07729328155517579, 0.07730207824707032, 0.0775393295288086, 0.07709852600097657, 0.07691107177734376, 0.07715020751953125, 0.0775816650390625, 0.07750624084472656, 0.07719216156005859, 0.07756739044189453, 0.0770456314086914, 0.0769625244140625, 0.07702877044677735, 0.077357666015625, 0.07743385314941406, 0.077591552734375, 0.07781715393066406, 0.07729427337646484, 0.07706009674072266, 0.07716659545898437, 0.07661347198486328, 0.0773625259399414, 0.0777347183227539, 0.07736844635009765, 0.07698883056640625, 0.07718147277832031, 0.07757817840576171, 0.07745331573486328, 0.07797350311279297, 0.07805084991455079, 0.07749430084228516, 0.07738582611083984, 0.07759702301025391, 0.0771618881225586, 0.07743472290039062, 0.07792511749267578, 0.07750569915771484, 0.07709782409667969, 0.07816397094726563, 0.07770870208740234, 0.07772220611572266, 0.07810368347167969, 0.0782610855102539, 0.07814726257324218, 0.07851404571533203, 0.07828015899658203, 0.07789055633544922, 0.07801446533203125, 0.07845273590087891, 0.0780738525390625, 0.07861622619628907, 0.07850019073486328, 0.07801036834716797, 0.08047411346435547, 0.07664435577392578, 0.07627571105957032, 0.0767070083618164, 0.07681497955322265, 0.07661353302001953, 0.07698255920410156, 0.076943359375, 0.07654605102539062, 0.07659878540039063, 0.07819725036621093, 0.07687570953369141, 0.07705811309814453, 0.07754259490966797, 0.07717356872558594, 0.07697612762451173, 0.07704576110839843, 0.07674620819091797, 0.07683945465087891, 0.07684710693359376, 0.07660323333740235, 0.07662403106689453, 0.07783382415771484, 0.07744713592529297, 0.07716643524169922, 0.07719967651367188, 0.07778870391845703, 0.07753139495849609, 0.07767501068115235, 0.07744921875, 0.07753727722167969, 0.0773855972290039, 0.07753536224365234, 0.07713942718505859, 0.07718351745605469, 0.0776495361328125, 0.07766668701171875, 0.0770169906616211, 0.07768271636962891, 0.07738492584228515, 0.07785763549804688, 0.07750656127929688, 0.07743852996826171, 0.07776227569580078, 0.07817084503173828, 0.0773570556640625, 0.07717887878417969, 0.07799993896484375, 0.07849116516113282, 0.07814995574951172, 0.07773423767089843, 0.07813734436035157, 0.0781740493774414, 0.07786454772949218, 0.07808016204833984, 0.07775379180908203, 0.07755260467529297, 0.07823072052001953, 0.07785104370117188, 0.07732265472412109, 0.07851213073730469, 0.07898316955566406, 0.07861820983886719, 0.08027935791015625, 0.07674899291992188, 0.07671097564697266, 0.0763544921875, 0.07640838623046875, 0.07664070129394532, 0.0769269790649414, 0.07666204833984375, 0.07673474884033203, 0.07739167785644531, 0.07709513854980468, 0.07702365112304688, 0.07747328186035156, 0.07767005157470704, 0.07717501068115235, 0.07696835327148438, 0.0770390396118164, 0.07671459197998047, 0.07682067108154297, 0.07663177490234375, 0.07666000366210937, 0.07721453094482422, 0.0771725082397461, 0.07710147094726562, 0.07824111938476562, 0.07744989013671875, 0.07724031829833984, 0.07784770965576172, 0.07711782073974609, 0.07745763397216797, 0.0771361312866211, 0.07749836730957031, 0.07702063751220703, 0.0776115493774414, 0.07762710571289062, 0.07722217559814454, 0.07788748931884766, 0.07796940612792969, 0.07777030181884766, 0.0775269775390625, 0.07794739532470703, 0.0780389404296875, 0.07736329650878906, 0.07764166259765624, 0.07728125, 0.07769916534423828, 0.07785676574707032, 0.0781496353149414, 0.07755554962158204, 0.07767382049560546, 0.07878489685058594, 0.07850972747802734, 0.0783081283569336, 0.07806361389160156, 0.07807718658447266, 0.07850879669189453, 0.07802432250976563, 0.07747212982177734, 0.07762671661376953, 0.07868073272705078, 0.07849779510498046, 0.07806742095947265, 0.07846736145019531]",tokens/s,12.910836616446788,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.747328,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3549.26592,4490.985472,0.0,4112.515072,3975.832064,s,1,9.7483212890625,9.7483212890625,0.0,9.7483212890625,9.7483212890625,9.7483212890625,9.7483212890625,[9.7483212890625],,kWh,8.036910985833382e-05,8.857929561269235e-06,2.728752183000019e-05,0.00011651456124960323,,MB,1460.297728,4688.11776,0.0,4280.287232,4101.544448,s,10,3.156000061035156,0.31560000610351563,0.0014197231289403167,0.3160218505859375,0.31709080505371096,0.317126579284668,0.3171551986694336,"[0.3167412109375, 0.3138393859863281, 0.3138292846679688, 0.31702584838867187, 0.31340982055664063, 0.3163779296875, 0.3148656005859375, 0.317162353515625, 0.3170828552246094, 0.315665771484375]",tokens/s,811.1533429946543,kWh,9.521249536693518e-06,1.0498812832248867e-06,6.354906517612905e-06,1.6926037337531308e-05,tokens/kWh,15124626.92211797,MB,1471.7952,4698.60352,0.0,4290.772992,4101.547008,s,10,23.867243652343753,2.3867243652343753,0.012184962469175376,2.3887918701171875,2.3986919921874996,2.401724658203125,2.404150791015625,"[2.367410888671875, 2.395916748046875, 2.40475732421875, 2.392780517578125, 2.396465576171875, 2.39801806640625, 2.38480322265625, 2.3697744140625, 2.381960693359375, 2.375356201171875]",tokens/s,26.396009911187807,kWh,6.981038930372296e-05,7.700269786290388e-06,4.4233967286787135e-05,0.0001217446263768005,tokens/kWh,517476.63839399815,,s,630,23.864826679229736,0.03788067726861863,0.0006657831223007035,0.037784208297729494,0.03847223854064941,0.03878150005340576,0.04094101547241213,"[0.038617599487304685, 0.03872358322143555, 0.037953056335449216, 0.037543998718261716, 0.037867935180664065, 0.03717907333374024, 0.037150432586669925, 0.037335296630859376, 0.03702819061279297, 0.03713616180419922, 0.0369870719909668, 0.03721971130371094, 0.0374477424621582, 0.03710940933227539, 0.03709014511108399, 0.03710985565185547, 0.03709952163696289, 0.03725423812866211, 0.03704892730712891, 0.03720428848266601, 0.037548030853271484, 0.04331907272338867, 0.03881929779052734, 0.03794800186157227, 0.037546142578125, 0.037746688842773435, 0.03780022430419922, 0.03706403350830078, 0.03736614227294922, 0.03753104019165039, 0.037730911254882815, 0.03717324829101563, 0.03718143844604492, 0.037617664337158206, 0.03740671920776367, 0.03711555099487305, 0.037111328125, 0.03732592010498047, 0.037770782470703125, 0.03729347229003906, 0.03736441421508789, 0.03736912155151367, 0.03717171096801758, 0.03729792022705078, 0.03718406295776367, 0.03712614440917969, 0.03739836883544922, 0.03773004913330078, 0.037638526916503906, 0.038455329895019534, 0.037384193420410154, 0.03752959823608398, 0.037615615844726565, 0.03749273681640625, 0.03767494583129883, 0.03751119995117187, 0.03742108917236328, 0.037959678649902344, 0.03755324935913086, 0.03740348815917969, 0.03733612823486328, 0.03756748962402344, 0.037478111267089845, 0.03877478408813476, 0.038752254486083985, 0.03825158309936524, 0.03790124893188476, 0.03872972869873047, 0.0377239990234375, 0.03765852737426758, 0.0377182731628418, 0.037520992279052735, 0.037626270294189454, 0.037565982818603516, 0.03745840072631836, 0.03787980651855469, 0.03753372955322266, 0.037676193237304687, 0.03769631958007812, 0.037738494873046875, 0.03762688064575195, 0.0381102066040039, 0.03781836700439453, 0.0382033920288086, 0.037920768737792966, 0.037814273834228515, 0.03786342239379883, 0.03765862274169922, 0.037577953338623044, 0.03753209686279297, 0.037644607543945316, 0.038102783203125, 0.03795142364501953, 0.03868636703491211, 0.0385338249206543, 0.03799193572998047, 0.03815030288696289, 0.03820899200439453, 0.03796057510375977, 0.038338558197021484, 0.038309856414794924, 0.03872108840942383, 0.03796543884277344, 0.037763679504394534, 0.03762128067016601, 0.0375437126159668, 0.0380568962097168, 0.03797401428222656, 0.03777536010742188, 0.03787776184082031, 0.037235904693603515, 0.03707900619506836, 0.03728265762329101, 0.0376258544921875, 0.03771187210083008, 0.03774816131591797, 0.04137631988525391, 0.04167679977416992, 0.03841843032836914, 0.03982089614868164, 0.03804367828369141, 0.037476318359375, 0.03778704071044922, 0.0378798713684082, 0.03732575988769531, 0.037689342498779296, 0.03847212982177734, 0.03798425674438476, 0.03778559875488281, 0.03756582260131836, 0.037290622711181644, 0.03724924850463867, 0.037281566619873044, 0.03740671920776367, 0.03721609497070313, 0.037505184173583984, 0.038109184265136715, 0.037601280212402347, 0.039042560577392575, 0.037776161193847656, 0.03816624069213867, 0.03829555130004883, 0.037814273834228515, 0.03790233612060547, 0.0380489616394043, 0.0377943344116211, 0.038835582733154295, 0.03781929779052735, 0.038152191162109376, 0.03845452880859375, 0.0378858871459961, 0.037880638122558596, 0.03789004898071289, 0.038950782775878906, 0.038082687377929685, 0.03777705764770508, 0.037869918823242185, 0.038012928009033206, 0.03855142211914062, 0.03824617767333984, 0.03829699325561523, 0.038497119903564456, 0.0383499526977539, 0.03813475036621094, 0.03916185760498047, 0.03792812728881836, 0.03803398513793945, 0.03862268829345703, 0.03772848129272461, 0.03784352111816406, 0.03800243377685547, 0.038145759582519534, 0.03871120071411133, 0.038163070678710935, 0.03809075164794922, 0.0383559684753418, 0.038142654418945314, 0.03818726348876953, 0.03794563293457031, 0.03791030502319336, 0.038182910919189454, 0.04315955352783203, 0.03846553421020508, 0.03833446502685547, 0.03824844741821289, 0.03798339080810547, 0.03790934371948242, 0.038156192779541014, 0.03909231948852539, 0.039147872924804684, 0.038664127349853514, 0.03829971313476563, 0.03805952072143555, 0.03842697525024414, 0.03809296035766602, 0.038006591796875, 0.03784268951416016, 0.038163105010986326, 0.037844768524169924, 0.03772415924072266, 0.03810508728027344, 0.038084606170654296, 0.03816556930541992, 0.03803823852539062, 0.03806639862060547, 0.03829759979248047, 0.0382006721496582, 0.038531425476074216, 0.038784961700439456, 0.0377470703125, 0.03788508987426758, 0.038133983612060544, 0.037431934356689456, 0.03786735916137695, 0.037932350158691404, 0.03779875183105469, 0.038152065277099606, 0.03748806381225586, 0.03740947341918945, 0.03728326416015625, 0.03738886260986328, 0.041082878112792966, 0.03774044799804688, 0.03744992065429688, 0.03774044799804688, 0.037119998931884765, 0.037713920593261716, 0.03743648147583008, 0.03794387054443359, 0.037676929473876956, 0.037638656616210936, 0.03759308624267578, 0.03818096160888672, 0.03784489440917969, 0.03870876693725586, 0.037744670867919924, 0.03777171325683594, 0.037821857452392575, 0.03805654525756836, 0.03766886520385742, 0.037908672332763675, 0.037782817840576174, 0.03795817565917969, 0.03777536010742188, 0.037881118774414066, 0.03751510238647461, 0.03768511962890625, 0.03766284942626953, 0.03817356872558594, 0.0380252799987793, 0.0383361587524414, 0.037826847076416016, 0.041443134307861326, 0.03885075378417969, 0.03840409469604492, 0.03844694519042969, 0.0385431022644043, 0.03819356918334961, 0.03812351989746094, 0.03804774475097656, 0.038340606689453126, 0.03795337677001953, 0.038086273193359374, 0.03827561569213867, 0.038223167419433594, 0.03793171310424805, 0.03806412887573242, 0.03798204803466797, 0.03826704025268555, 0.0381030387878418, 0.03815407943725586, 0.03791622543334961, 0.03775859069824219, 0.037835742950439455, 0.03760879898071289, 0.037692062377929686, 0.03873283386230469, 0.03832880020141602, 0.04035395050048828, 0.03817654418945313, 0.0377534065246582, 0.03865740966796875, 0.03832076644897461, 0.037902015686035156, 0.03796118545532227, 0.038216545104980466, 0.03884051132202149, 0.037921600341796875, 0.037763679504394534, 0.03792323303222656, 0.03783190536499023, 0.037643039703369144, 0.037549823760986326, 0.03732710266113281, 0.03789823913574219, 0.03803948974609375, 0.03767529678344726, 0.03773417663574219, 0.03813580703735352, 0.037889278411865235, 0.037743358612060546, 0.03740172958374023, 0.037454719543457034, 0.037591007232666014, 0.037291969299316406, 0.038241790771484374, 0.037403167724609374, 0.03739039993286133, 0.03775692749023438, 0.03721388626098633, 0.03747257614135742, 0.037443584442138675, 0.0375316162109375, 0.03803548812866211, 0.03744480133056641, 0.038703102111816406, 0.03810508728027344, 0.037483840942382815, 0.037300289154052736, 0.0371569938659668, 0.03723929595947266, 0.037803680419921874, 0.03711008071899414, 0.038775905609130856, 0.037886913299560544, 0.03817267227172851, 0.037919807434082034, 0.03752758407592773, 0.03761859130859375, 0.03767091369628906, 0.03775078582763672, 0.037946750640869144, 0.037646976470947266, 0.03753894424438477, 0.037775646209716796, 0.03806787109375, 0.03756060791015625, 0.03763267135620117, 0.03757670211791992, 0.03806208038330078, 0.037574657440185545, 0.0376360969543457, 0.0377077751159668, 0.03757231903076172, 0.037464351654052735, 0.03748044967651367, 0.04123817443847656, 0.03886656188964844, 0.0382696647644043, 0.04059369659423828, 0.03824099349975586, 0.038230785369873045, 0.03819692611694336, 0.03825078582763672, 0.03838390350341797, 0.038672351837158205, 0.038297630310058596, 0.03860684967041016, 0.03810844802856445, 0.03784076690673828, 0.03787235260009766, 0.03802096176147461, 0.03816016006469727, 0.038226238250732424, 0.03812473678588867, 0.0381921272277832, 0.03843686294555664, 0.03818700790405274, 0.03804361724853516, 0.03817475128173828, 0.03830579376220703, 0.037822463989257815, 0.03791782379150391, 0.03818175888061524, 0.03799766540527344, 0.03878390502929688, 0.038100990295410156, 0.037904384613037106, 0.03861811065673828, 0.03797670364379883, 0.03751766586303711, 0.03755011367797852, 0.03784294509887695, 0.0380428466796875, 0.03786771011352539, 0.037630207061767576, 0.03742345428466797, 0.037429054260253905, 0.03752569580078125, 0.03743539047241211, 0.03781340789794922, 0.0378639030456543, 0.037566848754882816, 0.037720062255859374, 0.03824639892578125, 0.037609088897705076, 0.03836467361450195, 0.03868137741088867, 0.03813167953491211, 0.03810275268554687, 0.0378737907409668, 0.03767728042602539, 0.03799660873413086, 0.03741689682006836, 0.03741856002807617, 0.037669376373291014, 0.03745177459716797, 0.03738800048828125, 0.037082977294921875, 0.037361793518066407, 0.037642017364501956, 0.03993449783325195, 0.03873382568359375, 0.037572608947753904, 0.037260478973388675, 0.03743827056884766, 0.03787324905395508, 0.0376569938659668, 0.037889057159423825, 0.03755510330200195, 0.03805356979370117, 0.03737004852294922, 0.03751955032348633, 0.037236160278320315, 0.037839424133300784, 0.039479297637939455, 0.039799041748046875, 0.038045440673828125, 0.038239551544189454, 0.03769606399536133, 0.03753792190551758, 0.03748659133911133, 0.03770163345336914, 0.03797401428222656, 0.038391807556152346, 0.03812870407104492, 0.038083518981933594, 0.037506977081298826, 0.03755401611328125, 0.03746432113647461, 0.037614784240722655, 0.03924870300292969, 0.03814723205566406, 0.038222625732421876, 0.03811334228515625, 0.037918846130371095, 0.03823756790161133, 0.03756304168701172, 0.03733196640014649, 0.037604190826416015, 0.03753334426879883, 0.03730803298950195, 0.037832672119140626, 0.03788671875, 0.037920768737792966, 0.0376297607421875, 0.03797011184692383, 0.03770272064208984, 0.03800371170043945, 0.0376888313293457, 0.03760172653198242, 0.037492385864257814, 0.03768681716918945, 0.03734611129760742, 0.038117374420166016, 0.037599231719970705, 0.03743660736083984, 0.037603614807128906, 0.03790492630004883, 0.037222400665283206, 0.037493824005126956, 0.03749369430541992, 0.037302528381347656, 0.03739798355102539, 0.03750121688842773, 0.03728384017944336, 0.03743539047241211, 0.0373043212890625, 0.037326751708984376, 0.037342529296875, 0.03733174514770508, 0.03816243362426758, 0.03712598419189453, 0.03721846389770508, 0.037197662353515626, 0.03713248062133789, 0.03705238342285156, 0.03700924682617188, 0.03694607925415039, 0.03760332870483398, 0.03952844619750977, 0.038485790252685545, 0.037216190338134766, 0.03725446319580078, 0.03763504028320312, 0.038108959197998046, 0.03759535980224609, 0.0373043212890625, 0.037365665435791014, 0.037642433166503904, 0.037257118225097655, 0.03714451217651367, 0.03718560028076172, 0.037294078826904296, 0.03863663864135742, 0.037733280181884765, 0.03736598587036133, 0.03816147232055664, 0.037501152038574216, 0.03825305557250976, 0.03767055892944336, 0.037455806732177736, 0.03760547256469727, 0.03816851043701172, 0.037613121032714844, 0.03762252807617188, 0.037803359985351566, 0.03807100677490234, 0.03868374252319336, 0.03753871917724609, 0.037634048461914066, 0.037715713500976564, 0.03779759979248047, 0.03749327850341797, 0.03747020721435547, 0.03794659042358398, 0.03826182556152344, 0.03788684844970703, 0.037776222229003904, 0.03784089660644531, 0.03779286575317383, 0.0376135368347168, 0.03750105667114258, 0.038009662628173825, 0.038217727661132815, 0.037838848114013675, 0.03881369781494141, 0.03828940963745117, 0.038168575286865236, 0.03909542465209961, 0.03793913650512695, 0.03784783935546875, 0.037752639770507815, 0.037763423919677734, 0.037986305236816405, 0.037621440887451174, 0.037833023071289065, 0.03790796661376953, 0.0378064956665039, 0.037529502868652344, 0.03749292755126953, 0.037625598907470706, 0.037867774963378904, 0.03746390533447266, 0.03738159942626953, 0.03745248031616211, 0.03741247940063477, 0.03793139266967773, 0.037332672119140625, 0.03730422210693359, 0.03790233612060547, 0.037267871856689457, 0.03803750228881836, 0.037482494354248046, 0.03720191955566406, 0.03759718322753906, 0.03794739151000977, 0.038240734100341796, 0.03747856140136719, 0.037015552520751956, 0.037326847076416016, 0.03701964950561523, 0.037101566314697264, 0.03706633758544922, 0.036865825653076174, 0.03696902465820313, 0.037007423400878904, 0.03992153549194336, 0.03817279815673828, 0.03718739318847656, 0.039765918731689456, 0.03724921417236328, 0.037152862548828124, 0.037263233184814455, 0.037769153594970704, 0.03732704162597656, 0.037426624298095706, 0.03738886260986328, 0.03772012710571289, 0.03724889755249024, 0.03720198440551758, 0.037117023468017575, 0.03705654525756836, 0.03707571029663086, 0.03721539306640625, 0.03715580749511719, 0.03727360153198242, 0.03806604766845703, 0.03782463836669922, 0.037758945465087894, 0.038458881378173826, 0.03807696151733399, 0.0378768310546875, 0.03747318267822266, 0.03795558547973633, 0.03839385604858398, 0.038473217010498044, 0.03883817672729492, 0.037865184783935545, 0.038353790283203126, 0.03849385452270508, 0.03846793746948242, 0.03832012939453125, 0.03799859237670898, 0.03807436752319336, 0.03877856063842773, 0.03746416091918945, 0.03774998474121094, 0.037790271759033205, 0.03757817459106445, 0.037577728271484374, 0.03746326446533203, 0.037720577239990234, 0.037671329498291016, 0.03754940795898438, 0.037439998626708985, 0.03759516906738281, 0.03735551834106445, 0.03742105484008789, 0.037384193420410154]",tokens/s,26.39868323654358,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4917.706752,7227.768832,0.0,6849.298432,6444.4416,s,1,11.1887177734375,11.1887177734375,0.0,11.1887177734375,11.1887177734375,11.1887177734375,11.1887177734375,[11.1887177734375],,kWh,0.00012569621067492942,1.385766822740644e-05,4.088281048403308e-05,0.00018043668938636895,,MB,2799.86176,7559.118848,0.0,7151.28832,6822.664192,s,10,4.020862365722656,0.4020862365722656,0.003901422977759983,0.4032820281982422,0.4047774627685547,0.40488236846923825,0.40496629302978515,"[0.39109088134765624, 0.4041576232910156, 0.4045347595214844, 0.40344390869140623, 0.404754150390625, 0.40246356201171873, 0.40498727416992186, 0.4005614929199219, 0.40312014770507815, 0.4017485656738281]",tokens/s,636.679340686634,kWh,1.1629275077404182e-05,1.2824702195391937e-06,7.747196368695065e-06,2.065894166563844e-05,tokens/kWh,12391728.68307185,MB,2799.86176,7561.216,0.0,7153.385472,6822.666752,s,10,26.55316528320312,2.655316528320312,0.022885641151044064,2.650044189453125,2.68346796875,2.6880259765625,2.6916723828125,"[2.6769189453125, 2.654231201171875, 2.692583984375, 2.672745361328125, 2.631368896484375, 2.645857177734375, 2.629407470703125, 2.627463623046875, 2.640133544921875, 2.682455078125]",tokens/s,23.725984954363334,kWh,7.841556979218403e-05,8.649395373571439e-06,5.191951803130231e-05,0.00013898448319705776,tokens/kWh,453288.01137229166,,s,630,26.55058037567139,0.042143778374081564,0.0007510202816334307,0.04209764862060547,0.042799008941650396,0.04308290672302246,0.04451943481445315,"[0.043050369262695315, 0.042723968505859376, 0.04230339050292969, 0.042414176940917966, 0.043726081848144534, 0.04287907028198242, 0.042371681213378906, 0.042048927307128905, 0.042234046936035156, 0.04298908615112305, 0.04257068634033203, 0.04478297424316406, 0.04221948623657226, 0.04212393569946289, 0.04221116638183594, 0.042196640014648436, 0.04222224044799805, 0.04278403091430664, 0.04237097549438477, 0.042602977752685546, 0.04230297470092773, 0.04273017501831055, 0.04251798248291016, 0.042531360626220704, 0.042122337341308595, 0.04204022216796875, 0.04191436767578125, 0.042202911376953124, 0.04231516647338867, 0.04235756683349609, 0.042188800811767575, 0.04215555191040039, 0.042215648651123046, 0.04220691299438477, 0.042168033599853515, 0.042410144805908205, 0.04224892807006836, 0.04223344039916992, 0.04226271820068359, 0.042250431060791016, 0.04224844741821289, 0.04248899078369141, 0.04231388854980469, 0.04243289566040039, 0.04228921508789062, 0.042982654571533205, 0.04274975967407227, 0.04255020904541015, 0.04237334442138672, 0.04236265563964844, 0.04246028900146484, 0.04233919906616211, 0.04254540634155273, 0.04230857467651367, 0.04268435287475586, 0.04262588882446289, 0.0425615348815918, 0.04299139022827148, 0.042422367095947267, 0.04227699279785156, 0.04309401702880859, 0.04243046569824219, 0.04290707015991211, 0.043122432708740235, 0.04241129684448242, 0.0425618896484375, 0.04256422424316406, 0.04281935882568359, 0.04258793640136719, 0.04271558380126953, 0.04284128189086914, 0.04260265731811524, 0.04257244873046875, 0.042379264831542966, 0.04228300857543945, 0.042031105041503904, 0.04188150405883789, 0.04207974243164062, 0.04235939025878906, 0.042176513671875, 0.042008575439453126, 0.04190617752075195, 0.04215155029296875, 0.0420665283203125, 0.04206979370117187, 0.04171712112426758, 0.04163238525390625, 0.04157030487060547, 0.04158464050292969, 0.04142899322509765, 0.04173366546630859, 0.04183087921142578, 0.0420843505859375, 0.04209267044067383, 0.04279897689819336, 0.04198323059082031, 0.04194153594970703, 0.04198329544067383, 0.041847713470458986, 0.04185878372192383, 0.0421255989074707, 0.04196761703491211, 0.042017951965332034, 0.041939807891845704, 0.04199532699584961, 0.041931137084960934, 0.042439231872558596, 0.04205567932128906, 0.04214492797851563, 0.04246409606933594, 0.0419488639831543, 0.042539329528808595, 0.042090496063232424, 0.04209183883666992, 0.04198060989379883, 0.04171305465698242, 0.042199649810791016, 0.042076160430908206, 0.042016864776611325, 0.04171084976196289, 0.04182819366455078, 0.04194319915771484, 0.04198467254638672, 0.042016769409179686, 0.04209616088867187, 0.04237673568725586, 0.043112449645996094, 0.04229097747802734, 0.042266849517822266, 0.04307558441162109, 0.04255744171142578, 0.04215193557739258, 0.04266995239257813, 0.04195500946044922, 0.042191295623779296, 0.04217657470703125, 0.04233824157714844, 0.04226233673095703, 0.0424071044921875, 0.04232908630371094, 0.04239680099487304, 0.042457984924316405, 0.042534912109375, 0.042688190460205076, 0.043727169036865236, 0.042883071899414066, 0.04316515350341797, 0.04265407943725586, 0.04280335998535156, 0.04683366394042969, 0.04271820831298828, 0.04327731323242188, 0.04293759918212891, 0.04272800064086914, 0.04279929733276367, 0.04272556686401367, 0.04281734466552734, 0.04264934539794922, 0.04271539306640625, 0.04273356628417969, 0.04729967880249023, 0.04230646514892578, 0.04247539138793945, 0.04347711944580078, 0.04261273574829102, 0.04236697769165039, 0.042323455810546876, 0.04280166244506836, 0.042396705627441404, 0.04268569564819336, 0.04256739044189453, 0.04236675262451172, 0.04247328186035156, 0.042234272003173826, 0.042272735595703125, 0.042342430114746095, 0.04218991851806641, 0.042171104431152344, 0.042248382568359374, 0.043146656036376956, 0.04248636627197266, 0.042452991485595705, 0.04235468673706055, 0.042555328369140624, 0.042743873596191403, 0.04249599838256836, 0.04244275283813476, 0.04252671813964844, 0.04344627380371094, 0.04331520080566406, 0.04274319839477539, 0.04243251037597656, 0.042345054626464845, 0.04277248001098633, 0.04229321670532227, 0.04238700866699219, 0.04232758331298828, 0.04252377700805664, 0.04239545440673828, 0.042446849822998046, 0.042323680877685545, 0.04352643203735351, 0.04284415817260742, 0.04274892807006836, 0.0422691535949707, 0.042570270538330075, 0.0426618881225586, 0.04286054229736328, 0.042395038604736326, 0.04224470520019531, 0.04248166275024414, 0.0430458869934082, 0.04288963317871094, 0.04252531051635742, 0.042584030151367185, 0.04259430313110352, 0.04363673782348633, 0.042551296234130856, 0.04229232025146484, 0.042142623901367186, 0.042255775451660156, 0.04233071899414063, 0.042167713165283206, 0.0423983039855957, 0.04231577682495117, 0.04255923080444336, 0.04224230575561523, 0.042426368713378904, 0.04289945602416992, 0.042590206146240234, 0.04287510299682617, 0.04308889770507812, 0.04238415908813477, 0.04267766571044922, 0.04229180908203125, 0.04269055938720703, 0.04225772857666016, 0.04254175949096679, 0.042409984588623044, 0.04238729476928711, 0.04258627319335938, 0.04204553604125977, 0.041775009155273435, 0.0416313591003418, 0.04174256134033203, 0.041781406402587894, 0.041734142303466795, 0.041836544036865236, 0.04170867156982422, 0.04156502532958985, 0.04154985427856445, 0.04154982376098633, 0.04301548767089844, 0.04262368011474609, 0.042694496154785155, 0.042135711669921874, 0.042176513671875, 0.042164222717285156, 0.04244198226928711, 0.04195817565917969, 0.04172796630859375, 0.041543678283691404, 0.04157574462890625, 0.041379745483398435, 0.04132479858398438, 0.04139471817016602, 0.041940990447998046, 0.04202489471435547, 0.04152873611450195, 0.042224288940429684, 0.04172185516357422, 0.04159801483154297, 0.041406944274902345, 0.04151350402832031, 0.04144918441772461, 0.04143894577026367, 0.04147577667236328, 0.041364288330078124, 0.04150614547729492, 0.04143894577026367, 0.04143404769897461, 0.04128768157958984, 0.041420799255371094, 0.041457279205322266, 0.041255294799804686, 0.04150579071044922, 0.04203766250610352, 0.04175455856323242, 0.04232233428955078, 0.0415830078125, 0.042880767822265624, 0.041592479705810544, 0.04178579330444336, 0.04185075378417969, 0.04178137588500976, 0.04175807952880859, 0.041553825378417966, 0.041799617767333985, 0.04163836669921875, 0.04159718322753906, 0.04169276809692383, 0.04152099227905273, 0.042199710845947265, 0.04173206329345703, 0.04155136108398438, 0.04149654388427734, 0.04235318374633789, 0.04148601531982422, 0.0413570556640625, 0.04136783981323242, 0.041614719390869144, 0.041599903106689456, 0.041766368865966796, 0.04191267013549805, 0.042381057739257814, 0.04323331069946289, 0.04211507034301758, 0.04473187255859375, 0.04230319976806641, 0.04167353439331055, 0.04179148864746094, 0.04196556854248047, 0.04187136077880859, 0.042141792297363284, 0.042209121704101564, 0.042469375610351565, 0.04298144149780273, 0.042426368713378904, 0.04247283172607422, 0.04238809585571289, 0.04270284652709961, 0.042602497100830077, 0.04253081512451172, 0.04262070465087891, 0.042135326385498044, 0.042272289276123046, 0.042019519805908206, 0.04195673751831055, 0.04172867202758789, 0.04188179016113281, 0.04255097579956055, 0.04166892623901367, 0.04165222549438476, 0.042495391845703126, 0.04205424118041992, 0.04177305603027344, 0.04156415939331055, 0.04155324935913086, 0.0418939208984375, 0.041517696380615234, 0.041315391540527345, 0.04148857498168945, 0.04197030258178711, 0.04140390396118164, 0.041644161224365234, 0.041810176849365235, 0.041982433319091794, 0.0451929931640625, 0.04190185546875, 0.0417624626159668, 0.04150944137573242, 0.041498207092285154, 0.04141507339477539, 0.04143718338012695, 0.041506431579589845, 0.041320926666259764, 0.041235488891601564, 0.04125350570678711, 0.04121392059326172, 0.043397407531738284, 0.041603073120117184, 0.04145356750488281, 0.04136934280395508, 0.041266910552978514, 0.04120835113525391, 0.04130944061279297, 0.041593601226806644, 0.04162518310546875, 0.043474720001220706, 0.042169567108154296, 0.04163631820678711, 0.04159337615966797, 0.041350177764892575, 0.04148083114624023, 0.04142489624023438, 0.041363807678222654, 0.04153753662109375, 0.043300254821777344, 0.04158723068237305, 0.041689151763916014, 0.04187136077880859, 0.041844608306884766, 0.04157392120361328, 0.04181667327880859, 0.041529342651367186, 0.041578495025634765, 0.04151705551147461, 0.04137955093383789, 0.04145148849487305, 0.041877120971679685, 0.041554622650146485, 0.04151910400390625, 0.041532672882080075, 0.04144355010986328, 0.041202014923095706, 0.041445022583007814, 0.0413721923828125, 0.041332736968994144, 0.04141449737548828, 0.04128924942016601, 0.04290828704833984, 0.04304921722412109, 0.041860992431640626, 0.04144329452514649, 0.04167782211303711, 0.04138713455200195, 0.0413917121887207, 0.041256736755371094, 0.041412479400634764, 0.041476638793945315, 0.04153139114379883, 0.043845630645751955, 0.04167084884643555, 0.04147897720336914, 0.04163273620605469, 0.04160428619384766, 0.04209561538696289, 0.04138275146484375, 0.04224012756347656, 0.04136025619506836, 0.0413306884765625, 0.04162736129760742, 0.04194079971313477, 0.041578975677490235, 0.04140630340576172, 0.04205923080444336, 0.041492191314697266, 0.041439361572265625, 0.04139299011230469, 0.043151168823242186, 0.04190399932861328, 0.04269571304321289, 0.04172079849243164, 0.041441280364990236, 0.04161280059814453, 0.04169779205322265, 0.04142489624023438, 0.04139436721801758, 0.04137100982666016, 0.04162534332275391, 0.04133763122558594, 0.04151395034790039, 0.04177606582641601, 0.04171980667114258, 0.04170137786865234, 0.04142489624023438, 0.0414343376159668, 0.041517345428466794, 0.04150505447387695, 0.04149174499511719, 0.041393089294433597, 0.041406688690185545, 0.04138166427612305, 0.04145529556274414, 0.04275436782836914, 0.04150406265258789, 0.04141030502319336, 0.04116144180297852, 0.04178761672973633, 0.04143513488769531, 0.041267200469970705, 0.04135273742675781, 0.04127324676513672, 0.04118710327148437, 0.04124716949462891, 0.041252670288085935, 0.04137337493896484, 0.041220638275146486, 0.041243038177490234, 0.04116265487670898, 0.04134912109375, 0.04136140823364258, 0.041383167266845704, 0.041341022491455076, 0.04177945709228516, 0.04209913635253906, 0.042399105072021485, 0.042089054107666016, 0.042247840881347656, 0.042213119506835935, 0.04214435195922851, 0.041944225311279296, 0.04212966537475586, 0.042162784576416014, 0.04201267242431641, 0.041987998962402344, 0.04198102569580078, 0.04316057586669922, 0.04195270538330078, 0.041963680267333985, 0.04219865417480469, 0.04232886505126953, 0.04232191848754883, 0.041973758697509765, 0.04295068740844726, 0.041785953521728515, 0.04151897430419922, 0.04136153411865234, 0.04135116958618164, 0.04126924896240235, 0.04138598251342773, 0.041348991394042967, 0.04158272171020508, 0.04145971298217774, 0.041809921264648435, 0.04164966583251953, 0.04158310317993164, 0.04150886535644531, 0.041793537139892575, 0.04179574584960938, 0.04170940780639648, 0.04191641616821289, 0.04182204818725586, 0.04187561416625977, 0.04194713592529297, 0.041569599151611326, 0.04247372817993164, 0.04175302505493164, 0.04176105499267578, 0.041620193481445314, 0.0416286735534668, 0.04220848083496094, 0.042175521850585936, 0.042108097076416016, 0.042175041198730466, 0.04249599838256836, 0.04202060699462891, 0.0424901123046875, 0.042512577056884764, 0.042028255462646484, 0.04193679809570312, 0.04195603179931641, 0.04208025741577148, 0.042063648223876954, 0.041883777618408204, 0.04206396865844726, 0.04194246292114258, 0.04239023971557617, 0.04175881576538086, 0.041809566497802736, 0.04310230255126953, 0.04223590469360351, 0.04257177734375, 0.04195942306518555, 0.04192371368408203, 0.04188051223754883, 0.04179142379760742, 0.04176428985595703, 0.04182278442382813, 0.042210399627685545, 0.041821086883544925, 0.04180377578735352, 0.041867263793945314, 0.04162355041503906, 0.04171494293212891, 0.04165219116210937, 0.04181628799438476, 0.042844352722167967, 0.04187590408325195, 0.04186928176879883, 0.042772510528564456, 0.04235449600219727, 0.04203683090209961, 0.04184124755859375, 0.042022911071777344, 0.041992416381835936, 0.04208351898193359, 0.0420533447265625, 0.04233919906616211, 0.04211312103271484, 0.04234640121459961, 0.042450111389160154, 0.042805248260498044, 0.042377662658691403, 0.042434944152832034, 0.042485759735107424, 0.042358016967773436, 0.04210492706298828, 0.04296540832519531, 0.04227827072143555, 0.04251091384887695, 0.04224646377563476, 0.042145790100097655, 0.042063072204589845, 0.041952030181884765, 0.04200243377685547, 0.04235059356689453, 0.0461578254699707, 0.05191017532348633, 0.04245078277587891, 0.04257036972045898, 0.0424136962890625, 0.04256192016601563, 0.04248950576782227, 0.04231439971923828, 0.04212047958374023, 0.04206131362915039, 0.04213238525390625, 0.04233420944213867, 0.0422542724609375, 0.042117183685302734, 0.04207001495361328, 0.04223696136474609, 0.04243555068969727, 0.042176097869873044, 0.04202950286865234, 0.04280112075805664, 0.04214076614379883, 0.042224544525146485, 0.04330495834350586, 0.04399932861328125, 0.0425634880065918, 0.042362846374511716, 0.042598430633544924, 0.04242544174194336, 0.04326902389526367, 0.04240153503417969, 0.04241843032836914, 0.042504417419433595, 0.04223680114746094]",tokens/s,23.728294865346015,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,889.417728,638.517248,0.0,260.046848,253.520896,s,1,7.95947119140625,7.95947119140625,0.0,7.95947119140625,7.95947119140625,7.95947119140625,7.95947119140625,[7.95947119140625],,kWh,2.1573645654151127e-05,2.372538863577728e-06,7.558894936005034e-06,3.150507945373389e-05,,MB,1197.993984,749.666304,0.0,341.835776,312.39168,s,13,0.16876528072357178,0.012981944671043983,0.0001479043615388075,0.012968255996704101,0.013094316673278808,0.013222700881958007,0.0133575772857666,"[0.01339129638671875, 0.012816160202026368, 0.01293564796447754, 0.012968255996704101, 0.01281935977935791, 0.012997568130493164, 0.013030367851257324, 0.013029151916503907, 0.01294924831390381, 0.013014816284179687, 0.01311030387878418, 0.012891039848327637, 0.012812064170837403]",tokens/s,19719.69581498863,kWh,3.806948366362395e-07,4.198411301527668e-08,2.527649421799144e-07,6.754438918314305e-07,tokens/kWh,379010015.6296765,MB,1209.102336,776.92928,0.0,369.098752,313.0496,s,13,9.86955401611328,0.7591964627779446,0.008135141971594635,0.7592740478515625,0.7656478393554688,0.7727370849609375,0.7804448291015625,"[0.7600068359375, 0.7629833374023437, 0.7575670166015624, 0.7509777221679688, 0.7530004272460937, 0.76631396484375, 0.7823717651367188, 0.7542471313476562, 0.76058740234375, 0.7592740478515625, 0.751735595703125, 0.7594671630859375, 0.7510216064453125]",tokens/s,82.98247303402769,kWh,2.2066482475545642e-05,2.433513627452451e-06,8.384656497666155e-06,3.288465260066424e-05,tokens/kWh,1915787.305556862,,s,819,9.863349944114677,0.012043162324926355,0.00031496773920116716,0.011961824417114257,0.012284908485412598,0.012567424011230469,0.012816493759155272,"[0.011796287536621094, 0.011968704223632812, 0.011915200233459473, 0.011891776084899903, 0.011961440086364745, 0.011865344047546387, 0.011878496170043945, 0.011895359992980957, 0.011929599761962891, 0.011890687942504884, 0.011993087768554688, 0.012023807525634766, 0.01209059238433838, 0.012116864204406738, 0.012681119918823243, 0.01223475170135498, 0.012143936157226563, 0.012116543769836426, 0.012095616340637208, 0.012186816215515137, 0.012251551628112794, 0.012153216361999512, 0.01212009620666504, 0.012156928062438965, 0.012240927696228027, 0.012166943550109863, 0.012101823806762696, 0.01239641571044922, 0.012196991920471191, 0.01226854419708252, 0.012060064315795899, 0.01197935962677002, 0.012107775688171387, 0.012236127853393555, 0.012231231689453125, 0.012126303672790528, 0.011988991737365723, 0.01231065559387207, 0.011973919868469239, 0.011962976455688477, 0.0119716796875, 0.012004256248474121, 0.012066720008850097, 0.012164671897888184, 0.012074943542480468, 0.012098143577575684, 0.012099424362182616, 0.01206287956237793, 0.012087295532226563, 0.01205247974395752, 0.012009471893310546, 0.011985055923461914, 0.011939552307128906, 0.01194816017150879, 0.011915072441101075, 0.011905216217041015, 0.011914624214172363, 0.011859968185424804, 0.011929439544677735, 0.011903776168823241, 0.011952383995056153, 0.011964159965515137, 0.011898880004882812, 0.011673600196838378, 0.01203536033630371, 0.01191539192199707, 0.011936351776123047, 0.01189417552947998, 0.012104415893554687, 0.012018719673156738, 0.012157343864440917, 0.01210531234741211, 0.012163295745849609, 0.012044992446899414, 0.011990976333618165, 0.011879839897155763, 0.011856063842773438, 0.012040224075317383, 0.011984831809997558, 0.011981087684631348, 0.011946144104003907, 0.01189840030670166, 0.011926239967346191, 0.011904000282287597, 0.011936351776123047, 0.0118756160736084, 0.012030816078186036, 0.012051615715026855, 0.012051648139953613, 0.012008511543273925, 0.01205731201171875, 0.012085311889648438, 0.01203321647644043, 0.012001888275146484, 0.011900064468383788, 0.011935872077941894, 0.012069888114929199, 0.012303680419921874, 0.012181856155395509, 0.012059840202331543, 0.012696640014648437, 0.01202188777923584, 0.01209875202178955, 0.01203667163848877, 0.012064831733703613, 0.012107935905456543, 0.012174304008483888, 0.012284735679626464, 0.012240575790405274, 0.012396960258483887, 0.012373760223388671, 0.012519392013549804, 0.013121888160705567, 0.012318559646606445, 0.012291744232177734, 0.012163423538208007, 0.01221241569519043, 0.012189503669738769, 0.012237983703613281, 0.01216598415374756, 0.012154879570007325, 0.012143967628479004, 0.012050496101379395, 0.012118623733520508, 0.012170528411865235, 0.01211023998260498, 0.011717184066772461, 0.012006848335266114, 0.011885343551635742, 0.01196828842163086, 0.011910783767700195, 0.011936127662658691, 0.012124159812927245, 0.012184927940368653, 0.014021087646484375, 0.01225334358215332, 0.01212758445739746, 0.012059040069580078, 0.012007712364196777, 0.011972736358642579, 0.012008607864379883, 0.011986687660217285, 0.01210262393951416, 0.011972607612609864, 0.011913215637207031, 0.011999135971069335, 0.011972703933715821, 0.011902976036071777, 0.011894783973693847, 0.01200492763519287, 0.012003840446472168, 0.012072575569152833, 0.012294560432434083, 0.012018752098083497, 0.012030400276184083, 0.011909536361694336, 0.011890399932861328, 0.011886879920959473, 0.0118985595703125, 0.012220895767211915, 0.012071904182434081, 0.012048768043518067, 0.012084768295288087, 0.011950112342834473, 0.01187116813659668, 0.011939423561096191, 0.011938207626342774, 0.011877568244934082, 0.011938624382019043, 0.01189788818359375, 0.011901439666748047, 0.011956319808959961, 0.011951647758483886, 0.011873023986816406, 0.011915360450744629, 0.011867615699768066, 0.011923999786376953, 0.011913215637207031, 0.012261440277099609, 0.01193065643310547, 0.011951007843017579, 0.012047871589660645, 0.011944543838500977, 0.011912991523742675, 0.011993215560913087, 0.011926912307739257, 0.012007871627807616, 0.011999423980712891, 0.012043840408325195, 0.011690079689025879, 0.012065664291381836, 0.01202995204925537, 0.012007488250732421, 0.011931584358215333, 0.012000639915466309, 0.011931808471679687, 0.011929439544677735, 0.012057439804077148, 0.011970335960388184, 0.012031999588012696, 0.012247039794921874, 0.011903103828430175, 0.011935456275939942, 0.011903136253356934, 0.011843584060668945, 0.0120481595993042, 0.011945343971252442, 0.011853952407836914, 0.011864800453186036, 0.011882495880126954, 0.011941984176635742, 0.011857695579528808, 0.011934176445007324, 0.011855520248413087, 0.011881888389587402, 0.011866047859191894, 0.011974656105041503, 0.011963104248046875, 0.011917471885681152, 0.011865887641906738, 0.011892352104187012, 0.011868800163269043, 0.011867487907409668, 0.011838080406188965, 0.011863295555114745, 0.011938528060913086, 0.011845151901245117, 0.011835519790649413, 0.011936127662658691, 0.011857695579528808, 0.0118471040725708, 0.011836992263793946, 0.011950528144836426, 0.011917887687683105, 0.012009471893310546, 0.011849727630615235, 0.011982848167419433, 0.01186406421661377, 0.011933504104614258, 0.011888895988464356, 0.011974623680114746, 0.011925472259521484, 0.011928768157958984, 0.01182812786102295, 0.011878303527832031, 0.011860223770141601, 0.011855615615844726, 0.01188003158569336, 0.011858336448669434, 0.011831295967102052, 0.011861503601074219, 0.011872223854064941, 0.011595552444458007, 0.011878911972045898, 0.011882559776306152, 0.011912256240844727, 0.011907327651977538, 0.011921279907226563, 0.011884448051452636, 0.01183836841583252, 0.012226559638977052, 0.01205452823638916, 0.011907072067260742, 0.011988415718078613, 0.011868032455444336, 0.011828927993774415, 0.01193391990661621, 0.011887424468994141, 0.011937760353088378, 0.011988991737365723, 0.011914976119995117, 0.011918911933898926, 0.012101375579833984, 0.011877344131469727, 0.012055744171142578, 0.01193231964111328, 0.011871583938598633, 0.011891167640686035, 0.011878591537475586, 0.01187446403503418, 0.011993087768554688, 0.01192572784423828, 0.01205020809173584, 0.012428447723388672, 0.011961312294006347, 0.01199295997619629, 0.011984479904174805, 0.011856287956237792, 0.011925439834594726, 0.01201683235168457, 0.011930496215820312, 0.01204428768157959, 0.011921407699584961, 0.011878399848937989, 0.011929599761962891, 0.011974656105041503, 0.011941184043884277, 0.011958080291748047, 0.01191164779663086, 0.012096991539001465, 0.011932607650756835, 0.011857024192810059, 0.01186905574798584, 0.011876352310180664, 0.011873791694641114, 0.011930272102355957, 0.011832448005676269, 0.011913344383239747, 0.011966624259948731, 0.011904800415039062, 0.011954208374023438, 0.012016256332397461, 0.011935232162475586, 0.01195263957977295, 0.012209792137145997, 0.012643424034118653, 0.012405664443969726, 0.012048383712768555, 0.01202995204925537, 0.011958271980285644, 0.011884544372558594, 0.012037887573242187, 0.011903231620788574, 0.012236063957214355, 0.015813568115234374, 0.015257375717163086, 0.012349184036254884, 0.011911423683166503, 0.01226137638092041, 0.011943936347961426, 0.011907072067260742, 0.012075136184692383, 0.01188758373260498, 0.011994048118591309, 0.011930944442749024, 0.01191817569732666, 0.01186796760559082, 0.011938048362731933, 0.01195315170288086, 0.011954208374023438, 0.01206550407409668, 0.0120664644241333, 0.012370400428771973, 0.012207488059997558, 0.012003840446472168, 0.012050496101379395, 0.012001215934753418, 0.012102848052978515, 0.012135231971740722, 0.012031744003295899, 0.012077312469482421, 0.012263263702392578, 0.012036352157592774, 0.012062560081481934, 0.012180831909179688, 0.012092127799987792, 0.012392543792724609, 0.012042143821716308, 0.012011520385742188, 0.011927424430847168, 0.011945695877075195, 0.01189622402191162, 0.011872960090637206, 0.011923775672912598, 0.012156031608581542, 0.011987071990966797, 0.011911231994628907, 0.011958975791931153, 0.01191868782043457, 0.011885055541992188, 0.012044447898864746, 0.012005599975585937, 0.011896608352661132, 0.012722175598144531, 0.011873888015747071, 0.01189897632598877, 0.011823424339294434, 0.011895903587341309, 0.011773951530456543, 0.01196675205230713, 0.012023520469665528, 0.011955679893493653, 0.011954624176025391, 0.01195734405517578, 0.011969280242919923, 0.01195638370513916, 0.011962688446044922, 0.011916095733642579, 0.011948960304260254, 0.01206505584716797, 0.012142111778259277, 0.012351743698120117, 0.012275712013244629, 0.012285599708557129, 0.012326975822448731, 0.012243231773376466, 0.012250783920288087, 0.012288319587707519, 0.012472352027893067, 0.012429568290710449, 0.01258675193786621, 0.012627200126647949, 0.012683039665222167, 0.01261683177947998, 0.012819552421569824, 0.012567296028137206, 0.012645888328552245, 0.012640768051147461, 0.012564640045166015, 0.012610527992248535, 0.01258790397644043, 0.012578816413879394, 0.012632063865661621, 0.012557600021362305, 0.012640992164611816, 0.012568575859069824, 0.012626976013183593, 0.01259823989868164, 0.012595104217529298, 0.012771424293518066, 0.012698816299438477, 0.012534496307373046, 0.012490303993225097, 0.012632543563842774, 0.012637984275817872, 0.012572959899902343, 0.012560319900512696, 0.012621888160705566, 0.012640128135681153, 0.012676959991455078, 0.01262816047668457, 0.012609631538391113, 0.0125665283203125, 0.012475520133972169, 0.012473088264465332, 0.01231443214416504, 0.01257091236114502, 0.012365119934082031, 0.01224163246154785, 0.012280832290649414, 0.0122391357421875, 0.01170188808441162, 0.01203651237487793, 0.012059231758117676, 0.012019136428833007, 0.011962080001831054, 0.011969344139099121, 0.011931488037109375, 0.012035296440124512, 0.011969311714172363, 0.011905023574829102, 0.011999232292175293, 0.011906271934509277, 0.0120164155960083, 0.012120287895202636, 0.012250240325927735, 0.012197919845581055, 0.012138751983642579, 0.012009696006774902, 0.01195638370513916, 0.011898880004882812, 0.011984095573425294, 0.011922207832336426, 0.011831232070922851, 0.011870271682739257, 0.011886752128601075, 0.011933376312255859, 0.011950240135192872, 0.011943936347961426, 0.011961824417114257, 0.011911711692810058, 0.012017279624938966, 0.011888447761535645, 0.011899135589599609, 0.011864383697509766, 0.012031807899475097, 0.012130496025085449, 0.012304384231567383, 0.012040032386779784, 0.012049792289733886, 0.011929599761962891, 0.011918111801147461, 0.011866080284118652, 0.011887840270996094, 0.011983776092529297, 0.012021439552307129, 0.011908991813659668, 0.011862400054931641, 0.011869600296020508, 0.011872832298278808, 0.011794431686401367, 0.01222383975982666, 0.012304544448852539, 0.011903488159179687, 0.011964415550231934, 0.011926591873168945, 0.011877311706542968, 0.011905152320861817, 0.011880319595336914, 0.011838815689086913, 0.01189136028289795, 0.011876352310180664, 0.011868160247802734, 0.011898880004882812, 0.011673312187194824, 0.011946751594543456, 0.011980223655700683, 0.011911808013916015, 0.011909119606018067, 0.011898880004882812, 0.011835519790649413, 0.011906175613403321, 0.01189740753173828, 0.01193183994293213, 0.011907135963439942, 0.012041952133178711, 0.01194825553894043, 0.011958271980285644, 0.01198095989227295, 0.011949888229370117, 0.011972352027893067, 0.011893024444580078, 0.011951199531555176, 0.011942496299743652, 0.011923744201660156, 0.01196835231781006, 0.011909503936767579, 0.011907072067260742, 0.011881279945373534, 0.011901951789855958, 0.01190112018585205, 0.011925312042236328, 0.011890015602111816, 0.011889311790466308, 0.01201257610321045, 0.011926495552062988, 0.012007295608520508, 0.011995264053344726, 0.011913215637207031, 0.011890687942504884, 0.011890687942504884, 0.011933024406433106, 0.012010144233703614, 0.011953760147094726, 0.011973024368286133, 0.01201142406463623, 0.011999327659606934, 0.012079008102416992, 0.012086848258972167, 0.012038687705993652, 0.012044159889221192, 0.01220348834991455, 0.012141056060791015, 0.012124544143676757, 0.01215056037902832, 0.012082464218139648, 0.012057120323181153, 0.012175552368164063, 0.01218563175201416, 0.012179295539855957, 0.01218511962890625, 0.012284416198730469, 0.012294015884399413, 0.012101280212402343, 0.012441856384277344, 0.01573100757598877, 0.01247436809539795, 0.011737407684326172, 0.011955615997314453, 0.011999551773071288, 0.011993663787841797, 0.01204428768157959, 0.012085247993469238, 0.012088895797729492, 0.012001440048217774, 0.011989279747009278, 0.011958271980285644, 0.011978752136230468, 0.01194604778289795, 0.011929535865783691, 0.011936863899230958, 0.012002207756042481, 0.01200879955291748, 0.01208131217956543, 0.012352000236511231, 0.012308480262756348, 0.011918432235717773, 0.011904959678649902, 0.01196720027923584, 0.011929951667785644, 0.011988896369934082, 0.01189254379272461, 0.011997376441955566, 0.011952128410339356, 0.011976192474365235, 0.011975168228149414, 0.01202790355682373, 0.012148063659667968, 0.012120736122131348, 0.012296192169189453, 0.012259327888488769, 0.012199808120727539, 0.012089056015014648, 0.012001376152038574, 0.012028223991394043, 0.012066816329956055, 0.011963839530944825, 0.012165696144104003, 0.012064448356628418, 0.012214591979980469, 0.012103679656982422, 0.012042400360107421, 0.012054368019104004, 0.01203711986541748, 0.012601856231689454, 0.012132672309875489, 0.012161215782165528, 0.012103615760803222, 0.012074080467224121, 0.012055520057678223, 0.012033663749694824, 0.012139967918395996, 0.012018624305725098, 0.012072959899902343, 0.01201801586151123, 0.011916288375854492, 0.01185244846343994, 0.011902015686035156, 0.011854111671447754, 0.012061344146728516, 0.011669504165649413, 0.012023072242736816, 0.012059359550476075, 0.012035327911376953, 0.012178175926208496, 0.0120381441116333, 0.01194934368133545, 0.011917887687683105, 0.011980159759521485, 0.0120449275970459, 0.01226972770690918, 0.011993280410766602, 0.011974047660827637, 0.012023391723632813, 0.011928383827209472, 0.011915264129638671, 0.011881983757019043, 0.011950143814086914, 0.011898719787597656, 0.011842016220092774, 0.01192784023284912, 0.011874367713928222, 0.011918720245361327, 0.011880831718444824, 0.011874079704284668, 0.011860223770141601, 0.01186406421661377, 0.011841535568237305, 0.011925503730773926, 0.012042240142822265, 0.012006815910339355, 0.011985504150390625, 0.011866111755371094, 0.01187564754486084, 0.011848383903503418, 0.011824864387512207, 0.011868448257446288, 0.011992992401123047, 0.011879551887512206, 0.011831775665283203, 0.011915776252746582, 0.011939552307128906, 0.01188595199584961, 0.011850655555725098, 0.01181056022644043, 0.011869983673095703, 0.011866592407226562, 0.011806719779968262, 0.011823103904724122, 0.011915264129638671, 0.011938048362731933, 0.011978367805480956, 0.011982848167419433, 0.011883872032165527, 0.011957183837890625, 0.011882335662841796, 0.011937055587768555, 0.011866847991943359, 0.011966464042663574, 0.012013567924499511, 0.011846879959106445, 0.011884480476379395, 0.011881600379943848, 0.01163475227355957, 0.012802559852600098, 0.01192313575744629, 0.011937376022338866, 0.012113408088684082, 0.013355487823486329, 0.01422985553741455, 0.012040384292602539, 0.012047743797302246, 0.011976672172546386, 0.011903552055358887, 0.011902976036071777, 0.012000800132751465, 0.011916928291320801, 0.012075712203979492, 0.011957792282104493, 0.011977343559265137, 0.011864031791687012, 0.011888671875, 0.011991040229797363, 0.012156448364257812, 0.011907456398010254, 0.011886591911315919, 0.011956319808959961, 0.012056032180786132, 0.011946528434753419, 0.011976703643798828, 0.012105728149414062, 0.01242643165588379, 0.012192383766174316, 0.012044544219970704, 0.012025792121887207, 0.011939840316772461, 0.011904959678649902, 0.011911231994628907, 0.011898207664489745, 0.011936415672302246, 0.011995327949523927, 0.011955904006958009, 0.011962495803833007, 0.011935615539550781, 0.01193996810913086, 0.01194803237915039, 0.011952383995056153, 0.012101280212402343, 0.011931936264038085, 0.011913215637207031, 0.011883584022521973, 0.011879263877868652, 0.011839391708374024, 0.011886591911315919, 0.011970560073852539, 0.011876352310180664, 0.011917407989501954, 0.011861760139465332, 0.011882016181945801, 0.01194048023223877, 0.011950079917907714, 0.01185971164703369, 0.011864319801330566, 0.011874303817749024, 0.011874303817749024, 0.013195263862609862, 0.011663455963134766, 0.011964735984802246, 0.011970879554748536, 0.011921152114868164, 0.012010016441345216, 0.012117247581481933, 0.012348447799682617, 0.011914976119995117, 0.011877920150756836, 0.011937824249267578, 0.011895071983337403, 0.011890591621398926, 0.01189094352722168, 0.011842911720275878, 0.011985024452209473, 0.011895071983337403, 0.011905088424682617, 0.011905280113220215, 0.011857983589172363, 0.011873439788818359, 0.011854559898376464, 0.011886591911315919, 0.011874272346496581, 0.011886367797851563, 0.011839743614196777, 0.011968511581420899, 0.011878399848937989, 0.011890687942504884, 0.011857536315917969, 0.011872480392456055, 0.011886752128601075, 0.01187820816040039, 0.011870271682739257, 0.011880576133728027, 0.0118920316696167, 0.01185478401184082, 0.011855680465698243, 0.01186348819732666, 0.011888768196105957, 0.011813440322875977, 0.011849535942077636, 0.011845088005065918, 0.011878432273864746, 0.011926176071166993, 0.011876192092895508, 0.011941887855529786, 0.011907008171081543, 0.011929311752319337, 0.01208080005645752, 0.012118528366088867, 0.01195638370513916, 0.011902015686035156, 0.011908160209655762, 0.011849632263183594, 0.011920672416687011, 0.012040063858032227, 0.012040672302246093, 0.011929216384887696, 0.01189782428741455, 0.011894559860229493, 0.011890560150146485, 0.011901056289672852, 0.011907072067260742]",tokens/s,83.0346692189184,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1542.725632,1546.584064,0.0,1168.113664,1154.613248,s,1,8.49350390625,8.49350390625,0.0,8.49350390625,8.49350390625,8.49350390625,8.49350390625,[8.49350390625],,kWh,4.117946988750039e-05,4.534977710168172e-06,1.3313066206013802e-05,5.902751380368236e-05,,MB,1532.3136,1777.270784,0.0,1369.440256,1323.44832,s,10,0.7752161865234374,0.07752161865234375,0.0008892316428154286,0.07745612716674805,0.07800702743530273,0.07897365608215332,0.0797469589996338,"[0.07994028472900391, 0.0769590072631836, 0.07756575775146485, 0.07749362945556641, 0.07741862487792969, 0.07645967864990234, 0.07779222106933593, 0.07695645141601562, 0.07706829071044922, 0.07756224060058593]",tokens/s,3302.304627410669,kWh,2.483743855932555e-06,2.73911777137436e-07,1.6543563140676495e-06,4.41201194713764e-06,tokens/kWh,58023414.95609138,MB,1535.668224,1798.242304,0.0,1390.411776,1377.26208,s,10,14.564807617187501,1.4564807617187499,0.012264741907122768,1.4549331665039062,1.4655761474609377,1.475623583984375,1.483661533203125,"[1.447943603515625, 1.4488612060546875, 1.4510828857421876, 1.45186474609375, 1.4580015869140626, 1.4856710205078125, 1.4610059814453125, 1.4633433837890626, 1.46022216796875, 1.43681103515625]",tokens/s,43.254948266982645,kWh,4.219580011198176e-05,4.653848486550735e-06,2.0553828119132958e-05,6.740347671766544e-05,tokens/kWh,934669.8874879943,,s,630,14.562426942825304,0.023114963401310026,0.0011230676053886127,0.022984591484069825,0.023376509284973146,0.023573486328125,0.02451666646957398,"[0.02337593650817871, 0.0228723201751709, 0.02276927947998047, 0.022599359512329102, 0.022639583587646485, 0.022726367950439454, 0.02265497589111328, 0.02265907287597656, 0.022983999252319337, 0.02302227210998535, 0.023000608444213866, 0.022987232208251954, 0.022992416381835936, 0.02297494316101074, 0.02308710479736328, 0.02312598419189453, 0.02322640037536621, 0.02326118469238281, 0.02336355209350586, 0.023187007904052735, 0.02308691215515137, 0.023191904067993162, 0.02300124740600586, 0.023142559051513672, 0.02305830383300781, 0.02309065628051758, 0.022999711990356445, 0.022925216674804686, 0.02297612762451172, 0.022876127243041992, 0.022929567337036133, 0.022913536071777343, 0.022762399673461914, 0.02286262321472168, 0.023071935653686523, 0.023212640762329102, 0.023346912384033202, 0.023265439987182616, 0.023296512603759766, 0.02336992073059082, 0.023311552047729493, 0.02293209648132324, 0.022820768356323243, 0.022863967895507813, 0.022796287536621093, 0.022816768646240236, 0.02287820816040039, 0.022775264739990236, 0.022740831375122072, 0.022685407638549804, 0.02294268798828125, 0.022877216339111328, 0.022834144592285156, 0.023060480117797853, 0.022962175369262695, 0.022964319229125976, 0.02293494415283203, 0.022858240127563476, 0.0229182071685791, 0.022844160079956054, 0.02286947250366211, 0.022984895706176758, 0.023154943466186524, 0.022935552597045897, 0.022816543579101563, 0.022840896606445314, 0.022944639205932617, 0.022894079208374024, 0.022872159957885742, 0.02289628791809082, 0.022957727432250975, 0.022783039093017578, 0.02294700813293457, 0.022874080657958984, 0.022942527770996094, 0.022834463119506834, 0.02288902473449707, 0.02274492835998535, 0.022851423263549806, 0.02287593650817871, 0.022805023193359374, 0.02280243110656738, 0.02275328063964844, 0.02290483283996582, 0.023350400924682616, 0.023073408126831056, 0.022804000854492187, 0.022870752334594728, 0.022829055786132812, 0.022861024856567384, 0.022836095809936525, 0.02292665672302246, 0.023068351745605467, 0.023069599151611327, 0.02309459114074707, 0.023175615310668946, 0.023474431991577147, 0.023201791763305665, 0.023103551864624025, 0.023156639099121093, 0.0231506233215332, 0.023220224380493162, 0.023183359146118163, 0.023052352905273438, 0.022943008422851564, 0.02309596824645996, 0.02295568084716797, 0.02279360008239746, 0.02298588752746582, 0.02301670455932617, 0.022811199188232423, 0.022668960571289063, 0.02285955238342285, 0.02282080078125, 0.02313279914855957, 0.024370975494384765, 0.023226591110229493, 0.022960128784179686, 0.02322329521179199, 0.022854496002197265, 0.0229803524017334, 0.02316739273071289, 0.023166431427001952, 0.023052831649780274, 0.02288844871520996, 0.022986751556396484, 0.023299808502197265, 0.023183839797973633, 0.022881376266479493, 0.02290777587890625, 0.022827072143554686, 0.023562240600585937, 0.02286988830566406, 0.02276355171203613, 0.022849824905395506, 0.022986207962036133, 0.022966623306274414, 0.022844800949096678, 0.022899072647094728, 0.022976768493652343, 0.02327142333984375, 0.02331443214416504, 0.02314854431152344, 0.023322816848754882, 0.023074623107910155, 0.022945823669433593, 0.02283839988708496, 0.022856544494628907, 0.023005184173583985, 0.023003135681152344, 0.022984575271606446, 0.02290496063232422, 0.02305023956298828, 0.02305788803100586, 0.023017311096191408, 0.02287481689453125, 0.023076160430908203, 0.022993600845336915, 0.02317679977416992, 0.02306038475036621, 0.023136768341064453, 0.023179264068603517, 0.023119295120239258, 0.02306105613708496, 0.023126016616821288, 0.023142463684082033, 0.023147455215454103, 0.023860160827636718, 0.02317478370666504, 0.023144575119018556, 0.023148672103881836, 0.02309872055053711, 0.023065439224243166, 0.022890592575073244, 0.022982751846313477, 0.02287392044067383, 0.022978559494018554, 0.022839296340942384, 0.022915071487426757, 0.022804479598999023, 0.022814495086669922, 0.022771999359130858, 0.022900672912597658, 0.022976512908935546, 0.022961536407470704, 0.022882944107055665, 0.022976255416870116, 0.023054592132568358, 0.02299660873413086, 0.023850208282470704, 0.023499551773071288, 0.023224319458007812, 0.023085216522216796, 0.022882144927978517, 0.02271392059326172, 0.02318777656555176, 0.022712575912475587, 0.022955904006958006, 0.023119871139526366, 0.022779903411865234, 0.022771135330200195, 0.022722272872924804, 0.023109952926635743, 0.022785856246948243, 0.02279702377319336, 0.022779903411865234, 0.022923263549804687, 0.02268505668640137, 0.022676095962524415, 0.023016992568969726, 0.023009664535522462, 0.02286367988586426, 0.022856191635131837, 0.02268547248840332, 0.022783647537231444, 0.022747488021850587, 0.022990720748901367, 0.02276335906982422, 0.022884639739990234, 0.02290480041503906, 0.022716255187988282, 0.022917024612426756, 0.025352319717407226, 0.02406825637817383, 0.023017471313476562, 0.022939647674560547, 0.023185407638549805, 0.022837247848510742, 0.023500864028930663, 0.023045984268188477, 0.02293071937561035, 0.022890560150146483, 0.022833919525146483, 0.02279167938232422, 0.02279270362854004, 0.022773439407348633, 0.02314067268371582, 0.0228121280670166, 0.022825504302978517, 0.022836671829223634, 0.022905471801757813, 0.023947200775146484, 0.024237503051757814, 0.023089727401733397, 0.023217248916625976, 0.02305683135986328, 0.023052768707275392, 0.02288640022277832, 0.02291472053527832, 0.022946048736572265, 0.02317283248901367, 0.023240095138549806, 0.023355871200561523, 0.023154560089111327, 0.02306255912780762, 0.023037567138671874, 0.02321830368041992, 0.023339263916015623, 0.023220096588134766, 0.02331660842895508, 0.023150463104248047, 0.0233123836517334, 0.02324287986755371, 0.023373983383178712, 0.023221376419067383, 0.02344799995422363, 0.023371904373168946, 0.023096864700317382, 0.023118463516235352, 0.02306892776489258, 0.02286755180358887, 0.02303104019165039, 0.022936319351196287, 0.02305449676513672, 0.02289254379272461, 0.02287311935424805, 0.022838239669799806, 0.022768735885620117, 0.022746015548706054, 0.022833120346069335, 0.022777055740356444, 0.022899520874023437, 0.02327552032470703, 0.02287161636352539, 0.023136703491210938, 0.023207935333251953, 0.02320191955566406, 0.02312716865539551, 0.023067392349243165, 0.023160831451416015, 0.023248895645141602, 0.023248096466064454, 0.0235467529296875, 0.023404447555541993, 0.023469760894775392, 0.02346015930175781, 0.02334671974182129, 0.02331286430358887, 0.023235647201538086, 0.022999807357788084, 0.02305766487121582, 0.023177568435668944, 0.023159391403198244, 0.0236844482421875, 0.02314931106567383, 0.022984607696533203, 0.022900928497314454, 0.022994752883911132, 0.022939647674560547, 0.023172384262084962, 0.023284000396728517, 0.022935840606689455, 0.023328927993774413, 0.023029312133789063, 0.023019968032836916, 0.023623680114746092, 0.02327398490905762, 0.023121599197387696, 0.023060192108154298, 0.022912639617919922, 0.0229385929107666, 0.022853023529052736, 0.023147232055664064, 0.022777727127075195, 0.022757375717163086, 0.0228002872467041, 0.022846944808959962, 0.022687711715698243, 0.02271504020690918, 0.02285977554321289, 0.023002527236938478, 0.0230380802154541, 0.02281929588317871, 0.02287311935424805, 0.02289148712158203, 0.022845439910888672, 0.022833152770996092, 0.023181312561035155, 0.02304332733154297, 0.023036672592163087, 0.023112768173217772, 0.023235519409179686, 0.028821504592895508, 0.02369049644470215, 0.02336582374572754, 0.023416608810424806, 0.023180063247680665, 0.023355295181274414, 0.023199840545654295, 0.02408995246887207, 0.02315523147583008, 0.02322649574279785, 0.023013408660888673, 0.02327142333984375, 0.023065696716308592, 0.04848057556152344, 0.02325721549987793, 0.022942144393920897, 0.0231014404296875, 0.02313852882385254, 0.022983552932739258, 0.022958112716674806, 0.022935487747192382, 0.023325632095336914, 0.02326540756225586, 0.022957952499389648, 0.023332639694213866, 0.023236671447753907, 0.0232509765625, 0.022951616287231445, 0.0230382080078125, 0.022960319519042968, 0.023162879943847657, 0.023000864028930663, 0.02290915107727051, 0.022943296432495118, 0.023144544601440428, 0.023014944076538087, 0.02294304084777832, 0.022931455612182617, 0.022923967361450196, 0.022955968856811525, 0.023376224517822265, 0.02324799919128418, 0.02322697639465332, 0.022962175369262695, 0.022934816360473634, 0.022895328521728514, 0.02304614448547363, 0.02301247978210449, 0.023050752639770508, 0.02296460723876953, 0.022960351943969726, 0.029336639404296875, 0.023007743835449217, 0.02288377571105957, 0.022813472747802734, 0.02288025665283203, 0.02275481605529785, 0.02290892791748047, 0.022696447372436524, 0.02289232063293457, 0.02292348861694336, 0.022808576583862306, 0.02324390411376953, 0.022875007629394532, 0.022988800048828126, 0.02289004707336426, 0.022872512817382812, 0.023019039154052734, 0.02288844871520996, 0.02276736068725586, 0.02457059288024902, 0.022874111175537108, 0.02292086410522461, 0.022814624786376952, 0.022804927825927735, 0.022854816436767577, 0.022891359329223634, 0.022920192718505858, 0.02288342475891113, 0.023138208389282225, 0.02308255958557129, 0.023251392364501952, 0.02331363105773926, 0.0232774715423584, 0.02352627182006836, 0.023633184432983397, 0.023511775970458983, 0.0243056640625, 0.023411903381347656, 0.023240703582763672, 0.023163904190063478, 0.023212928771972657, 0.02328057670593262, 0.023088640213012695, 0.023326911926269532, 0.023109312057495116, 0.0232825927734375, 0.023053600311279298, 0.02311827278137207, 0.024211328506469728, 0.023583072662353516, 0.0237922248840332, 0.02358268737792969, 0.02338822364807129, 0.023433183670043944, 0.023676351547241212, 0.023712255477905272, 0.023502464294433593, 0.023333087921142578, 0.023361759185791017, 0.023233535766601563, 0.023204864501953124, 0.02297964859008789, 0.023082975387573243, 0.022985471725463866, 0.0231364803314209, 0.022999040603637694, 0.023185407638549805, 0.023185440063476562, 0.023343072891235352, 0.023379072189331055, 0.023451679229736327, 0.023409439086914063, 0.023543487548828124, 0.0234520320892334, 0.023334144592285156, 0.023358240127563476, 0.023336864471435546, 0.023412799835205077, 0.02345599937438965, 0.02353036880493164, 0.023509536743164063, 0.0236527042388916, 0.023326719284057617, 0.023236320495605468, 0.022974752426147462, 0.023013343811035158, 0.022898687362670898, 0.022997024536132813, 0.02287740707397461, 0.023017696380615234, 0.02301804733276367, 0.022821983337402343, 0.02282588768005371, 0.02288844871520996, 0.022883903503417968, 0.02270867156982422, 0.022660224914550782, 0.022754175186157227, 0.022674495697021485, 0.02278495979309082, 0.022846687316894532, 0.022923263549804687, 0.023967872619628905, 0.02315945625305176, 0.022968063354492186, 0.022994943618774414, 0.02350921630859375, 0.023500831604003906, 0.023023616790771483, 0.023031679153442383, 0.023061824798583985, 0.02417292785644531, 0.023533376693725586, 0.02342233657836914, 0.023140607833862306, 0.0231976318359375, 0.02328985595703125, 0.023386560440063476, 0.02368707275390625, 0.023027807235717773, 0.022915071487426757, 0.02307891273498535, 0.023126016616821288, 0.023053407669067383, 0.022919551849365234, 0.02294432067871094, 0.022805503845214844, 0.0227906551361084, 0.022791776657104492, 0.023489343643188478, 0.027436479568481446, 0.023142688751220702, 0.02330793571472168, 0.023106143951416015, 0.02308924865722656, 0.023268608093261717, 0.02329267120361328, 0.023144447326660156, 0.024950464248657225, 0.02336185646057129, 0.023273664474487303, 0.022911903381347656, 0.024384639739990235, 0.02286732864379883, 0.022967744827270507, 0.02281705665588379, 0.022784704208374022, 0.022882303237915038, 0.02307030487060547, 0.022860191345214845, 0.02284774398803711, 0.022797311782836914, 0.022838016510009766, 0.022743040084838868, 0.02287820816040039, 0.02278825569152832, 0.023799232482910156, 0.022923679351806642, 0.022775808334350587, 0.022845439910888672, 0.023152639389038086, 0.023003135681152344, 0.022939647674560547, 0.023152639389038086, 0.022841344833374022, 0.022761472702026365, 0.023048383712768555, 0.022982463836669922, 0.022978559494018554, 0.022791360855102537, 0.022876991271972656, 0.022796287536621093, 0.022812671661376953, 0.02292643165588379, 0.02308336067199707, 0.02277827262878418, 0.022908319473266603, 0.02288252830505371, 0.02325337600708008, 0.022784255981445314, 0.022810016632080078, 0.02369571113586426, 0.023177248001098633, 0.022845407485961915, 0.022749183654785156, 0.022749343872070314, 0.022617952346801758, 0.022651136398315428, 0.022681343078613282, 0.022603967666625976, 0.022742847442626953, 0.02259542465209961, 0.022501535415649414, 0.022633663177490236, 0.022536544799804686, 0.0228353271484375, 0.02269164848327637, 0.022741247177124023, 0.02271571159362793, 0.02271673583984375, 0.022686368942260744, 0.022636287689208983, 0.022687744140625, 0.022933216094970704, 0.0228887996673584, 0.022800703048706055, 0.022812416076660156, 0.023159008026123046, 0.022841119766235353, 0.02279187202453613, 0.023216032028198243, 0.022848159790039062, 0.022879232406616212, 0.022786943435668946, 0.022773759841918945, 0.022806528091430665, 0.02273689651489258, 0.022940895080566407, 0.022823711395263672, 0.022765567779541016, 0.02285875129699707, 0.022852031707763672, 0.022818784713745117, 0.022704736709594726, 0.022758943557739258, 0.022662912368774414, 0.022827743530273437, 0.022662784576416014, 0.022700416564941407, 0.022691295623779296, 0.02273539161682129, 0.02283033561706543, 0.02265983963012695, 0.02272051239013672, 0.02277097511291504, 0.02277459144592285, 0.0227542724609375]",tokens/s,43.262019612080614,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1844.768768,2718.892032,0.0,2340.421632,2284.9536,s,1,9.19907421875,9.19907421875,0.0,9.19907421875,9.19907421875,9.19907421875,9.19907421875,[9.19907421875],,kWh,6.098281516240907e-05,6.719490604809796e-06,2.0080571620006982e-05,8.778287738722584e-05,,MB,1678.987264,3067.019264,0.0,2659.188736,2578.241536,s,10,0.8648788452148437,0.08648788452148438,0.0008393809032997067,0.08629409790039062,0.08777180709838868,0.08789060707092285,0.08798564704895019,"[0.08774540710449219, 0.08606150054931641, 0.08800940704345703, 0.08627216339111328, 0.08684835052490235, 0.08557577514648437, 0.08529590606689454, 0.08686914825439453, 0.08631603240966797, 0.08588515472412109]",tokens/s,2959.952153025633,kWh,2.7864079055557844e-06,3.070940271263117e-07,1.8510438088761718e-06,4.944545741558267e-06,tokens/kWh,51774220.19749016,MB,1678.987264,3067.019264,0.0,2659.188736,2578.244096,s,10,19.281598876953126,1.9281598876953123,0.015830657136111852,1.9316155395507812,1.942700866699219,1.9435855407714844,1.944293280029297,"[1.931883544921875, 1.9333028564453125, 1.9425042724609376, 1.8866256103515624, 1.9407467041015625, 1.922415771484375, 1.9283448486328125, 1.94447021484375, 1.9313475341796875, 1.91995751953125]",tokens/s,32.67363894562838,kWh,5.6074434942366776e-05,6.1849856945941064e-06,2.7192284981325268e-05,8.945170561828615e-05,tokens/kWh,704290.651190459,,s,630,19.27919442558289,0.03060189591362363,0.0005375087792078687,0.03058670425415039,0.030940896224975586,0.031229864120483397,0.03296664829254151,"[0.03162393569946289, 0.030367712020874023, 0.03023865509033203, 0.030432256698608398, 0.030716800689697267, 0.030396543502807617, 0.030440959930419922, 0.030578304290771484, 0.030530431747436523, 0.030824256896972657, 0.03241164779663086, 0.030447071075439452, 0.030411487579345704, 0.030443519592285157, 0.03015782356262207, 0.03042780876159668, 0.030471904754638672, 0.030531455993652343, 0.030479328155517578, 0.030258975982666015, 0.030298175811767577, 0.031333536148071287, 0.03057539176940918, 0.030471551895141603, 0.030570335388183593, 0.030421791076660157, 0.0304323844909668, 0.030411680221557616, 0.030556127548217772, 0.03056447982788086, 0.03060223960876465, 0.030667423248291015, 0.03064271926879883, 0.030603967666625976, 0.030726144790649414, 0.03066889572143555, 0.03089664077758789, 0.03064681625366211, 0.030690176010131836, 0.03054080009460449, 0.030851264953613282, 0.030590400695800782, 0.030561792373657228, 0.030532480239868164, 0.030511104583740234, 0.030635135650634766, 0.030540191650390625, 0.030582847595214843, 0.030654880523681642, 0.030564512252807617, 0.030965599060058593, 0.03058835220336914, 0.030640703201293945, 0.030715904235839843, 0.03080601692199707, 0.030869312286376953, 0.030895488739013672, 0.030747167587280272, 0.03092508888244629, 0.03101817512512207, 0.030965887069702148, 0.030780223846435546, 0.031192096710205078, 0.03163539123535156, 0.0309333438873291, 0.030966848373413087, 0.03063065528869629, 0.030729536056518555, 0.03073523139953613, 0.03076268768310547, 0.03066659164428711, 0.030513280868530272, 0.030716255187988283, 0.03059891128540039, 0.030762367248535157, 0.03060211181640625, 0.03060531234741211, 0.030728191375732423, 0.030611455917358397, 0.03052342414855957, 0.030445728302001953, 0.031063871383666994, 0.030330368041992187, 0.030378496170043946, 0.030439231872558595, 0.030519487380981446, 0.03060083198547363, 0.030787456512451173, 0.030685152053833008, 0.030718496322631836, 0.030768991470336914, 0.03066304016113281, 0.030856639862060546, 0.03069372749328613, 0.03060870361328125, 0.030638784408569337, 0.030744575500488282, 0.03076688003540039, 0.03080828857421875, 0.030800031661987304, 0.030653823852539064, 0.030704095840454103, 0.030817440032958984, 0.030892736434936525, 0.030779552459716798, 0.030707712173461913, 0.03168460845947266, 0.030651552200317383, 0.030595455169677734, 0.030623743057250977, 0.03058953666687012, 0.03049033546447754, 0.030574304580688477, 0.03062009620666504, 0.030537727355957032, 0.030500095367431642, 0.030624767303466797, 0.030404352188110353, 0.030693023681640626, 0.030563871383666993, 0.030489408493041992, 0.030529535293579102, 0.030470144271850585, 0.03058406448364258, 0.030689823150634767, 0.030555551528930663, 0.031258016586303713, 0.03082681655883789, 0.030720256805419923, 0.03059916877746582, 0.030604896545410157, 0.03077571105957031, 0.030674943923950194, 0.03082147216796875, 0.030631839752197267, 0.030450687408447266, 0.030487552642822265, 0.030483455657958985, 0.030702655792236327, 0.030747583389282226, 0.030594879150390625, 0.03046009635925293, 0.030619647979736327, 0.03304447937011719, 0.031723007202148434, 0.030788095474243164, 0.030766592025756836, 0.030718719482421875, 0.031735551834106444, 0.031083871841430664, 0.03096028709411621, 0.031151391983032226, 0.030845375061035157, 0.030824192047119142, 0.03089161682128906, 0.030827455520629883, 0.030724096298217773, 0.031659456253051756, 0.03237331390380859, 0.03091472053527832, 0.03070911979675293, 0.03092323112487793, 0.030801216125488282, 0.030869600296020507, 0.031363679885864255, 0.03129046440124512, 0.030651296615600586, 0.030504127502441407, 0.03030713653564453, 0.030259199142456054, 0.030038015365600586, 0.029995008468627928, 0.03003392028808594, 0.030117984771728515, 0.03017923164367676, 0.030354528427124022, 0.03034921646118164, 0.03035238456726074, 0.030517248153686522, 0.030418943405151368, 0.03043062400817871, 0.030419551849365234, 0.033509632110595704, 0.030666496276855467, 0.03078963279724121, 0.033481822967529294, 0.03021113586425781, 0.030112703323364257, 0.030148992538452147, 0.03061625671386719, 0.03014656066894531, 0.03027916717529297, 0.030331615447998048, 0.029944992065429686, 0.029802879333496093, 0.02980454444885254, 0.02985580825805664, 0.029597888946533202, 0.02961408042907715, 0.030299232482910155, 0.02972559928894043, 0.02967081642150879, 0.029565536499023437, 0.029518848419189454, 0.030112768173217775, 0.029701152801513673, 0.029557727813720704, 0.02961408042907715, 0.029732864379882814, 0.029563936233520507, 0.029585887908935547, 0.029708320617675782, 0.029587200164794922, 0.029524703979492188, 0.029577215194702147, 0.029554975509643554, 0.029615840911865234, 0.029644800186157227, 0.029708288192749024, 0.02984761619567871, 0.029880287170410157, 0.029467615127563476, 0.02979327964782715, 0.029615488052368164, 0.02980928039550781, 0.029847455978393556, 0.02988435173034668, 0.029804704666137695, 0.03039232063293457, 0.029898975372314455, 0.030296895980834963, 0.02995689582824707, 0.029902591705322265, 0.029915584564208984, 0.029839487075805665, 0.030096511840820312, 0.0301841926574707, 0.030148256301879884, 0.030205856323242186, 0.03015110397338867, 0.030101631164550783, 0.030284896850585937, 0.03038643264770508, 0.030315040588378906, 0.030379615783691406, 0.030185567855834962, 0.030273855209350584, 0.030345216751098632, 0.030361728668212892, 0.03026521682739258, 0.03056230354309082, 0.0304388484954834, 0.031356927871704104, 0.030468095779418947, 0.03058652877807617, 0.031006080627441406, 0.03051152038574219, 0.030614080429077147, 0.030650367736816408, 0.030447391510009764, 0.030522720336914062, 0.030446399688720704, 0.0305982723236084, 0.03055619239807129, 0.030626720428466796, 0.030655807495117187, 0.030726591110229493, 0.03093222427368164, 0.03074764823913574, 0.03097964859008789, 0.030908031463623045, 0.031369855880737305, 0.031080671310424805, 0.030842912673950194, 0.030832607269287108, 0.030887903213500975, 0.030833984375, 0.030866111755371094, 0.030803840637207033, 0.030736127853393556, 0.031062623977661134, 0.030719776153564454, 0.03088934326171875, 0.030574943542480467, 0.030617887496948243, 0.030875648498535156, 0.030668800354003906, 0.030932352066040038, 0.03080665588378906, 0.030643583297729492, 0.031063840866088866, 0.03129430389404297, 0.03082659149169922, 0.030871519088745115, 0.030613439559936523, 0.03064374351501465, 0.03073686408996582, 0.030698591232299805, 0.03081679916381836, 0.030869728088378907, 0.030763391494750977, 0.030796768188476563, 0.030679872512817383, 0.030901376724243163, 0.031050624847412108, 0.030867008209228514, 0.030931392669677735, 0.030758399963378907, 0.0306177921295166, 0.030951072692871093, 0.03079132843017578, 0.030950399398803712, 0.030806175231933595, 0.0310864315032959, 0.030732288360595703, 0.031123519897460938, 0.030662176132202148, 0.030794431686401367, 0.030455520629882812, 0.030303808212280275, 0.03035500717163086, 0.03035171127319336, 0.030447168350219725, 0.03052774429321289, 0.03034409523010254, 0.0304965763092041, 0.030370912551879882, 0.030421920776367187, 0.030559520721435546, 0.030446239471435547, 0.030533695220947267, 0.030496768951416016, 0.03059027290344238, 0.030939840316772462, 0.03148185539245606, 0.030375680923461913, 0.03037980842590332, 0.03030473518371582, 0.030341119766235353, 0.030195711135864257, 0.030331903457641602, 0.03053260803222656, 0.031339967727661136, 0.03046236801147461, 0.030590944290161133, 0.030484575271606446, 0.03044361686706543, 0.030451967239379884, 0.03052864074707031, 0.030521984100341796, 0.030414848327636718, 0.030543872833251953, 0.03078054428100586, 0.031298431396484375, 0.030380096435546875, 0.03034720039367676, 0.030291967391967774, 0.030240768432617186, 0.03038928031921387, 0.03019593620300293, 0.030255168914794923, 0.03036150360107422, 0.030085760116577147, 0.03015305519104004, 0.029876031875610352, 0.0298590087890625, 0.0299835205078125, 0.029755327224731447, 0.029857887268066406, 0.029902687072753908, 0.030128288269042968, 0.03040460777282715, 0.03058892822265625, 0.030547775268554688, 0.030521472930908202, 0.030564416885375978, 0.03269232177734375, 0.03277609634399414, 0.031184192657470702, 0.03063007926940918, 0.030554239273071288, 0.030486848831176756, 0.030451295852661132, 0.030519775390625, 0.030537023544311523, 0.030216896057128906, 0.030244863510131836, 0.03022585678100586, 0.03019219207763672, 0.03014656066894531, 0.030031040191650392, 0.03032966423034668, 0.03038768005371094, 0.030464544296264648, 0.030553279876708986, 0.030472671508789063, 0.030650527954101562, 0.03060140800476074, 0.030668800354003906, 0.03085433578491211, 0.030780223846435546, 0.03045782470703125, 0.030582815170288085, 0.030584159851074218, 0.030639936447143554, 0.031035327911376955, 0.030835424423217773, 0.03070518493652344, 0.030652223587036134, 0.030702112197875976, 0.03054732894897461, 0.030742496490478517, 0.030712352752685548, 0.030611263275146485, 0.030621631622314453, 0.030632640838623045, 0.030928735733032228, 0.03054761505126953, 0.030585344314575196, 0.03160678482055664, 0.03049852752685547, 0.030660160064697267, 0.030676767349243163, 0.030969823837280273, 0.030618560791015624, 0.030593055725097656, 0.03044668769836426, 0.03051817512512207, 0.030445632934570314, 0.030434272766113282, 0.030636287689208983, 0.030355360031127928, 0.03064313507080078, 0.030672767639160155, 0.030775104522705078, 0.030727487564086914, 0.030581567764282228, 0.030801376342773436, 0.030718559265136718, 0.03070332717895508, 0.030697216033935548, 0.03150819206237793, 0.030886367797851564, 0.030672191619873047, 0.03074105644226074, 0.030719392776489256, 0.03082499122619629, 0.030739519119262697, 0.03066975975036621, 0.030654464721679688, 0.03054719924926758, 0.030495487213134765, 0.03058687973022461, 0.030776927947998047, 0.03079542350769043, 0.030403327941894532, 0.031096832275390625, 0.03361715316772461, 0.03105254364013672, 0.03043440055847168, 0.03037481689453125, 0.030119935989379884, 0.030072959899902343, 0.03007798385620117, 0.030026752471923827, 0.030109535217285155, 0.029954048156738283, 0.03026848030090332, 0.030963712692260743, 0.030704160690307618, 0.030707647323608398, 0.03456169509887695, 0.030657344818115235, 0.03073859214782715, 0.03069443130493164, 0.03099945640563965, 0.030631711959838867, 0.030590463638305664, 0.03153126335144043, 0.030701728820800783, 0.03058095932006836, 0.030652095794677734, 0.03502889633178711, 0.030787359237670897, 0.030534271240234376, 0.030633983612060548, 0.030547840118408203, 0.03043084716796875, 0.030441408157348634, 0.030343744277954103, 0.030431232452392577, 0.030418943405151368, 0.030494720458984374, 0.03046723175048828, 0.03063484764099121, 0.030445119857788087, 0.030537343978881835, 0.030567232131958007, 0.03058278465270996, 0.03502870559692383, 0.030623647689819337, 0.03066713523864746, 0.03093212890625, 0.030712671279907226, 0.03119545555114746, 0.030795616149902345, 0.0307476806640625, 0.030598112106323242, 0.03040777587890625, 0.03067750358581543, 0.030562719345092772, 0.030641279220581054, 0.030601503372192383, 0.03063868713378906, 0.03058073616027832, 0.030406656265258788, 0.03059663963317871, 0.030582815170288085, 0.030540224075317382, 0.03057391929626465, 0.030496543884277343, 0.030622592926025392, 0.03058070373535156, 0.0305677433013916, 0.030711904525756836, 0.030603168487548828, 0.030640863418579103, 0.03057459259033203, 0.03069241523742676, 0.0306279354095459, 0.030767200469970703, 0.030696191787719727, 0.030780960083007812, 0.03075267219543457, 0.030714431762695314, 0.03055411148071289, 0.030470144271850585, 0.030674591064453124, 0.030677087783813478, 0.030664831161499023, 0.03131609535217285, 0.030697471618652345, 0.030813440322875977, 0.030712575912475587, 0.030732288360595703, 0.030860639572143553, 0.030681184768676758, 0.030545503616333007, 0.030561248779296876, 0.030621696472167968, 0.030664703369140626, 0.03060326385498047, 0.030674112319946288, 0.03049951934814453, 0.030468223571777343, 0.030658559799194338, 0.031041536331176758, 0.03068070411682129, 0.030466432571411132, 0.030406656265258788, 0.03038559913635254, 0.030427743911743164, 0.0304434871673584, 0.031024864196777344, 0.030576383590698242, 0.030730207443237303, 0.03074924850463867, 0.03132809638977051, 0.030873920440673826, 0.03086067199707031, 0.030976320266723634, 0.03078963279724121, 0.030793727874755858, 0.030485952377319336, 0.030503488540649413, 0.030469568252563476, 0.03051558494567871, 0.030570688247680663, 0.030410751342773438, 0.03057823944091797, 0.030414751052856445, 0.030298112869262695, 0.030147104263305663, 0.030310400009155275, 0.030357503890991212, 0.030319936752319337, 0.030581439971923828, 0.03057663917541504, 0.030695423126220703, 0.0306213436126709, 0.030503263473510744, 0.030596895217895506, 0.03051884841918945, 0.03096847915649414, 0.03055820846557617, 0.030533184051513673, 0.03062009620666504, 0.030720064163208008, 0.030610879898071288, 0.0307391357421875, 0.030515008926391602, 0.030526464462280273, 0.030655263900756836, 0.03069593620300293, 0.030392032623291015, 0.030280799865722657, 0.03018844795227051, 0.030281728744506835, 0.03026857566833496, 0.03013104057312012, 0.02986160087585449, 0.02984783935546875, 0.029971744537353514, 0.030116575241088867, 0.02997657585144043, 0.03015884780883789, 0.03037772750854492, 0.030201536178588867, 0.0303374080657959, 0.030507200241088866, 0.030435327529907227, 0.030307424545288085, 0.030792608261108398, 0.030426719665527343, 0.030299999237060546, 0.030218656539916993, 0.030310272216796875, 0.030639583587646485, 0.030550848007202147, 0.03058073616027832]",tokens/s,32.67771391754884,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4254.9248,4725.866496,0.0,4347.396096,4328.833024,s,1,10.1469296875,10.1469296875,0.0,10.1469296875,10.1469296875,10.1469296875,10.1469296875,[10.1469296875],,kWh,8.911540633337153e-05,9.822733647001379e-06,3.0221413065995018e-05,0.00012915955304636793,,MB,1385.30816,5212.40576,0.0,4804.575232,4748.27776,s,10,3.353415740966797,0.3353415740966797,0.002830763496809025,0.3359424896240234,0.3385506988525391,0.3389105422973633,0.3391984170532226,"[0.33013101196289063, 0.3346091613769531, 0.3317735290527344, 0.33279052734375, 0.3358599853515625, 0.3360249938964844, 0.33715487670898436, 0.3373305358886719, 0.3392703857421875, 0.3384707336425781]",tokens/s,763.4007226500185,kWh,9.747006747915444e-06,1.0745657501540521e-06,6.462662577533251e-06,1.7284235075602748e-05,tokens/kWh,14811184.809754886,MB,1388.466176,5321.457664,0.0,4913.627136,4878.043648,s,10,18.212191162109374,1.8212191162109375,0.004827572697774101,1.8191776123046877,1.827504541015625,1.8289905517578124,1.8301793603515624,"[1.8162979736328124, 1.8251033935546874, 1.818055419921875, 1.8238345947265624, 1.8304765625, 1.8177127685546874, 1.8202998046875, 1.8162591552734375, 1.8169771728515625, 1.82717431640625]",tokens/s,34.59221322641949,kWh,5.342437244375331e-05,5.892963448966416e-06,3.5353204208465884e-05,9.467054010118562e-05,tokens/kWh,665465.729176832,,s,630,18.20843064689638,0.028902270868089466,0.0012119222805392464,0.028918832778930665,0.029754111671447755,0.03004606399536133,0.03278376026153565,"[0.03177004814147949, 0.02955731201171875, 0.028858367919921874, 0.02851840019226074, 0.027850751876831056, 0.02754867172241211, 0.02753638458251953, 0.027578367233276366, 0.028203008651733398, 0.029016063690185546, 0.029050464630126952, 0.028816831588745116, 0.029090656280517577, 0.028483680725097656, 0.02849795150756836, 0.028086271286010742, 0.027801599502563477, 0.028544736862182618, 0.029761024475097656, 0.029354784011840822, 0.029095775604248048, 0.028499647140502928, 0.028977632522583008, 0.029050880432128907, 0.028440576553344726, 0.027906047821044923, 0.0275883846282959, 0.028242143630981445, 0.029832672119140625, 0.029891103744506837, 0.029163520812988283, 0.02832156753540039, 0.028291296005249024, 0.03003392028808594, 0.03014656066894531, 0.02898739242553711, 0.02931692886352539, 0.029257919311523436, 0.029247488021850586, 0.02872230339050293, 0.02824665641784668, 0.028120479583740234, 0.028841951370239257, 0.02945849609375, 0.029219648361206055, 0.028964351654052735, 0.029443904876708983, 0.029078208923339843, 0.028456960678100586, 0.02815385627746582, 0.02800230407714844, 0.028374208450317382, 0.02902300834655762, 0.029276191711425783, 0.02873520088195801, 0.02784079933166504, 0.02813542366027832, 0.02848348808288574, 0.029728511810302734, 0.029505887985229493, 0.029249536514282228, 0.029326847076416016, 0.029399551391601563, 0.03220675277709961, 0.029395040512084962, 0.028828832626342775, 0.02757923126220703, 0.027123647689819334, 0.028083904266357422, 0.028434528350830077, 0.027760223388671876, 0.027682880401611328, 0.027588735580444335, 0.02812774467468262, 0.028112895965576173, 0.027661504745483397, 0.02776927947998047, 0.028299455642700196, 0.02780998420715332, 0.027906047821044923, 0.028274688720703125, 0.027883520126342775, 0.02781795120239258, 0.028302560806274413, 0.02907804870605469, 0.028835744857788087, 0.02902396774291992, 0.028891679763793945, 0.028144960403442384, 0.027810623168945312, 0.02791993522644043, 0.028352960586547852, 0.028080127716064454, 0.028243967056274414, 0.028663808822631837, 0.030811775207519532, 0.0302325439453125, 0.029684127807617186, 0.029362016677856446, 0.029407392501831053, 0.02933145523071289, 0.02958745574951172, 0.029133920669555665, 0.028629919052124024, 0.029247488021850586, 0.02949734306335449, 0.028338176727294922, 0.02771968078613281, 0.027678752899169923, 0.0284704647064209, 0.029706687927246095, 0.02988172721862793, 0.029518592834472657, 0.029104352951049805, 0.028655071258544922, 0.02846352005004883, 0.028051584243774415, 0.02840332794189453, 0.027990400314331056, 0.049065662384033204, 0.027023231506347656, 0.029034143447875978, 0.02974595260620117, 0.02917328071594238, 0.02923535919189453, 0.028907487869262696, 0.03366329574584961, 0.03047235107421875, 0.028802879333496095, 0.0285534725189209, 0.02906924819946289, 0.028940288543701172, 0.02834432029724121, 0.027701248168945314, 0.028097631454467774, 0.02898012733459473, 0.02907916831970215, 0.02828940773010254, 0.02789580726623535, 0.028030975341796875, 0.028017887115478514, 0.02787942314147949, 0.02818332862854004, 0.029312160491943358, 0.029037055969238282, 0.029102432250976563, 0.028553216934204102, 0.028193855285644533, 0.027881824493408203, 0.028227231979370118, 0.029346879959106446, 0.029349760055541994, 0.028366367340087892, 0.027639776229858398, 0.02810220718383789, 0.028253120422363283, 0.028112895965576173, 0.028084224700927734, 0.02901558494567871, 0.0298111686706543, 0.029353696823120116, 0.02922224044799805, 0.028330944061279298, 0.028077184677124025, 0.028137887954711914, 0.028527008056640626, 0.029323328018188478, 0.02897100830078125, 0.029474815368652343, 0.02949718475341797, 0.029005247116088866, 0.029147872924804686, 0.02975334358215332, 0.029123584747314454, 0.02845737648010254, 0.02769366455078125, 0.028088319778442384, 0.028704767227172853, 0.02875596809387207, 0.029290496826171877, 0.029501312255859374, 0.028831872940063476, 0.029509632110595704, 0.02961961555480957, 0.029037151336669922, 0.029675519943237305, 0.02915328025817871, 0.029683935165405274, 0.029357215881347657, 0.03265171051025391, 0.02995840072631836, 0.028925952911376954, 0.028827104568481445, 0.029544992446899412, 0.028612607955932616, 0.027246368408203124, 0.027302112579345703, 0.02760704040527344, 0.02852060890197754, 0.029222751617431642, 0.028579679489135743, 0.02761334419250488, 0.027280895233154297, 0.02804377555847168, 0.02812886428833008, 0.027892127990722656, 0.02814771270751953, 0.027674623489379883, 0.028676095962524413, 0.029519872665405275, 0.029224960327148438, 0.02853273582458496, 0.028516128540039064, 0.02798944091796875, 0.027879199981689452, 0.028999935150146483, 0.029727487564086913, 0.029016063690185546, 0.02873958396911621, 0.027637407302856444, 0.028309856414794922, 0.030441087722778322, 0.0302923526763916, 0.029014015197753908, 0.029320735931396485, 0.02987571144104004, 0.029895647048950195, 0.029648895263671874, 0.029173536300659178, 0.02855891227722168, 0.028868511199951173, 0.030233343124389647, 0.029853055953979492, 0.028887359619140626, 0.029366592407226562, 0.02937001609802246, 0.02905241584777832, 0.02918454360961914, 0.02913862419128418, 0.02927199935913086, 0.028176544189453125, 0.028082719802856447, 0.02814566421508789, 0.029540096282958984, 0.029859296798706053, 0.029446943283081055, 0.02875596809387207, 0.02950553512573242, 0.029699232101440428, 0.029166431427001954, 0.028903423309326173, 0.030087167739868165, 0.03285168075561523, 0.029694303512573243, 0.028843551635742187, 0.02815433692932129, 0.028008447647094727, 0.028094463348388672, 0.02776380729675293, 0.02816092872619629, 0.028049407958984376, 0.028196096420288086, 0.02897724723815918, 0.0291843204498291, 0.02885443115234375, 0.029200576782226564, 0.02886409568786621, 0.029180320739746093, 0.028979103088378907, 0.028541023254394532, 0.027439104080200196, 0.027725887298583985, 0.02868332862854004, 0.029538944244384767, 0.02894054412841797, 0.02797772789001465, 0.02775859260559082, 0.027602336883544923, 0.02832035255432129, 0.029104127883911132, 0.028778495788574218, 0.029462528228759766, 0.029470720291137696, 0.029100032806396486, 0.028841567993164063, 0.030443935394287108, 0.035036224365234375, 0.02868729591369629, 0.0283504638671875, 0.029900800704956054, 0.02959564781188965, 0.028652544021606444, 0.029189023971557617, 0.02941961669921875, 0.02958950424194336, 0.028628095626831055, 0.02955763244628906, 0.029542335510253905, 0.02910214424133301, 0.028231679916381838, 0.02815782356262207, 0.029218944549560547, 0.029035968780517576, 0.029295167922973632, 0.02936627197265625, 0.028807167053222657, 0.0296711368560791, 0.030003488540649412, 0.029646848678588866, 0.02964838409423828, 0.02926748847961426, 0.029231327056884766, 0.029106943130493165, 0.02851840019226074, 0.028858335494995117, 0.03283769607543945, 0.029829055786132812, 0.02857740783691406, 0.027896192550659178, 0.028024288177490236, 0.028090911865234373, 0.028628992080688476, 0.027662336349487306, 0.028188255310058592, 0.02768479919433594, 0.02809084892272949, 0.028700063705444336, 0.027802207946777343, 0.02736947250366211, 0.02789295959472656, 0.028508960723876955, 0.027998207092285156, 0.02856879997253418, 0.027625759124755858, 0.02765180778503418, 0.02817100715637207, 0.028051488876342772, 0.027772031784057617, 0.027865407943725586, 0.028711488723754883, 0.029710336685180663, 0.02876345634460449, 0.027781824111938476, 0.027627519607543945, 0.02832313537597656, 0.029420223236083985, 0.029388799667358398, 0.029212480545043946, 0.0292388801574707, 0.02929929542541504, 0.029246591567993165, 0.02997657585144043, 0.029932416915893555, 0.029324384689331056, 0.028967552185058594, 0.029036832809448243, 0.030055999755859375, 0.029450687408447265, 0.029083648681640626, 0.02895382308959961, 0.02847772789001465, 0.0295897274017334, 0.02974764823913574, 0.029611072540283202, 0.029274911880493165, 0.028730623245239256, 0.029360895156860353, 0.029306880950927733, 0.029175296783447265, 0.029204992294311522, 0.029278207778930664, 0.029259679794311523, 0.02896291160583496, 0.029479167938232423, 0.029036575317382813, 0.028708576202392578, 0.029732864379882814, 0.029398624420166015, 0.03294822311401367, 0.029867839813232423, 0.029040864944458008, 0.027706783294677736, 0.027068992614746094, 0.028096511840820314, 0.028366847991943358, 0.028911455154418945, 0.029558944702148437, 0.028681535720825196, 0.027322656631469728, 0.027701663970947265, 0.028106752395629882, 0.028239839553833006, 0.02887068748474121, 0.029417695999145507, 0.029037343978881837, 0.028327007293701172, 0.027469728469848635, 0.02838937568664551, 0.028026559829711913, 0.028200544357299805, 0.028002527236938475, 0.028614463806152343, 0.0285150089263916, 0.027967487335205078, 0.02826755142211914, 0.02788150405883789, 0.02845359992980957, 0.027803712844848633, 0.0280864315032959, 0.02958336067199707, 0.029605119705200196, 0.02921548843383789, 0.0294072322845459, 0.02944819259643555, 0.029899999618530272, 0.02978019142150879, 0.029477439880371093, 0.02913603210449219, 0.02920956802368164, 0.029604799270629884, 0.029619136810302735, 0.028831520080566407, 0.029516000747680664, 0.029054527282714845, 0.02801299285888672, 0.0281395206451416, 0.028411872863769533, 0.02853276824951172, 0.029585407257080077, 0.02992291259765625, 0.029622175216674804, 0.029356544494628906, 0.029077503204345705, 0.028767967224121095, 0.029638944625854494, 0.03003392028808594, 0.02965878486633301, 0.029339424133300783, 0.02908831977844238, 0.028753536224365234, 0.02960367965698242, 0.03287036895751953, 0.029923360824584962, 0.028441728591918944, 0.02773593521118164, 0.028050432205200194, 0.02817228889465332, 0.028622848510742187, 0.029095935821533202, 0.028794879913330077, 0.02750223922729492, 0.02761337661743164, 0.02830294418334961, 0.029206464767456055, 0.028797760009765624, 0.028532543182373048, 0.02761849594116211, 0.027848512649536132, 0.02840060806274414, 0.027805728912353514, 0.027852800369262694, 0.02829516792297363, 0.028991487503051756, 0.0293703670501709, 0.028854272842407228, 0.02759212875366211, 0.0277509765625, 0.027936767578125, 0.02906284713745117, 0.029065536499023437, 0.028401376724243164, 0.027666719436645507, 0.028359872817993164, 0.030460735321044922, 0.030402559280395508, 0.029736640930175782, 0.029134496688842774, 0.029231775283813478, 0.02906857681274414, 0.028383359909057618, 0.02871766471862793, 0.027420448303222655, 0.028801248550415038, 0.028780160903930666, 0.028649280548095703, 0.02897724723815918, 0.030394847869873048, 0.029533824920654296, 0.028655040740966798, 0.027595712661743165, 0.02750054359436035, 0.028804512023925782, 0.02987654495239258, 0.029278495788574218, 0.029147136688232423, 0.029703872680664063, 0.029681983947753905, 0.02930633544921875, 0.028993888854980467, 0.02929871940612793, 0.02915305519104004, 0.028952320098876952, 0.030128671646118165, 0.029552703857421876, 0.03210588836669922, 0.02944879913330078, 0.028563199996948244, 0.02880281639099121, 0.028264959335327147, 0.02757212829589844, 0.027670623779296875, 0.027527328491210937, 0.02826019287109375, 0.02908345603942871, 0.028698816299438476, 0.027584512710571288, 0.02759065628051758, 0.02790809631347656, 0.028243648529052735, 0.028757856369018554, 0.029156864166259764, 0.028928159713745117, 0.029430591583251953, 0.02897305679321289, 0.028297216415405273, 0.027876447677612305, 0.027960224151611326, 0.028499967575073244, 0.027682815551757813, 0.027893760681152343, 0.028526496887207032, 0.02773955154418945, 0.028209856033325195, 0.029057024002075195, 0.02964179229736328, 0.029565887451171877, 0.029013376235961914, 0.029913728713989257, 0.029327360153198243, 0.02859779167175293, 0.02831817626953125, 0.027596128463745116, 0.028641984939575194, 0.02966524887084961, 0.029534208297729493, 0.02853887939453125, 0.028006399154663086, 0.028925216674804688, 0.03010211181640625, 0.02967900848388672, 0.029370271682739257, 0.028912448883056642, 0.02792038345336914, 0.02856550407409668, 0.02892995262145996, 0.030169183731079102, 0.029775871276855468, 0.029261823654174804, 0.029005311965942384, 0.029069440841674805, 0.029980863571166992, 0.029673055648803712, 0.029737567901611327, 0.029198335647583007, 0.02897100830078125, 0.029057024002075195, 0.029582880020141603, 0.03211884689331055, 0.029147136688232423, 0.02846031951904297, 0.027697887420654297, 0.027885087966918944, 0.028352991104125976, 0.028851648330688477, 0.029515647888183595, 0.028973760604858397, 0.02792857551574707, 0.02781955146789551, 0.02912918472290039, 0.029400896072387696, 0.028452255249023437, 0.027707584381103517, 0.02823360061645508, 0.029260320663452147, 0.02954022407531738, 0.02847148895263672, 0.02872563171386719, 0.02875366401672363, 0.029419519424438476, 0.029122560501098634, 0.028255807876586915, 0.027680831909179686, 0.02866419219970703, 0.030228063583374022, 0.029621696472167967, 0.029305728912353515, 0.028821599960327147, 0.028109920501708983, 0.02910095977783203, 0.030103551864624024, 0.029800447463989257, 0.028933696746826172, 0.029706687927246095, 0.029026304244995117, 0.029446016311645506, 0.029431583404541016, 0.030105951309204102, 0.030003200531005858, 0.029421567916870117, 0.02910310363769531, 0.029535232543945314, 0.029419519424438476, 0.028901376724243165, 0.027729055404663087, 0.027479904174804688, 0.027491167068481447, 0.02749251174926758, 0.027922431945800782, 0.029416479110717774, 0.029783008575439453, 0.029648832321166992, 0.029902048110961914, 0.029571456909179686, 0.029051359176635743, 0.028383232116699218, 0.028180383682250978, 0.029373823165893556, 0.03010633659362793, 0.02998067283630371, 0.029626047134399414]",tokens/s,34.599357419492044,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1301.057536,1083.113472,0.0,706.740224,675.13344,s,1,8.02769287109375,8.02769287109375,0.0,8.02769287109375,8.02769287109375,8.02769287109375,8.02769287109375,[8.02769287109375],,kWh,3.416082585419341e-05,3.75901378485943e-06,1.1313064606000012e-05,4.923290424505285e-05,,MB,1338.507264,1403.977728,0.0,996.1472,942.731264,s,10,0.32046176147460936,0.032046176147460935,0.0002986242411204077,0.031976847648620604,0.032185202789306644,0.03253535423278808,0.03281547538757324,"[0.03288550567626953, 0.03196047973632812, 0.03190598487854004, 0.031801536560058595, 0.032107391357421876, 0.032062145233154295, 0.03206054306030273, 0.031993215560913085, 0.03177020835876465, 0.03191475105285645]",tokens/s,7988.472597230083,kWh,1.0090062612644241e-06,1.1127539921727499e-07,6.678242890483014e-07,1.7881059495300003e-06,tokens/kWh,143168250.21877983,MB,1356.914688,1412.366336,0.0,1004.535808,942.733824,s,10,13.1520458984375,1.3152045898437499,0.01372797262127614,1.311053955078125,1.3325357055664062,1.3414024963378908,1.3484959289550782,"[1.3132799072265624, 1.3078309326171875, 1.304052734375, 1.3305653076171875, 1.350269287109375, 1.311547607421875, 1.310560302734375, 1.306673583984375, 1.3040201416015624, 1.31324609375]",tokens/s,47.90129268594218,kWh,3.860332580165441e-05,4.257519750925985e-06,1.6431994946351634e-05,5.929284049893203e-05,tokens/kWh,1062522.8858977458,,s,630,13.149944005966175,0.02087292699359712,0.000397384396262056,0.02076681613922119,0.021191869163513184,0.0216299822807312,0.022143090381622318,"[0.020803615570068358, 0.02102272033691406, 0.020967424392700194, 0.020914176940917968, 0.02078268814086914, 0.02070515251159668, 0.020525600433349608, 0.020752384185791017, 0.02056300735473633, 0.020646848678588868, 0.020609024047851563, 0.02064787292480469, 0.02072991943359375, 0.02066022491455078, 0.020592639923095703, 0.020717567443847656, 0.021561344146728514, 0.021295103073120117, 0.020914176940917968, 0.020957183837890626, 0.020715520858764647, 0.020868576049804688, 0.02091881561279297, 0.020616735458374023, 0.020797920227050782, 0.020897184371948242, 0.020827968597412108, 0.020982175827026366, 0.020805376052856445, 0.020769407272338867, 0.020746240615844725, 0.02074790382385254, 0.02074457550048828, 0.020785120010375975, 0.020730047225952147, 0.02086425590515137, 0.021059648513793945, 0.020776992797851564, 0.02080758476257324, 0.0208056640625, 0.02102739143371582, 0.020893024444580077, 0.020856735229492187, 0.02086579132080078, 0.020836383819580077, 0.02096691131591797, 0.02086960029602051, 0.020856096267700196, 0.02085308837890625, 0.02098419189453125, 0.020908031463623047, 0.020965375900268556, 0.02168185615539551, 0.020817760467529298, 0.02070694351196289, 0.020800159454345702, 0.020863168716430663, 0.020792448043823242, 0.02071340751647949, 0.020703487396240235, 0.02085536003112793, 0.02086720085144043, 0.020762624740600585, 0.020867071151733398, 0.02082614326477051, 0.02055369567871094, 0.02063148880004883, 0.020639808654785156, 0.020736000061035157, 0.020754432678222655, 0.020614463806152342, 0.020672639846801757, 0.020644416809082033, 0.020746240615844725, 0.020658176422119142, 0.02069231986999512, 0.020676416397094728, 0.020724159240722656, 0.02065558433532715, 0.020628416061401367, 0.02057164764404297, 0.020597248077392577, 0.020625408172607423, 0.020717567443847656, 0.020641792297363282, 0.02064134407043457, 0.020813728332519533, 0.02062748718261719, 0.02073651123046875, 0.02087321662902832, 0.02090595245361328, 0.02074012756347656, 0.02080499267578125, 0.020865055084228517, 0.020836959838867186, 0.020817920684814452, 0.02163644790649414, 0.02080732727050781, 0.02104422378540039, 0.020848096847534178, 0.020754655838012694, 0.020833856582641603, 0.0209783992767334, 0.020758560180664062, 0.020789247512817383, 0.020669567108154298, 0.020734848022460936, 0.0209039363861084, 0.0208035831451416, 0.020728063583374024, 0.02068454360961914, 0.02074844741821289, 0.020813663482666014, 0.02072985649108887, 0.020674560546875, 0.0207127685546875, 0.020679359436035157, 0.02077846336364746, 0.02068534469604492, 0.020719615936279297, 0.020764671325683593, 0.020813440322875975, 0.020822399139404296, 0.02069708824157715, 0.020731008529663086, 0.02071436882019043, 0.02080143928527832, 0.02072380828857422, 0.021032960891723632, 0.022084768295288087, 0.021023584365844727, 0.020783103942871094, 0.02090825653076172, 0.020676383972167967, 0.02065407943725586, 0.020543487548828124, 0.020557535171508788, 0.02057347106933594, 0.020583423614501953, 0.020663936614990233, 0.020803903579711912, 0.02055766487121582, 0.020647584915161134, 0.020502464294433594, 0.020701311111450196, 0.020656639099121094, 0.020600831985473633, 0.020508672714233397, 0.020559871673583984, 0.02069708824157715, 0.020666048049926757, 0.020547903060913086, 0.02057580757141113, 0.02057200050354004, 0.020648544311523437, 0.02055513572692871, 0.020718208312988283, 0.020772447586059572, 0.020740480422973634, 0.020709407806396483, 0.020699071884155273, 0.020598848342895507, 0.02057823944091797, 0.020733280181884764, 0.02078323173522949, 0.020533855438232423, 0.02054960060119629, 0.02059062385559082, 0.020626911163330076, 0.020720352172851564, 0.0205980167388916, 0.020513343811035156, 0.020663423538208006, 0.020732799530029298, 0.020684799194335936, 0.020565376281738282, 0.020602655410766602, 0.020541343688964844, 0.02082912063598633, 0.02059395217895508, 0.020698944091796876, 0.020650911331176757, 0.020665760040283202, 0.02083008003234863, 0.02077743911743164, 0.020668672561645507, 0.020789247512817383, 0.02072166442871094, 0.020929887771606447, 0.021078399658203125, 0.022013952255249023, 0.021489919662475584, 0.021057376861572264, 0.020739871978759764, 0.020908447265625, 0.02084614372253418, 0.020856927871704102, 0.02090867233276367, 0.020772863388061523, 0.020774751663208007, 0.02591926383972168, 0.023018911361694337, 0.020912864685058593, 0.020879392623901368, 0.020823904037475586, 0.02080191993713379, 0.020745439529418944, 0.021135520935058594, 0.020911775588989257, 0.02091107177734375, 0.020965375900268556, 0.02169036865234375, 0.021616384506225585, 0.021528831481933595, 0.02105753517150879, 0.020997280120849608, 0.020928352355957032, 0.02085785675048828, 0.02090297508239746, 0.020949951171875, 0.02089129638671875, 0.020957504272460938, 0.020938495635986328, 0.020939008712768555, 0.02085068893432617, 0.020867103576660155, 0.020839744567871094, 0.021121728897094728, 0.02084390449523926, 0.020898431777954102, 0.020811904907226564, 0.02091948890686035, 0.021083839416503908, 0.020925312042236327, 0.02085411262512207, 0.020980512619018555, 0.02100223922729492, 0.021044832229614258, 0.020879711151123047, 0.02097158432006836, 0.020872352600097656, 0.020938655853271485, 0.020904895782470703, 0.020973567962646485, 0.020914335250854493, 0.021046367645263672, 0.021612831115722656, 0.021154272079467774, 0.021102592468261717, 0.021004064559936524, 0.02105686378479004, 0.021160959243774414, 0.021361919403076173, 0.02235228729248047, 0.021129728317260742, 0.021142879486083985, 0.020959903717041015, 0.020858879089355468, 0.020883455276489257, 0.02105446434020996, 0.020976512908935548, 0.021299327850341797, 0.02111692810058594, 0.020996095657348633, 0.020992000579833983, 0.020917375564575194, 0.020846559524536134, 0.02095110321044922, 0.021003103256225585, 0.02090332794189453, 0.020832191467285155, 0.020806304931640623, 0.021170175552368165, 0.02271027183532715, 0.021127296447753907, 0.020954399108886718, 0.020978271484375, 0.020914176940917968, 0.021139455795288087, 0.02106368064880371, 0.021132736206054686, 0.02128748893737793, 0.021555200576782226, 0.022030336380004883, 0.021534719467163087, 0.0239553279876709, 0.021263711929321288, 0.021317888259887695, 0.02102022361755371, 0.021119232177734374, 0.02120163154602051, 0.02128505516052246, 0.02160950469970703, 0.021902111053466795, 0.021782527923583983, 0.0218603515625, 0.021876735687255858, 0.02216691207885742, 0.021895807266235353, 0.02181318473815918, 0.021825599670410156, 0.02185558319091797, 0.021772159576416015, 0.021752607345581054, 0.02170591926574707, 0.021715776443481445, 0.021817344665527344, 0.021817344665527344, 0.021654624938964844, 0.021722047805786134, 0.02175587272644043, 0.021741567611694337, 0.021366239547729492, 0.021272863388061523, 0.021209344863891602, 0.021219327926635743, 0.021141504287719725, 0.020916223526000977, 0.020963327407836914, 0.02084000015258789, 0.02099580764770508, 0.020850528717041017, 0.020757024765014648, 0.020594303131103515, 0.02068332862854004, 0.02059833526611328, 0.02054640007019043, 0.02058527946472168, 0.020648895263671876, 0.02048204803466797, 0.020477951049804686, 0.020573919296264648, 0.020549919128417967, 0.020658176422119142, 0.02074777603149414, 0.02076691246032715, 0.0207108154296875, 0.020607904434204103, 0.020580352783203124, 0.02061296081542969, 0.02066201591491699, 0.020648351669311525, 0.020586496353149415, 0.021059135437011718, 0.020550079345703125, 0.020560991287231444, 0.020691871643066406, 0.02060073661804199, 0.02049648094177246, 0.021190656661987304, 0.020704639434814452, 0.020609184265136717, 0.020635200500488282, 0.02083113670349121, 0.020450496673583986, 0.020671295166015624, 0.02061520004272461, 0.020684768676757812, 0.020862943649291994, 0.020811807632446288, 0.02094895935058594, 0.020916255950927734, 0.021028863906860353, 0.02109775924682617, 0.02096988868713379, 0.021184383392333986, 0.02109075164794922, 0.021116352081298827, 0.02091270446777344, 0.0210882568359375, 0.021079296112060546, 0.020920799255371093, 0.020689184188842774, 0.0213087043762207, 0.02101113510131836, 0.021622079849243164, 0.02142486381530762, 0.02087936019897461, 0.020807071685791014, 0.020793760299682617, 0.02073139190673828, 0.020726463317871095, 0.02125004768371582, 0.020942720413208007, 0.020748416900634767, 0.020708799362182617, 0.020720191955566406, 0.020760576248168947, 0.02080143928527832, 0.021173856735229493, 0.02105094337463379, 0.021190784454345704, 0.020891584396362305, 0.020949440002441408, 0.020898208618164063, 0.0209932804107666, 0.021003040313720703, 0.021030080795288085, 0.02100511932373047, 0.021112831115722656, 0.021168127059936523, 0.0209849910736084, 0.021091039657592774, 0.020893375396728517, 0.020830656051635744, 0.02068396759033203, 0.020615999221801757, 0.020604703903198244, 0.020633823394775392, 0.020742143630981445, 0.020703231811523438, 0.020700319290161133, 0.020585311889648437, 0.020696992874145507, 0.02064188766479492, 0.02049228858947754, 0.020684223175048828, 0.020550207138061525, 0.020590591430664062, 0.020690944671630858, 0.020584320068359373, 0.020574335098266602, 0.020724960327148437, 0.020566112518310548, 0.02079404830932617, 0.020940799713134766, 0.02094499206542969, 0.02079270362854004, 0.02078505516052246, 0.020875904083251955, 0.020819583892822267, 0.020882015228271485, 0.02067375946044922, 0.020548160552978516, 0.02066431999206543, 0.02069001579284668, 0.02077343940734863, 0.020811807632446288, 0.02072403144836426, 0.020684768676757812, 0.020637727737426757, 0.020763071060180664, 0.020762048721313476, 0.02056687927246094, 0.020629215240478515, 0.020628896713256836, 0.020533855438232423, 0.020657279968261718, 0.02068160057067871, 0.020766719818115235, 0.020760576248168947, 0.020736000061035157, 0.02069424057006836, 0.020801471710205077, 0.020730720520019532, 0.020692895889282227, 0.020665887832641602, 0.021047712326049805, 0.020658336639404296, 0.0207391357421875, 0.020646751403808592, 0.02066236877441406, 0.020700927734375, 0.020684703826904297, 0.020674911499023438, 0.020676607131958007, 0.02064102363586426, 0.02074025535583496, 0.020656736373901367, 0.021345888137817383, 0.020754112243652343, 0.020714208602905272, 0.021070016860961913, 0.02098361587524414, 0.02083430480957031, 0.02077004814147949, 0.020766464233398438, 0.020815008163452147, 0.020731071472167968, 0.020806304931640623, 0.020620288848876952, 0.02067059135437012, 0.02057744026184082, 0.020620800018310546, 0.02083452796936035, 0.02095235252380371, 0.02064627265930176, 0.020801887512207032, 0.020750335693359375, 0.020727455139160158, 0.020674911499023438, 0.02074336051940918, 0.02079155158996582, 0.02084716796875, 0.020736000061035157, 0.02075651168823242, 0.02071945571899414, 0.02062553596496582, 0.020670015335083006, 0.020748735427856446, 0.020679935455322266, 0.020771583557128905, 0.020645280838012696, 0.020669023513793947, 0.020709695816040038, 0.020816095352172853, 0.020520959854125977, 0.020619359970092774, 0.020574111938476563, 0.020596031188964845, 0.020681631088256835, 0.02088118362426758, 0.020633600234985353, 0.020629024505615233, 0.02059107208251953, 0.02065407943725586, 0.020609024047851563, 0.020727807998657227, 0.0206376953125, 0.020715648651123048, 0.020621183395385743, 0.02060054397583008, 0.020879648208618165, 0.020702335357666017, 0.020713632583618163, 0.020615583419799806, 0.0207076473236084, 0.02065580749511719, 0.020682111740112304, 0.020675104141235353, 0.02061552047729492, 0.020784255981445312, 0.020546495437622072, 0.02061516761779785, 0.020619264602661135, 0.02065119934082031, 0.021478208541870117, 0.020673952102661132, 0.020574815750122072, 0.020725568771362304, 0.020566207885742187, 0.020578271865844728, 0.02075651168823242, 0.02066329574584961, 0.020506784439086913, 0.020626272201538086, 0.02059222412109375, 0.020634016036987304, 0.02064588737487793, 0.020633312225341798, 0.020590240478515626, 0.020664960861206054, 0.020686368942260742, 0.020732383728027343, 0.020889312744140624, 0.020748575210571288, 0.020641792297363282, 0.020774911880493165, 0.02063974380493164, 0.020696416854858398, 0.02066703987121582, 0.02071343994140625, 0.02073193550109863, 0.020725759506225586, 0.021323776245117186, 0.020731807708740235, 0.020926559448242187, 0.020727584838867188, 0.020941696166992187, 0.0209136962890625, 0.02063817596435547, 0.02067865562438965, 0.020590272903442383, 0.0206595516204834, 0.020636159896850585, 0.020827775955200196, 0.02055459213256836, 0.020594688415527345, 0.020800607681274414, 0.020808736801147462, 0.020670143127441407, 0.020702751159667968, 0.020721632003784178, 0.020671167373657227, 0.020701183319091796, 0.020762624740600585, 0.02085683250427246, 0.02127257537841797, 0.02107187271118164, 0.021112831115722656, 0.020774911880493165, 0.02068467140197754, 0.020879487991333007, 0.020747392654418946, 0.020759424209594726, 0.020760576248168947, 0.021456735610961914, 0.02203868865966797, 0.02136182403564453, 0.02090070343017578, 0.0208154239654541, 0.02091574478149414, 0.020660543441772462, 0.02063545608520508, 0.020566816329956054, 0.02078643226623535, 0.02131167984008789, 0.022221376419067383, 0.0209039363861084, 0.020841951370239257, 0.020844383239746092, 0.020768543243408204, 0.02071347236633301, 0.020675487518310547, 0.020710752487182616, 0.020646240234375, 0.020977983474731444, 0.020733631134033204, 0.02073334312438965, 0.020804319381713867, 0.02073382377624512, 0.020572128295898436, 0.020723487854003905, 0.02065171241760254, 0.02072822380065918, 0.02072198486328125, 0.02101807975769043, 0.021110687255859375, 0.020613536834716797, 0.020627840042114258]",tokens/s,47.908949248313625,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTJForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3798, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ self.model = DeciCoderModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ self.self_attn = DeciCoderAttention(config=config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ self._init_rope() File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1709, in __getattr__ raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2193.874944,2486.10816,0.0,2107.63776,1984.899072,s,1,8.646224609375,8.646224609375,0.0,8.646224609375,8.646224609375,8.646224609375,8.646224609375,[8.646224609375],,kWh,5.23868538958671e-05,5.771400339708608e-06,1.7365847226002984e-05,7.55241014615787e-05,,MB,2219.184128,2781.806592,0.0,2373.976064,2247.84384,s,10,1.5936590423583983,0.15936590423583982,0.0005326781635457881,0.15929850006103516,0.15965512542724608,0.1601829704284668,0.16060524642944335,"[0.1607108154296875, 0.15953782653808593, 0.1590572509765625, 0.15952720642089843, 0.15938755798339843, 0.15948480224609374, 0.15920944213867189, 0.15853050231933594, 0.15904937744140624, 0.1591642608642578]",tokens/s,1606.3661874697796,kWh,4.751470376276601e-06,5.239947012396316e-07,3.161127080870649e-06,8.436592158386882e-06,tokens/kWh,30344005.63567701,MB,2220.621824,2865.692672,0.0,2457.862144,2341.346816,s,10,15.49423681640625,1.549423681640625,0.01236509743194837,1.5456995849609374,1.5683415161132812,1.5730599792480469,1.5768347497558595,"[1.5434661865234376, 1.5443287353515625, 1.5384366455078125, 1.5373388671875, 1.5470704345703126, 1.547352783203125, 1.5412177734375, 1.5499539794921875, 1.5777784423828125, 1.56729296875]",tokens/s,40.660279526185974,kWh,4.6381450148306314e-05,5.1154417775501975e-06,2.863289835433217e-05,8.012979028018868e-05,tokens/kWh,786224.4463601966,,s,630,15.491415399551387,0.024589548253256176,0.0004737422126479425,0.024426400184631346,0.025178885650634766,0.025357644176483153,0.026356863498687745,"[0.025246559143066408, 0.024980703353881837, 0.025242399215698243, 0.02483305549621582, 0.0245831356048584, 0.02435686492919922, 0.024293535232543944, 0.02456150436401367, 0.024390783309936524, 0.024582719802856444, 0.025866783142089844, 0.025108192443847658, 0.024294496536254883, 0.028855264663696287, 0.024311040878295897, 0.024402687072753906, 0.02434252738952637, 0.024360383987426758, 0.02431648063659668, 0.02434867286682129, 0.024197120666503907, 0.024322080612182616, 0.024321216583251953, 0.024363103866577147, 0.024314111709594726, 0.02508025550842285, 0.024200992584228517, 0.024334560394287108, 0.02439129638671875, 0.02475881576538086, 0.02426358413696289, 0.024367456436157228, 0.024187200546264647, 0.024276479721069336, 0.02409347152709961, 0.02422809600830078, 0.024253215789794922, 0.024331104278564452, 0.024206911087036133, 0.02429142379760742, 0.024275583267211916, 0.02425040054321289, 0.024294240951538086, 0.024199552536010742, 0.024209280014038086, 0.024163040161132812, 0.02421785545349121, 0.024333311080932618, 0.02476214408874512, 0.02470832061767578, 0.024379135131835938, 0.024363008499145508, 0.024289472579956055, 0.024378368377685547, 0.024226015090942382, 0.024404640197753905, 0.024354751586914063, 0.02432598304748535, 0.02416860771179199, 0.024175872802734377, 0.024193504333496093, 0.02426233673095703, 0.024218208312988283, 0.02470889663696289, 0.024913536071777344, 0.024355424880981445, 0.024319072723388672, 0.024273792266845704, 0.024193056106567384, 0.024260608673095704, 0.02430771255493164, 0.026355136871337892, 0.02553094482421875, 0.024473600387573242, 0.024344415664672853, 0.02430521583557129, 0.024253023147583007, 0.024223583221435547, 0.024298784255981445, 0.024170944213867188, 0.024256959915161132, 0.024195072174072265, 0.024557567596435546, 0.02421753692626953, 0.025440319061279297, 0.024346752166748045, 0.024419551849365236, 0.02426128005981445, 0.024420352935791017, 0.024315967559814453, 0.02446329689025879, 0.024358816146850586, 0.024410207748413085, 0.024747648239135743, 0.024295808792114258, 0.024354591369628906, 0.02437763214111328, 0.02426985549926758, 0.02425539207458496, 0.024275136947631837, 0.024423295974731446, 0.024351680755615234, 0.02431590461730957, 0.02454313659667969, 0.02541391944885254, 0.024926015853881836, 0.024374656677246094, 0.02435696029663086, 0.024295167922973634, 0.024400672912597655, 0.024401920318603516, 0.024367103576660155, 0.02434867286682129, 0.02442144012451172, 0.024625503540039062, 0.024443487167358398, 0.024422399520874022, 0.02434662437438965, 0.02434662437438965, 0.024366559982299803, 0.024351232528686522, 0.024363040924072266, 0.024345983505249025, 0.02486534309387207, 0.026291648864746095, 0.02482035255432129, 0.025149375915527343, 0.024415903091430664, 0.02431420707702637, 0.02429955291748047, 0.02474611282348633, 0.0244466552734375, 0.024166528701782226, 0.024339935302734376, 0.024136127471923827, 0.024332416534423827, 0.024182912826538085, 0.02467593574523926, 0.024615232467651366, 0.025097536087036132, 0.024392383575439453, 0.02434662437438965, 0.02425587272644043, 0.024248735427856445, 0.024324607849121094, 0.024461023330688475, 0.024379392623901368, 0.024429567337036134, 0.02445414352416992, 0.0245534725189209, 0.02434867286682129, 0.024426496505737305, 0.02429952049255371, 0.024362016677856445, 0.02429641532897949, 0.024276031494140624, 0.024260671615600585, 0.024255231857299806, 0.024567712783813478, 0.02432841682434082, 0.024442880630493165, 0.024250431060791014, 0.02430988883972168, 0.024248128890991212, 0.024258560180664062, 0.02427903938293457, 0.024190080642700194, 0.02414886474609375, 0.024328351974487305, 0.024237791061401368, 0.0242642879486084, 0.024236576080322265, 0.025944063186645508, 0.02510032081604004, 0.024481311798095703, 0.024275392532348634, 0.02509414482116699, 0.024266752243041992, 0.0245467529296875, 0.024344768524169922, 0.024527040481567383, 0.024225984573364258, 0.02432204818725586, 0.024236192703247072, 0.024282976150512694, 0.02428927993774414, 0.024223743438720705, 0.024262752532958985, 0.024350624084472656, 0.025370624542236327, 0.024645631790161132, 0.024297664642333985, 0.02501817512512207, 0.02478489685058594, 0.02561027145385742, 0.024242143630981445, 0.024387199401855467, 0.024252799987792967, 0.024368991851806642, 0.024156320571899415, 0.024268096923828125, 0.02414252853393555, 0.024193023681640623, 0.024156160354614258, 0.024132863998413086, 0.024048383712768556, 0.0240614070892334, 0.024151935577392578, 0.024394399642944337, 0.024209407806396483, 0.02418217658996582, 0.024140031814575195, 0.024144128799438478, 0.024543039321899413, 0.02436534309387207, 0.024712928771972655, 0.02415817642211914, 0.02441379165649414, 0.024355552673339845, 0.024379392623901368, 0.024393184661865235, 0.024424896240234376, 0.02432419204711914, 0.02442844772338867, 0.02434467124938965, 0.02449612808227539, 0.02439596748352051, 0.02439676856994629, 0.024520959854125977, 0.024545888900756836, 0.024653120040893553, 0.024873023986816407, 0.024525312423706053, 0.024930559158325195, 0.02446112060546875, 0.02452444839477539, 0.024256927490234375, 0.02434016036987305, 0.02430803108215332, 0.024275968551635742, 0.0242653751373291, 0.024227807998657227, 0.024289663314819337, 0.024149984359741212, 0.024317983627319337, 0.024371072769165038, 0.02430169677734375, 0.024164287567138672, 0.02420128059387207, 0.024242176055908202, 0.02443449592590332, 0.024387136459350586, 0.024819616317749024, 0.024572671890258788, 0.024516639709472657, 0.024308000564575195, 0.024233951568603515, 0.02433612823486328, 0.02432601547241211, 0.02439993667602539, 0.02475372886657715, 0.02421811294555664, 0.02428313636779785, 0.024221216201782227, 0.024278656005859375, 0.02440892791748047, 0.024458463668823243, 0.024467679977416994, 0.025401920318603517, 0.024670015335083006, 0.024660160064697265, 0.02459020805358887, 0.024599679946899412, 0.024691551208496094, 0.0247031364440918, 0.024559135437011718, 0.024531423568725588, 0.02453913688659668, 0.024874176025390625, 0.024435712814331056, 0.024383295059204103, 0.02444623947143555, 0.024543584823608397, 0.024672607421875, 0.024671615600585936, 0.024601408004760742, 0.024595903396606444, 0.024658336639404296, 0.02469068717956543, 0.024567615509033202, 0.024674816131591795, 0.02454684829711914, 0.024537248611450194, 0.0245534725189209, 0.024532831192016602, 0.02458844757080078, 0.02448409652709961, 0.02450543975830078, 0.024470176696777344, 0.02459619140625, 0.024407808303833007, 0.024376960754394533, 0.024626272201538086, 0.024595808029174805, 0.024461055755615236, 0.02510041618347168, 0.024369760513305663, 0.024424448013305664, 0.024473600387573242, 0.02442038345336914, 0.024833343505859376, 0.024698944091796876, 0.02449468803405762, 0.024655487060546873, 0.02464192008972168, 0.025536672592163086, 0.024971200942993165, 0.02479315185546875, 0.02454732894897461, 0.024483840942382814, 0.02443878364562988, 0.02454332733154297, 0.02457740783691406, 0.024573631286621093, 0.024515424728393555, 0.024692607879638673, 0.024694175720214845, 0.02470528030395508, 0.024607200622558594, 0.02467148780822754, 0.024597375869750977, 0.024563743591308595, 0.024573631286621093, 0.024537248611450194, 0.024696832656860353, 0.024761888504028322, 0.025350624084472657, 0.02513068771362305, 0.025080095291137694, 0.024858591079711913, 0.024671775817871094, 0.02461955261230469, 0.025792768478393555, 0.0247825927734375, 0.024639488220214844, 0.024534879684448244, 0.02444304084777832, 0.02442630386352539, 0.024500415802001952, 0.024369152069091796, 0.024476959228515626, 0.024541120529174804, 0.024584991455078125, 0.02440575981140137, 0.024338304519653322, 0.024463743209838868, 0.024516511917114257, 0.0244716796875, 0.02452681541442871, 0.024399232864379884, 0.024306047439575196, 0.0242608642578125, 0.024313695907592775, 0.024574111938476563, 0.024700927734375, 0.02412348747253418, 0.024201120376586914, 0.02411667251586914, 0.024334911346435548, 0.024171775817871093, 0.02425916862487793, 0.024252864837646486, 0.024200927734375, 0.024188928604125977, 0.024252384185791016, 0.02418636894226074, 0.024271392822265626, 0.024359968185424803, 0.02487126350402832, 0.024457311630249022, 0.024449024200439453, 0.02427289581298828, 0.024389631271362306, 0.02734195137023926, 0.02472025680541992, 0.024532991409301756, 0.02444825553894043, 0.024443647384643555, 0.02447769546508789, 0.024649728775024415, 0.024551263809204103, 0.024307680130004884, 0.024291519165039063, 0.024276992797851563, 0.02494259262084961, 0.02440608024597168, 0.02461235237121582, 0.024854976654052733, 0.02467862319946289, 0.024991519927978517, 0.024319456100463866, 0.02418332862854004, 0.024120800018310545, 0.024226303100585937, 0.024477535247802735, 0.024207712173461914, 0.024223583221435547, 0.024186784744262696, 0.024238176345825195, 0.024193023681640623, 0.0241940803527832, 0.024247264862060545, 0.024186880111694335, 0.02421526336669922, 0.02421583938598633, 0.024213151931762697, 0.024205663681030273, 0.024167552947998047, 0.02474687957763672, 0.02444697570800781, 0.024252576828002928, 0.024254304885864258, 0.024218687057495115, 0.025066112518310545, 0.02431622314453125, 0.024550943374633788, 0.02436569595336914, 0.024405792236328126, 0.024270912170410157, 0.024377344131469726, 0.024268800735473633, 0.024467647552490233, 0.024198783874511718, 0.024325727462768554, 0.024309600830078125, 0.024326911926269533, 0.024576000213623047, 0.024647680282592774, 0.024285184860229493, 0.024236032485961914, 0.02505449676513672, 0.02641119956970215, 0.02457596778869629, 0.024829727172851562, 0.024139904022216798, 0.024119583129882813, 0.024205408096313476, 0.024174591064453126, 0.024262208938598633, 0.024100351333618163, 0.02424928092956543, 0.02412531280517578, 0.024602752685546875, 0.024165376663208008, 0.024378368377685547, 0.024246271133422852, 0.024389631271362306, 0.024155391693115234, 0.02539289665222168, 0.02426108741760254, 0.024259103775024413, 0.024098848342895506, 0.024203231811523437, 0.024145248413085938, 0.024246944427490234, 0.02410675239562988, 0.02415382385253906, 0.024087072372436524, 0.024180736541748047, 0.024297439575195312, 0.024319520950317385, 0.024280704498291016, 0.024281984329223634, 0.02535148811340332, 0.025165504455566406, 0.024405216217041014, 0.024579872131347658, 0.024223167419433592, 0.02435744094848633, 0.024252416610717774, 0.02434252738952637, 0.024211456298828125, 0.024178688049316405, 0.024305599212646484, 0.02529804801940918, 0.0246824951171875, 0.02462211227416992, 0.02464348793029785, 0.02485862350463867, 0.02485862350463867, 0.024854528427124024, 0.02500387191772461, 0.025178335189819337, 0.026527008056640624, 0.025773792266845702, 0.025045951843261718, 0.025046335220336915, 0.02502931213378906, 0.02500169563293457, 0.02495257568359375, 0.025258207321166994, 0.024747743606567382, 0.02476608085632324, 0.024672672271728514, 0.025593311309814452, 0.024751903533935547, 0.024936607360839844, 0.024850784301757814, 0.024780288696289062, 0.02468534469604492, 0.02543577575683594, 0.025192832946777342, 0.025194368362426757, 0.025610080718994142, 0.025256223678588867, 0.025023679733276367, 0.025128992080688476, 0.02500912094116211, 0.025163583755493164, 0.025203744888305665, 0.0251114559173584, 0.025186368942260742, 0.02514227294921875, 0.025025535583496093, 0.02495052719116211, 0.024944896697998046, 0.024853887557983397, 0.025063072204589844, 0.024977535247802735, 0.02520969581604004, 0.02518383979797363, 0.02514358329772949, 0.025106399536132813, 0.02521913528442383, 0.02521062469482422, 0.02529929542541504, 0.025132543563842775, 0.025138912200927736, 0.025273120880126954, 0.025053184509277345, 0.027054271697998046, 0.026357568740844727, 0.025272319793701172, 0.02539891242980957, 0.02535862350463867, 0.025250175476074218, 0.025337568283081056, 0.02497331237792969, 0.02484547233581543, 0.0258056640625, 0.02461926460266113, 0.024588031768798826, 0.02433228874206543, 0.024162336349487303, 0.024098432540893555, 0.024121408462524415, 0.024106304168701173, 0.024234975814819336, 0.027041791915893554, 0.02483206367492676, 0.02463737678527832, 0.024584192276000977, 0.02450227165222168, 0.024779935836791993, 0.024331104278564452, 0.02427903938293457, 0.024563711166381837, 0.025356447219848633, 0.0245897274017334, 0.024365503311157225, 0.02420265579223633, 0.024189184188842774, 0.024309280395507813, 0.024111648559570313, 0.024133535385131837, 0.024127872467041015, 0.024207040786743163, 0.024034912109375, 0.024121728897094727, 0.02412169647216797, 0.02427449607849121, 0.024228288650512696, 0.024516063690185545, 0.024500127792358398, 0.025204959869384765, 0.024859039306640626, 0.0245382080078125, 0.02447043228149414, 0.024432224273681642, 0.024629663467407227, 0.024705024719238283, 0.024913984298706053, 0.02523257637023926, 0.025219839096069337, 0.025083904266357423, 0.025038080215454103, 0.02522336006164551, 0.025292512893676757, 0.025165727615356445, 0.025095104217529297, 0.025092096328735353, 0.025612287521362305, 0.025286176681518554, 0.02520521545410156, 0.02549065589904785, 0.025444223403930665, 0.02547372817993164, 0.025364736557006835, 0.02517398452758789, 0.02521811294555664, 0.025170879364013674, 0.025108192443847658, 0.025094432830810545, 0.025233407974243165, 0.025078016281127928, 0.024940288543701172, 0.02507366371154785, 0.025114368438720704, 0.025034528732299804, 0.0249749755859375, 0.025289567947387695, 0.025075712203979493, 0.024958400726318358, 0.02485919952392578, 0.02512214469909668, 0.024824480056762695, 0.025102527618408203, 0.024988544464111327, 0.024968000411987306, 0.025106176376342774]",tokens/s,40.66768489199792,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,7394.029568,7808.679936,0.0,7430.209536,7414.23104,s,1,11.6087958984375,11.6087958984375,0.0,11.6087958984375,11.6087958984375,11.6087958984375,11.6087958984375,[11.6087958984375],,kWh,0.00013641785669166968,1.5039798885584721e-05,4.55058697379912e-05,0.0001969635253152456,,MB,1705.480192,8574.140416,0.0,8166.309888,8044.111872,s,10,7.060970458984374,0.7060970458984375,0.0011264210878099625,0.7062753601074219,0.7073542114257813,0.7075199401855469,0.7076525231933594,"[0.7037373046875, 0.7047920532226563, 0.7063265991210937, 0.70622412109375, 0.7055739135742187, 0.7057803955078125, 0.7065865478515625, 0.7073173828125, 0.7076856689453125, 0.7069464721679688]",tokens/s,362.5563957349032,kWh,2.063900946333206e-05,2.2761256128991876e-06,1.3666640562932467e-05,3.658177563916372e-05,tokens/kWh,6998020.06674415,MB,1711.550464,8741.912576,0.0,8334.082048,8265.686528,s,10,31.494411376953124,3.1494411376953124,0.00693604709210387,3.148900146484375,3.1596987304687496,3.16015576171875,3.16052138671875,"[3.13799658203125, 3.140945068359375, 3.146510009765625, 3.147779296875, 3.15002099609375, 3.145820068359375, 3.15088232421875, 3.16061279296875, 3.1542470703125, 3.15959716796875]",tokens/s,20.003548961737994,kWh,9.221209574208425e-05,1.0171450241856433e-05,6.134283611126853e-05,0.0001637263820952092,tokens/kWh,384788.3230166572,,s,630,31.49077796554566,0.04998536185007246,0.0006632363913626596,0.049945600509643556,0.05066940574645996,0.050867183113098145,0.051667712059021,"[0.05168320083618164, 0.049449089050292966, 0.04885289764404297, 0.0493548469543457, 0.0495302734375, 0.04914006423950195, 0.049039070129394534, 0.04963369750976562, 0.0493645133972168, 0.04901686477661133, 0.049473663330078126, 0.04911539077758789, 0.04906383895874023, 0.04924854278564453, 0.04908806228637695, 0.04898409652709961, 0.04977824020385742, 0.04954304122924805, 0.049492416381835935, 0.0499431037902832, 0.04966185760498047, 0.0500777587890625, 0.049903617858886716, 0.05008588790893555, 0.04970454406738281, 0.04927667236328125, 0.049652385711669925, 0.04968243026733398, 0.049604606628417966, 0.049390911102294925, 0.049656513214111325, 0.049495166778564456, 0.04974585723876953, 0.04943369674682617, 0.04935795211791992, 0.04980809783935547, 0.050355358123779295, 0.05007241439819336, 0.04991785430908203, 0.04990512084960937, 0.05016844940185547, 0.050405376434326174, 0.05003059387207031, 0.05042966461181641, 0.05003276824951172, 0.050092193603515626, 0.04993628692626953, 0.049860702514648435, 0.05023110580444336, 0.05003811264038086, 0.050068321228027346, 0.05019443130493164, 0.04966604614257813, 0.05013708877563477, 0.049903617858886716, 0.05021868896484375, 0.05029849624633789, 0.05022895812988281, 0.05016572952270508, 0.050237472534179685, 0.05043503952026367, 0.050751487731933595, 0.05056211090087891, 0.051413311004638675, 0.048828414916992184, 0.0487014389038086, 0.049326080322265625, 0.04962236785888672, 0.049164958953857425, 0.048774784088134765, 0.04921705627441406, 0.0492470703125, 0.04969062423706055, 0.04956774520874024, 0.049743072509765625, 0.04940639877319336, 0.049250656127929685, 0.04896982574462891, 0.04936080169677735, 0.0496978874206543, 0.049466209411621095, 0.049850433349609376, 0.05040304183959961, 0.049602848052978515, 0.04924415969848633, 0.04963328170776367, 0.049186302185058595, 0.04946380615234375, 0.0497336311340332, 0.050283905029296874, 0.04994521713256836, 0.05007155227661133, 0.04970086288452148, 0.05020876693725586, 0.04981939315795898, 0.04973571014404297, 0.0503216323852539, 0.0497437744140625, 0.049973182678222657, 0.04958163070678711, 0.04990012741088867, 0.050135040283203126, 0.050151424407958986, 0.04996860885620117, 0.049838623046875, 0.050157569885253904, 0.05041766357421875, 0.05004083251953125, 0.04992409515380859, 0.049962718963623046, 0.05043228912353516, 0.05007561492919922, 0.05003459167480469, 0.05007782363891602, 0.050067455291748046, 0.05017804718017578, 0.05008377456665039, 0.050085281372070314, 0.05063033676147461, 0.04981196975708008, 0.050105857849121097, 0.05012092971801758, 0.05027065658569336, 0.05068012619018555, 0.050851104736328125, 0.05056790542602539, 0.05162979125976563, 0.05000409698486328, 0.04920950317382813, 0.049186431884765625, 0.04907807922363281, 0.04903308868408203, 0.049737728118896485, 0.04938159942626953, 0.04901020812988281, 0.04925740814208984, 0.04916633605957031, 0.049844223022460936, 0.049870849609375, 0.049565696716308595, 0.050019710540771485, 0.049555103302001954, 0.04916118240356445, 0.04984832000732422, 0.049721343994140625, 0.04975526428222656, 0.05010316848754883, 0.04927644729614258, 0.049721088409423825, 0.04964425659179687, 0.04961280059814453, 0.04936294555664063, 0.049697822570800784, 0.04952163314819336, 0.04970025634765625, 0.05017660903930664, 0.05008793640136719, 0.04959555053710937, 0.04969270324707031, 0.049738559722900394, 0.05038022232055664, 0.050340415954589844, 0.05045033645629883, 0.05035964965820312, 0.050590465545654294, 0.050323455810546876, 0.050014270782470706, 0.04992812728881836, 0.05027123260498047, 0.04977875137329101, 0.04986566543579102, 0.049756160736083986, 0.049796768188476566, 0.05049792098999024, 0.05000188827514648, 0.05035007858276367, 0.04976844787597656, 0.05038035202026367, 0.05016361618041992, 0.05038723373413086, 0.050202880859375, 0.05039228820800781, 0.05042460632324219, 0.05066870498657226, 0.05028268814086914, 0.051010208129882814, 0.05084979248046875, 0.05076287841796875, 0.05012700653076172, 0.0512710075378418, 0.049090496063232424, 0.04951679992675781, 0.04945558547973633, 0.04943814468383789, 0.049021503448486325, 0.04946649551391601, 0.049407905578613284, 0.049207454681396486, 0.04904995346069336, 0.04907260894775391, 0.04974777603149414, 0.04933036804199219, 0.04933603286743164, 0.04995510482788086, 0.04976435089111328, 0.04981145477294922, 0.04953702545166016, 0.05057126235961914, 0.05037583923339844, 0.049504287719726564, 0.04955215835571289, 0.04959958267211914, 0.04964771270751953, 0.04965868759155274, 0.04903071975708008, 0.04960918426513672, 0.04929536056518555, 0.04955955123901367, 0.049377281188964846, 0.050165374755859374, 0.05013900756835937, 0.04975395202636719, 0.049668033599853514, 0.05018454360961914, 0.049858943939208984, 0.05026201629638672, 0.05123891067504883, 0.05153177642822265, 0.049754112243652344, 0.05034393692016602, 0.050206718444824217, 0.04997024154663086, 0.05053740692138672, 0.050261344909667965, 0.049944225311279296, 0.04981622314453125, 0.05009779357910156, 0.04978316879272461, 0.050166110992431644, 0.05036812973022461, 0.0500184326171875, 0.050315521240234376, 0.0504087028503418, 0.05038156890869141, 0.05053849411010742, 0.05050540924072266, 0.05120150375366211, 0.05035452651977539, 0.05086259078979492, 0.05051801681518555, 0.05039888000488281, 0.050487648010253905, 0.052324352264404295, 0.04948918533325195, 0.04959305572509766, 0.04871097564697266, 0.049199806213378904, 0.04910489654541016, 0.048879520416259765, 0.04945110321044922, 0.04939324951171875, 0.049335807800292966, 0.04993040084838867, 0.049849086761474606, 0.049909759521484375, 0.049565696716308595, 0.049239486694335935, 0.049609119415283204, 0.04965353775024414, 0.049692031860351565, 0.05013401412963867, 0.05034947204589844, 0.049850975036621094, 0.050032638549804685, 0.049874942779541014, 0.049838081359863284, 0.04965478515625, 0.04956671905517578, 0.0498603515625, 0.049715297698974606, 0.049645729064941406, 0.050179904937744144, 0.04979321670532227, 0.04965964889526367, 0.050245407104492185, 0.049625152587890624, 0.05020099258422851, 0.05016115188598633, 0.05034649658203125, 0.05062364959716797, 0.051006305694580076, 0.050896385192871096, 0.05030553436279297, 0.04994598388671875, 0.04986124801635742, 0.04945500946044922, 0.049932224273681644, 0.049505695343017575, 0.04991769790649414, 0.05014220809936523, 0.05047235107421875, 0.050067840576171876, 0.050509727478027344, 0.04970323181152344, 0.05035827255249024, 0.05038639831542969, 0.05028668975830078, 0.050316928863525394, 0.05061510467529297, 0.050245311737060545, 0.050704704284667966, 0.05068790435791016, 0.05046691131591797, 0.05086931228637695, 0.05076678466796875, 0.05190063858032227, 0.049793441772460936, 0.04917452621459961, 0.049192958831787106, 0.049698783874511716, 0.049203166961669924, 0.049156158447265626, 0.049235969543457034, 0.04935411071777344, 0.04877961730957031, 0.04930384063720703, 0.04952988815307617, 0.049759105682373045, 0.04952892684936523, 0.04979097747802735, 0.04943772888183594, 0.04947452926635742, 0.04963129425048828, 0.05036435317993164, 0.05003468704223633, 0.04993356704711914, 0.049604736328125, 0.04950080108642578, 0.050116127014160156, 0.049705440521240235, 0.04948096084594727, 0.049388286590576175, 0.049547264099121094, 0.04937932968139649, 0.050132896423339846, 0.04970476913452149, 0.05012713623046875, 0.05027814483642578, 0.049877246856689456, 0.05024739074707031, 0.05021699142456055, 0.05025740814208984, 0.049842945098876955, 0.04978870391845703, 0.05033801651000976, 0.05048115158081055, 0.050333278656005856, 0.050006046295166015, 0.04995663833618164, 0.05006396865844726, 0.05048713684082031, 0.050049312591552736, 0.05020044708251953, 0.050106273651123044, 0.04979926300048828, 0.05048844909667969, 0.049881057739257814, 0.0503276481628418, 0.05059052658081055, 0.05036441421508789, 0.05011215972900391, 0.05012847900390625, 0.05019315338134766, 0.05079859161376953, 0.05048524856567383, 0.05068310546875, 0.04982156753540039, 0.05033667373657227, 0.051605312347412106, 0.04896380615234375, 0.04879833602905274, 0.049352191925048826, 0.049162464141845705, 0.048858463287353514, 0.049103809356689454, 0.049084415435791014, 0.04901852798461914, 0.04932460784912109, 0.04978665542602539, 0.04935270309448242, 0.04948787307739258, 0.04986044692993164, 0.049793182373046876, 0.04946137619018555, 0.050263935089111325, 0.04962876892089844, 0.05004534530639648, 0.050525726318359374, 0.04994486236572265, 0.04952083206176758, 0.04993843078613281, 0.04989257431030274, 0.0495624008178711, 0.0495043830871582, 0.049554847717285154, 0.04959689712524414, 0.04959436798095703, 0.049516544342041016, 0.04962508773803711, 0.05018009567260742, 0.05046886444091797, 0.04995686340332031, 0.05008115386962891, 0.05004352188110352, 0.05002239990234375, 0.05038489532470703, 0.050528160095214845, 0.05065532684326172, 0.05026153564453125, 0.05076425552368164, 0.05024764633178711, 0.050147361755371093, 0.049797119140625, 0.05073100662231445, 0.0502108154296875, 0.05001203155517578, 0.05018342590332031, 0.050582401275634764, 0.05033552169799805, 0.05099708938598633, 0.0503702392578125, 0.050588352203369144, 0.050392894744873046, 0.0503152961730957, 0.049957023620605466, 0.05094790267944336, 0.05131660842895508, 0.050506046295166016, 0.05030912017822266, 0.0510648307800293, 0.05047091293334961, 0.052408065795898434, 0.04942515182495117, 0.04926892852783203, 0.04899225616455078, 0.04948342514038086, 0.049359199523925784, 0.04941619110107422, 0.048947200775146485, 0.0490967025756836, 0.04946739196777344, 0.04978857421875, 0.049746273040771484, 0.0495118408203125, 0.049785438537597655, 0.04974393463134766, 0.04991993713378906, 0.049698432922363284, 0.04990163040161133, 0.05032787322998047, 0.05064908981323242, 0.04998963165283203, 0.04991385650634766, 0.049635326385498044, 0.049430526733398435, 0.04948112106323242, 0.04969123077392578, 0.04928220748901367, 0.0493105583190918, 0.04966998291015625, 0.049825824737548825, 0.050092159271240236, 0.05050281524658203, 0.04983075332641602, 0.04965580749511719, 0.04962428665161133, 0.05804316711425781, 0.04798463821411133, 0.05063065719604492, 0.050865951538085936, 0.050402622222900394, 0.050439071655273435, 0.050539775848388674, 0.050244350433349606, 0.05008118438720703, 0.050098175048828124, 0.05058969497680664, 0.050180606842041016, 0.050018398284912106, 0.05020390319824219, 0.05087433624267578, 0.05069903945922852, 0.05045248031616211, 0.05017910385131836, 0.05162902450561523, 0.051059711456298826, 0.0504060173034668, 0.050356609344482425, 0.05069823837280273, 0.05055692672729492, 0.05040947341918945, 0.0508223991394043, 0.05053094482421875, 0.0504136962890625, 0.05153209686279297, 0.049323936462402344, 0.04909423828125, 0.04907251358032227, 0.04957187271118164, 0.04921286392211914, 0.049113502502441404, 0.04911324691772461, 0.049653759002685545, 0.049375232696533204, 0.048998401641845706, 0.04972857666015625, 0.04980012893676758, 0.04990083312988281, 0.05034467315673828, 0.05004886245727539, 0.049880577087402345, 0.04987152099609375, 0.05073715209960938, 0.05083955383300781, 0.04918272018432617, 0.049307647705078124, 0.04960870361328125, 0.049941856384277346, 0.04980188751220703, 0.04909414291381836, 0.050019840240478515, 0.04952576065063476, 0.04927078247070313, 0.0496448974609375, 0.05005583953857422, 0.04970003128051758, 0.05021974563598633, 0.050708606719970704, 0.05086819076538086, 0.05094604873657226, 0.05078764724731445, 0.05039583969116211, 0.05083955383300781, 0.050427902221679685, 0.0500401611328125, 0.04979369735717774, 0.05041766357421875, 0.0500398063659668, 0.0499681282043457, 0.04983388900756836, 0.049969249725341794, 0.04980121612548828, 0.050480510711669924, 0.050145919799804685, 0.05003984069824219, 0.05050601577758789, 0.05055088043212891, 0.05075209426879883, 0.05067571258544922, 0.05071567916870117, 0.05059804916381836, 0.050629440307617186, 0.050864158630371095, 0.050666942596435546, 0.05044483184814453, 0.05060310363769531, 0.05083843231201172, 0.05221532821655273, 0.04985289764404297, 0.04919091033935547, 0.049172576904296876, 0.04880783843994141, 0.04886483383178711, 0.049252704620361326, 0.04947158432006836, 0.04931958389282227, 0.04891683197021485, 0.04982374572753906, 0.0498172492980957, 0.04980259323120117, 0.04993040084838867, 0.04983280181884766, 0.04987670516967774, 0.05052444839477539, 0.050282497406005856, 0.05050572967529297, 0.05065100860595703, 0.05048297500610351, 0.05008623886108399, 0.04994867324829102, 0.049654975891113284, 0.049722175598144534, 0.05009612655639648, 0.04937907028198242, 0.04958428955078125, 0.04928931045532227, 0.049626590728759766, 0.04940240097045898, 0.05013087844848633, 0.050161727905273436, 0.04994611358642578, 0.05017036819458008, 0.050716670989990234, 0.05066547012329101, 0.05095804977416992, 0.05059104156494141, 0.050522113800048826, 0.05029372787475586, 0.05053849411010742, 0.05005516815185547, 0.05006444931030273, 0.05069670486450195, 0.05029833602905273, 0.05002953720092773, 0.04989033508300781, 0.050113502502441405, 0.05043404769897461, 0.05066854476928711, 0.050648063659667966, 0.05026816177368164, 0.050444446563720706, 0.0508941421508789, 0.05104611206054688, 0.05186191940307617, 0.0512639045715332, 0.05083340835571289, 0.05086566543579102, 0.050049182891845706, 0.05027875137329101, 0.05039308929443359]",tokens/s,20.005856974676483,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,8184.229888,11080.237056,0.0,10701.766656,10468.923392,s,1,13.624958984375,13.624958984375,0.0,13.624958984375,13.624958984375,13.624958984375,13.624958984375,[13.624958984375],,kWh,0.00018848511120831972,2.078398916371034e-05,6.125671567197566e-05,0.00027052581604400573,,MB,3761.123328,11596.136448,0.0,11188.30592,10924.281856,s,10,7.1939533081054705,0.7193953308105469,0.0032105422254021195,0.7190292968750001,0.7231811035156249,0.7239746459960937,0.7246094799804688,"[0.7137408447265625, 0.7159876098632812, 0.7186414794921875, 0.71795703125, 0.7194171142578125, 0.7179682006835938, 0.719479248046875, 0.7230047607421874, 0.7247681884765625, 0.7229888305664063]",tokens/s,355.85440860668825,kWh,2.1005317179166615e-05,2.3152727821531963e-06,1.3990467541568391e-05,3.73110575028882e-05,tokens/kWh,6861236.778941025,MB,3761.123328,11598.2336,0.0,11190.403072,10924.284416,s,10,33.78362280273438,3.378362280273438,0.005775475230882454,3.3771402587890624,3.38703984375,3.387946435546875,3.388671708984375,"[3.37131494140625, 3.374577392578125, 3.373578857421875, 3.3742099609375, 3.372783447265625, 3.379703125, 3.38683837890625, 3.3818642578125, 3.3798994140625, 3.38885302734375]",tokens/s,18.64808885887185,kWh,9.891497501708878e-05,1.0910171633788415e-05,6.56877350898374e-05,0.00017551288174071458,tokens/kWh,358948.0121069973,,s,630,33.78057100677492,0.05361995397900779,0.0008795463550417683,0.053510686874389646,0.054382683181762696,0.054890793991088864,0.058187379913330076,"[0.05852979278564453, 0.05375174331665039, 0.052815937042236326, 0.053262016296386716, 0.05295135879516601, 0.052695072174072266, 0.05292192077636719, 0.05316159820556641, 0.052429214477539066, 0.05278348922729492, 0.052631553649902345, 0.052779006958007815, 0.05315980911254883, 0.05269926452636719, 0.052836353302001954, 0.052799518585205076, 0.053028129577636716, 0.05453689575195313, 0.05502153778076172, 0.05337478256225586, 0.053399776458740236, 0.053217281341552736, 0.052968734741210936, 0.0535497932434082, 0.05337443161010742, 0.05319705581665039, 0.053073566436767576, 0.05286361694335937, 0.052983680725097654, 0.0534749755859375, 0.05310512161254883, 0.053489505767822264, 0.053442718505859375, 0.053101726531982425, 0.054336353302001955, 0.05441948699951172, 0.053721057891845704, 0.054401023864746094, 0.053953536987304686, 0.053082271575927736, 0.05381792068481445, 0.05364559936523437, 0.05308620834350586, 0.053657535552978516, 0.05322963333129883, 0.053093631744384764, 0.053319873809814455, 0.0535107192993164, 0.05328076934814453, 0.05341798400878906, 0.05313740921020508, 0.05333782577514649, 0.05428047943115234, 0.05423513412475586, 0.05366556930541992, 0.054241310119628905, 0.05418963241577148, 0.053674625396728515, 0.05365526580810547, 0.053418270111083986, 0.05428224182128906, 0.053938175201416014, 0.05352243041992188, 0.05855366516113281, 0.05341891098022461, 0.05236684799194336, 0.05272627258300781, 0.0525814094543457, 0.05275875091552734, 0.05267718505859375, 0.05213782501220703, 0.05274454498291015, 0.05313315200805664, 0.05254159927368164, 0.053856254577636715, 0.05335647964477539, 0.05311494445800781, 0.053286369323730466, 0.053236255645751955, 0.05351830291748047, 0.055011009216308596, 0.05365932846069336, 0.05405148696899414, 0.0536220817565918, 0.052472511291503904, 0.05341798400878906, 0.053265567779541015, 0.05269347381591797, 0.052901535034179686, 0.05260355377197266, 0.053316864013671875, 0.05352841567993164, 0.05401699066162109, 0.05358796691894531, 0.053526527404785154, 0.053053279876708985, 0.054020256042480466, 0.05431897735595703, 0.05511999893188477, 0.05434774398803711, 0.05424335861206055, 0.053938175201416014, 0.05300617599487305, 0.053409950256347656, 0.05345280075073242, 0.053017921447753906, 0.05298988723754883, 0.05306867218017578, 0.053653377532958985, 0.053678081512451174, 0.053438720703125, 0.05385776138305664, 0.05369680023193359, 0.05338636779785156, 0.05387142562866211, 0.05435782241821289, 0.05431315231323242, 0.05453420639038086, 0.0539156494140625, 0.05380681610107422, 0.05349814224243164, 0.05382940673828125, 0.05356351852416992, 0.05434729766845703, 0.05338336181640625, 0.053404033660888674, 0.058878143310546874, 0.0534466552734375, 0.05230112075805664, 0.052619552612304686, 0.05253142547607422, 0.05233667373657227, 0.053123233795166015, 0.05260012817382813, 0.05261587142944336, 0.052703231811523435, 0.052794689178466796, 0.053279296875, 0.05377036666870117, 0.05356329727172852, 0.0532050895690918, 0.053433345794677733, 0.054096927642822264, 0.054876129150390626, 0.05483625411987305, 0.053421024322509766, 0.0534365119934082, 0.0529889907836914, 0.053033824920654296, 0.05308528137207031, 0.05338393783569336, 0.05300588989257812, 0.052951648712158204, 0.05299369430541992, 0.05325993728637695, 0.05333647918701172, 0.05329135894775391, 0.05345004653930664, 0.05384041595458984, 0.053346176147460934, 0.053916927337646484, 0.054897087097167965, 0.054067745208740234, 0.05425356674194336, 0.05359427261352539, 0.05310038375854492, 0.05337235260009766, 0.05325267028808594, 0.05290105438232422, 0.05312527847290039, 0.05351065444946289, 0.05305155181884766, 0.05349740982055664, 0.05336489486694336, 0.05323519897460938, 0.05449100875854492, 0.05405129623413086, 0.05355155181884766, 0.054281566619873045, 0.05514473724365234, 0.05439731216430664, 0.05422079849243164, 0.05363663864135742, 0.053141983032226565, 0.05359206390380859, 0.05346928024291992, 0.05306972885131836, 0.05359622573852539, 0.05358774566650391, 0.05815836715698242, 0.05343920135498047, 0.05255299377441406, 0.052900577545166014, 0.05281587219238281, 0.052244094848632815, 0.05264169692993164, 0.05262384033203125, 0.0521973762512207, 0.053053024291992185, 0.05296140670776367, 0.053287200927734375, 0.05372723388671875, 0.05352985763549805, 0.054080257415771486, 0.05379072189331055, 0.05301971054077149, 0.05484844970703125, 0.05471846389770508, 0.05357158279418945, 0.05395455932617187, 0.053376705169677734, 0.0528919677734375, 0.05309846496582031, 0.05238582229614258, 0.05324185562133789, 0.052915775299072265, 0.05280575942993164, 0.05324627304077149, 0.05337686538696289, 0.053338272094726566, 0.0535838737487793, 0.05390681457519531, 0.05380774307250977, 0.05502361679077149, 0.05442502212524414, 0.05351891326904297, 0.053835777282714846, 0.05415935897827148, 0.053465087890625, 0.05338889694213867, 0.05296169662475586, 0.05301248168945313, 0.05294204711914063, 0.053142303466796874, 0.053526527404785154, 0.05405462265014648, 0.053526081085205075, 0.053359264373779296, 0.05363513565063477, 0.05407129669189453, 0.05362073516845703, 0.05422284698486328, 0.0542105598449707, 0.053739105224609375, 0.0549031982421875, 0.05432851028442383, 0.05377312088012695, 0.05324800109863281, 0.05362895965576172, 0.05325791931152344, 0.05323190307617188, 0.05363097763061524, 0.05791929626464844, 0.05348448181152344, 0.05300156784057617, 0.05262790298461914, 0.05257033538818359, 0.052039745330810544, 0.052592575073242186, 0.052959232330322265, 0.05227724838256836, 0.0526192626953125, 0.052678398132324215, 0.05326054382324219, 0.05372313690185547, 0.05350400161743164, 0.05343436813354492, 0.053956607818603515, 0.05410598373413086, 0.05449123382568359, 0.05499878311157227, 0.05396828842163086, 0.05338159942626953, 0.05321564865112305, 0.05256937789916992, 0.053176864624023434, 0.05300857543945312, 0.052784191131591794, 0.052911041259765625, 0.05330940628051758, 0.05285481643676758, 0.05334630584716797, 0.05332582473754883, 0.05326412963867187, 0.053932289123535156, 0.05341299057006836, 0.053942977905273436, 0.054114398956298826, 0.05410620880126953, 0.05396275329589844, 0.05375747299194336, 0.053735904693603516, 0.053354496002197264, 0.05325823974609375, 0.052994049072265625, 0.0530063362121582, 0.053440513610839846, 0.0531599349975586, 0.05341798400878906, 0.053405696868896485, 0.053106689453125, 0.053776382446289066, 0.05411542510986328, 0.05354595184326172, 0.05388076782226563, 0.054269248962402344, 0.05432572937011719, 0.05412681579589844, 0.0538309440612793, 0.05476220703125, 0.05358182525634766, 0.05354620742797851, 0.05313411331176758, 0.05427308654785156, 0.05387760162353516, 0.059109310150146484, 0.05351081466674805, 0.05310844802856445, 0.0527158088684082, 0.05307731246948242, 0.05214278411865234, 0.052631553649902345, 0.05255478286743164, 0.05241955184936523, 0.053343616485595706, 0.05266886520385742, 0.05370694351196289, 0.053337631225585935, 0.0531190071105957, 0.05376822280883789, 0.053518753051757816, 0.05351424026489258, 0.05533875274658203, 0.05542486572265625, 0.054604190826416016, 0.05353420639038086, 0.05328332901000977, 0.052876350402832034, 0.052853694915771486, 0.05273395156860351, 0.05250371170043945, 0.053281280517578126, 0.05320044708251953, 0.05322991943359375, 0.05392428970336914, 0.05354905700683594, 0.053037086486816404, 0.05392585754394531, 0.05394761657714844, 0.05394515228271484, 0.0545167350769043, 0.05504508972167969, 0.05408358383178711, 0.05352447891235351, 0.053370174407958985, 0.05345939254760742, 0.053399520874023436, 0.05303734588623047, 0.0534870719909668, 0.05308265686035156, 0.05291417694091797, 0.053166080474853515, 0.053628158569335935, 0.05340441513061524, 0.05358796691894531, 0.05382758331298828, 0.05382758331298828, 0.05436345672607422, 0.054460990905761716, 0.054628288269042966, 0.05535353469848633, 0.05384806442260742, 0.05399552154541016, 0.053851871490478515, 0.053494049072265626, 0.053679710388183595, 0.05323984146118164, 0.0536657600402832, 0.058199230194091796, 0.05353055953979492, 0.05304204940795899, 0.05260697555541992, 0.052763809204101564, 0.0526032943725586, 0.05305916976928711, 0.05294780731201172, 0.05227724838256836, 0.052776958465576174, 0.053104190826416015, 0.05322774505615235, 0.05361891174316406, 0.05391360092163086, 0.05398739242553711, 0.05386848068237305, 0.05415446472167969, 0.054141727447509766, 0.05487206268310547, 0.05403622436523438, 0.052934913635253905, 0.05329919815063477, 0.05325619125366211, 0.05276633453369141, 0.053014911651611325, 0.05358339309692383, 0.05341136169433594, 0.05299065780639649, 0.05303526306152344, 0.05394598388671875, 0.05393008041381836, 0.05346732711791992, 0.05416707229614258, 0.05420707321166992, 0.05500707244873047, 0.054376575469970705, 0.05428224182128906, 0.05381324768066406, 0.053741600036621096, 0.0542589111328125, 0.05400758361816406, 0.053275646209716795, 0.05359408187866211, 0.05354905700683594, 0.05324710464477539, 0.05369945526123047, 0.05353059387207031, 0.05353254318237305, 0.05379087829589844, 0.05360595321655273, 0.05406355285644531, 0.054639713287353515, 0.05440105438232422, 0.054415519714355466, 0.05456560134887695, 0.05499699020385742, 0.053690433502197266, 0.054298561096191404, 0.05383708953857422, 0.0537033920288086, 0.054296577453613284, 0.05394188690185547, 0.05366412734985351, 0.058063232421875, 0.05331763076782227, 0.05286092758178711, 0.053049343109130856, 0.05269270324707031, 0.05258054351806641, 0.0530228157043457, 0.052948928833007815, 0.05252921676635742, 0.05249587249755859, 0.05313792037963867, 0.05369241714477539, 0.053749759674072264, 0.053749183654785156, 0.053201473236083985, 0.05401190567016602, 0.053835777282714846, 0.054846622467041015, 0.05497494506835938, 0.05412803268432617, 0.05325104141235352, 0.05380300903320313, 0.05343436813354492, 0.05279743957519531, 0.053443904876708984, 0.053091007232666014, 0.05269504165649414, 0.05326438522338867, 0.05288937759399414, 0.0533260498046875, 0.053988544464111325, 0.05351059341430664, 0.05354095840454102, 0.05394870376586914, 0.053718208312988285, 0.05450425720214844, 0.055070240020751955, 0.054303199768066406, 0.054040576934814455, 0.05341958236694336, 0.0534266242980957, 0.05344870376586914, 0.053510143280029294, 0.05348508834838867, 0.053488094329833986, 0.05337071990966797, 0.053045406341552734, 0.05400985717773438, 0.05395455932617187, 0.053508129119873044, 0.053517631530761715, 0.054506145477294925, 0.05460172653198242, 0.05431500625610351, 0.054379680633544925, 0.054903648376464845, 0.05364038467407226, 0.05357241439819336, 0.05325209426879883, 0.0541921272277832, 0.05374156951904297, 0.05361663818359375, 0.053190208435058596, 0.05868153762817383, 0.05326847839355469, 0.0526561279296875, 0.052706302642822264, 0.052854881286621094, 0.05266934585571289, 0.0526965446472168, 0.05307651138305664, 0.05269417572021484, 0.053065631866455076, 0.05317504119873047, 0.05369590377807617, 0.05362489700317383, 0.053410526275634765, 0.05406719970703125, 0.05412659072875976, 0.05361449432373047, 0.05486396789550781, 0.05413273620605469, 0.05371903991699219, 0.05334220886230469, 0.053204544067382814, 0.05276716613769531, 0.052654079437255856, 0.05338252639770508, 0.052904384613037106, 0.05319494247436524, 0.05320908737182617, 0.05298588943481446, 0.053915073394775394, 0.05389081573486328, 0.053956607818603515, 0.05488310241699219, 0.05431283187866211, 0.05366796875, 0.0542138557434082, 0.054032833099365234, 0.05354703903198242, 0.054272319793701174, 0.054013950347900394, 0.05325414276123047, 0.05343231964111328, 0.05340726470947266, 0.05287974548339844, 0.05347132873535156, 0.05343958282470703, 0.053185440063476565, 0.053731040954589845, 0.05356777572631836, 0.053611808776855466, 0.054137569427490234, 0.05404828643798828, 0.05376252746582031, 0.055103488922119144, 0.05453209686279297, 0.05363711929321289, 0.05423276901245117, 0.053778751373291016, 0.0531517448425293, 0.05371680068969727, 0.05360614395141602, 0.05313753509521484, 0.05364358520507812, 0.05971004867553711, 0.05364640045166016, 0.053104705810546875, 0.05276950454711914, 0.05255593490600586, 0.053079425811767576, 0.053241790771484374, 0.05271993637084961, 0.05284688186645508, 0.0527176628112793, 0.05285823822021484, 0.05351897430419922, 0.05378051376342773, 0.05355926513671875, 0.05422694396972656, 0.05380815887451172, 0.05355324935913086, 0.055180160522460935, 0.05505023956298828, 0.05416960144042969, 0.05358796691894531, 0.05344655990600586, 0.05317846298217774, 0.05341593551635742, 0.05334364700317383, 0.05358796691894531, 0.05305615997314453, 0.05288748931884766, 0.05343231964111328, 0.053507102966308596, 0.053730270385742185, 0.05423513412475586, 0.054040576934814455, 0.053610496520996094, 0.05430681610107422, 0.05443916702270508, 0.054136737823486325, 0.054647296905517576, 0.05401430511474609, 0.053661697387695315, 0.05292819213867188, 0.05342444610595703, 0.053515296936035156, 0.05364835357666015, 0.05352447891235351, 0.05315379333496094, 0.05410172653198242, 0.05359439849853516, 0.05410819244384766, 0.053579742431640626, 0.05389267349243164, 0.05517475128173828, 0.053863262176513674, 0.05396275329589844, 0.055100513458251954, 0.05372406387329102, 0.054335487365722655, 0.05394563293457031, 0.053752513885498045, 0.05343356704711914, 0.05438105773925781, 0.05357395172119141, 0.05347635269165039]",tokens/s,18.649773559885933,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,808.792064,537.853952,0.0,159.383552,141.760512,s,1,7.6563134765625,7.6563134765625,0.0,7.6563134765625,7.6563134765625,7.6563134765625,7.6563134765625,[7.6563134765625],,kWh,1.79373740874856e-05,1.9668316973786543e-06,5.639448955985227e-06,2.5543654740849478e-05,,MB,1144.049664,607.059968,0.0,199.22944,184.771584,s,29,0.19300534534454347,0.006655356736018739,0.00011334521046294794,0.006623744010925293,0.006685113525390625,0.006855315113067626,0.007081322154998779,"[0.006968287944793701, 0.006623487949371338, 0.006574975967407227, 0.006549312114715576, 0.006638016223907471, 0.006614880084991455, 0.0066418561935424805, 0.007125279903411865, 0.006634367942810059, 0.006639647960662842, 0.006623744010925293, 0.006684927940368652, 0.006589568138122558, 0.006607135772705078, 0.006613344192504883, 0.006599296092987061, 0.006683296203613282, 0.0066440639495849605, 0.00661407995223999, 0.006594816207885742, 0.006597824096679687, 0.006613344192504883, 0.006680768013000488, 0.00660364818572998, 0.006643487930297852, 0.006685855865478515, 0.006675551891326905, 0.006609504222869873, 0.00663097620010376]",tokens/s,38465.25590649859,kWh,1.978375081271173e-07,2.1818001823691834e-08,1.1384039239050209e-07,3.334959023413112e-07,tokens/kWh,767625623.591623,MB,1154.94912,609.15712,0.0,201.326592,184.774144,s,29,10.026686584472657,0.34574781325767784,0.001936242525375772,0.34593734741210935,0.3482521545410156,0.3495382568359375,0.35064216064453124,"[0.3509805908203125, 0.34616665649414063, 0.34227490234375, 0.34410513305664064, 0.344109375, 0.34430694580078125, 0.34435601806640626, 0.3462296142578125, 0.3433265380859375, 0.34534359741210935, 0.3446510009765625, 0.34593777465820313, 0.34509738159179687, 0.3437035522460937, 0.34317913818359375, 0.3459658203125, 0.344365478515625, 0.3455206909179688, 0.34714877319335935, 0.3462953491210938, 0.3466032409667969, 0.34593734741210935, 0.3460842590332031, 0.3497719116210937, 0.3491877746582031, 0.3480182495117187, 0.34609002685546875, 0.34707467651367185, 0.3448547668457031]",tokens/s,182.2137337801398,kWh,9.914877986556458e-06,1.093428826163856e-06,3.7496588669900057e-06,1.4757965679710319e-05,tokens/kWh,4268881.048193129,,s,1827,10.013619136810306,0.005480908120859497,0.00010531538078863172,0.005462399959564209,0.005541747379302978,0.005588665628433227,0.005881040763854981,"[0.005810080051422119, 0.005703616142272949, 0.005703680038452149, 0.005683296203613282, 0.005652128219604492, 0.0056384320259094236, 0.005557439804077149, 0.005544223785400391, 0.005566976070404053, 0.005506847858428955, 0.005497056007385254, 0.00550710391998291, 0.0056394882202148435, 0.005673344135284424, 0.0055556797981262205, 0.005481279850006103, 0.005451200008392334, 0.005467711925506592, 0.005610496044158936, 0.00550707197189331, 0.005496831893920898, 0.0054651517868042, 0.005448863983154297, 0.005555935859680176, 0.005509183883666992, 0.0056273918151855465, 0.005583360195159912, 0.005475808143615723, 0.005519807815551758, 0.0054988799095153805, 0.0055168957710266114, 0.0054563841819763184, 0.005441343784332276, 0.00553718376159668, 0.005473055839538574, 0.005426688194274902, 0.005541920185089111, 0.005416416168212891, 0.005766143798828125, 0.0054579200744628905, 0.005432384014129639, 0.0054462399482727054, 0.005490752220153808, 0.005454112052917481, 0.005480447769165039, 0.005441535949707031, 0.005932831764221191, 0.005479648113250733, 0.006020095825195313, 0.0065977277755737305, 0.006096992015838623, 0.0054767999649047855, 0.005443808078765869, 0.005546144008636475, 0.005490464210510254, 0.00548905611038208, 0.005468160152435303, 0.005477920055389405, 0.0054848318099975585, 0.005479712009429932, 0.005444320201873779, 0.005472576141357422, 0.005423935890197754, 0.0053311681747436526, 0.005439871788024902, 0.005711775779724121, 0.005418335914611817, 0.005464767932891846, 0.0054429759979248045, 0.005789504051208496, 0.005429535865783691, 0.005415487766265869, 0.005432672023773193, 0.005425920009613037, 0.005413631916046143, 0.005444384098052978, 0.005429471969604492, 0.005399680137634278, 0.00542790412902832, 0.0054048638343811035, 0.005402624130249023, 0.005437727928161621, 0.005434144020080567, 0.00548140811920166, 0.007245535850524902, 0.005472544193267822, 0.005700607776641845, 0.0054241280555725096, 0.005466239929199219, 0.005427072048187256, 0.0055168957710266114, 0.005440127849578857, 0.005453343868255615, 0.005523231983184815, 0.005454559803009033, 0.005506239891052246, 0.005455967903137207, 0.0054268798828125, 0.00553056001663208, 0.005437183856964112, 0.005435679912567139, 0.005508512020111084, 0.005420928001403809, 0.005383776187896728, 0.005436351776123047, 0.005408768177032471, 0.005469503879547119, 0.0054627199172973635, 0.005439328193664551, 0.005402944087982178, 0.005465951919555664, 0.005462016105651855, 0.005470335960388184, 0.005431168079376221, 0.005437024116516113, 0.005433760166168213, 0.005418655872344971, 0.0054148478507995606, 0.005435967922210693, 0.005473663806915283, 0.005423583984375, 0.005472256183624268, 0.005432928085327148, 0.005423200130462646, 0.005400512218475342, 0.005701888084411621, 0.005315839767456055, 0.005440063953399658, 0.005402624130249023, 0.00539244794845581, 0.005534880161285401, 0.005401088237762451, 0.005458176136016846, 0.005422783851623535, 0.00541926383972168, 0.005443520069122314, 0.005419360160827637, 0.005421088218688965, 0.005428544044494629, 0.0054280319213867185, 0.005421055793762207, 0.0054290881156921385, 0.0054290881156921385, 0.005406496047973633, 0.005424960136413574, 0.0054330239295959475, 0.0054301438331604, 0.005406720161437988, 0.005418848037719726, 0.005441696166992188, 0.00542310380935669, 0.005433343887329102, 0.005418240070343017, 0.005421919822692871, 0.005440959930419922, 0.00543177604675293, 0.005419007778167725, 0.005398528099060058, 0.0054609918594360355, 0.005475327968597412, 0.005439295768737793, 0.005418496131896973, 0.005410624027252197, 0.005427231788635254, 0.005392320156097412, 0.005393311977386474, 0.005410816192626953, 0.005404928207397461, 0.0054494719505310055, 0.0054308481216430666, 0.005388735771179199, 0.005433343887329102, 0.005418464183807373, 0.005430975914001465, 0.005373983860015869, 0.005522240161895752, 0.005427040100097656, 0.005455967903137207, 0.0054293122291564945, 0.005396736145019531, 0.005426368236541748, 0.005493311882019043, 0.005399775981903076, 0.005463039875030518, 0.0053901119232177734, 0.00540064001083374, 0.005424895763397217, 0.0054273920059204105, 0.005455359935760498, 0.0053309440612792965, 0.005423295974731446, 0.005474112033843994, 0.005429535865783691, 0.005448416233062744, 0.005427584171295166, 0.0054503679275512695, 0.00539628791809082, 0.005384191989898681, 0.0054254398345947264, 0.005434463977813721, 0.005443520069122314, 0.0054148159027099605, 0.005407392024993896, 0.005439712047576904, 0.005394527912139893, 0.005818367958068848, 0.005437215805053711, 0.005400447845458984, 0.0054126400947570805, 0.005496895790100097, 0.005448192119598389, 0.005429247856140137, 0.005404672145843506, 0.005385663986206055, 0.005432064056396484, 0.005533504009246826, 0.005406367778778076, 0.005482944011688232, 0.005430304050445557, 0.005404672145843506, 0.00559222412109375, 0.005456831932067871, 0.0054505281448364255, 0.005473696231842041, 0.005423647880554199, 0.0054448962211608885, 0.0054135041236877445, 0.005408927917480468, 0.005437695980072021, 0.005500480175018311, 0.005529280185699463, 0.005446144104003906, 0.005446752071380615, 0.0054486079216003415, 0.005430592060089111, 0.005438144207000732, 0.005442944049835205, 0.005439680099487304, 0.005416607856750489, 0.00545462417602539, 0.005472256183624268, 0.005428864002227783, 0.005466495990753174, 0.005492256164550781, 0.005450208187103272, 0.005428351879119873, 0.005718463897705078, 0.005497280120849609, 0.005453824043273926, 0.005441023826599121, 0.005601439952850342, 0.005466464042663574, 0.005371903896331787, 0.005462016105651855, 0.005502975940704345, 0.0054596481323242185, 0.005472799777984619, 0.005477183818817138, 0.005458911895751953, 0.0054635519981384275, 0.005419648170471191, 0.00546127986907959, 0.005454432010650634, 0.005459551811218262, 0.005441952228546143, 0.005449120044708252, 0.005435872077941895, 0.005490496158599854, 0.005471776008605957, 0.005486720085144043, 0.005499711990356445, 0.005427296161651611, 0.005484000205993652, 0.005445600032806397, 0.0054520959854125976, 0.005455039978027344, 0.005459775924682617, 0.005409791946411133, 0.00557260799407959, 0.005485727787017823, 0.005534560203552246, 0.005451776027679443, 0.005438496112823486, 0.005470208168029785, 0.005441535949707031, 0.00543228816986084, 0.005456096172332764, 0.005431072235107422, 0.005418655872344971, 0.005505055904388428, 0.005451935768127442, 0.005463871955871582, 0.005429599761962891, 0.005424831867218018, 0.005455552101135254, 0.005445919990539551, 0.005408671855926514, 0.005422848224639893, 0.005434048175811767, 0.005403679847717285, 0.0054236159324646, 0.005400191783905029, 0.00540553617477417, 0.005412864208221436, 0.005428671836853027, 0.0055487041473388675, 0.0054098558425903324, 0.0053993921279907224, 0.005433343887329102, 0.005475808143615723, 0.0054605121612548825, 0.005562431812286377, 0.005502175807952881, 0.005489696025848389, 0.005455008029937744, 0.005437312126159668, 0.005457856178283691, 0.005652768135070801, 0.005601791858673095, 0.005492224216461182, 0.005468160152435303, 0.005437952041625976, 0.005467199802398682, 0.005464928150177002, 0.005426976203918457, 0.005449759960174561, 0.005429247856140137, 0.005433343887329102, 0.005480415821075439, 0.005420351982116699, 0.005446368217468261, 0.005445631980895996, 0.005442592144012451, 0.00542409610748291, 0.005429247856140137, 0.0054287037849426265, 0.0054297919273376466, 0.005435359954833984, 0.0054287037849426265, 0.005579328060150147, 0.005533376216888428, 0.005458079814910889, 0.005454336166381836, 0.005433055877685547, 0.005429183959960938, 0.005440991878509521, 0.005418879985809326, 0.005470016002655029, 0.005479040145874023, 0.005430784225463867, 0.0055092802047729495, 0.005440127849578857, 0.005494880199432373, 0.00543833589553833, 0.0055428800582885745, 0.0054906878471374515, 0.005462016105651855, 0.0054230718612670896, 0.0054672322273254395, 0.005425312042236328, 0.005467167854309082, 0.005473567962646485, 0.005444191932678223, 0.005445343971252441, 0.005415135860443115, 0.005392320156097412, 0.005428959846496582, 0.005441152095794678, 0.005411488056182861, 0.005471519947052002, 0.0054332160949707034, 0.005482463836669922, 0.005460864067077637, 0.005453343868255615, 0.0054709439277648925, 0.005441504001617432, 0.0054280638694763185, 0.005447840213775635, 0.005362880229949951, 0.005430335998535156, 0.005426784038543701, 0.005457695960998535, 0.005459680080413819, 0.005485343933105469, 0.005416831970214844, 0.005408768177032471, 0.005452991962432862, 0.005452928066253662, 0.005387392044067383, 0.005427775859832763, 0.005498623847961426, 0.005386752128601074, 0.005424032211303711, 0.005401055812835693, 0.005516831874847412, 0.005423967838287354, 0.005425280094146729, 0.005599103927612305, 0.0054347200393676754, 0.005415584087371826, 0.005529248237609864, 0.005418432235717773, 0.005452191829681397, 0.005456543922424316, 0.005416287899017334, 0.005409215927124024, 0.00542742395401001, 0.005439455986022949, 0.005445504188537597, 0.005464064121246338, 0.005470367908477783, 0.0054268798828125, 0.005414239883422852, 0.005420191764831543, 0.005459616184234619, 0.005458047866821289, 0.0054085440635681156, 0.005486432075500488, 0.005480703830718994, 0.005467391967773437, 0.005436192035675049, 0.005459936141967774, 0.005468160152435303, 0.00544268798828125, 0.0054563841819763184, 0.005683584213256836, 0.005474431991577149, 0.005611519813537597, 0.0054967041015625, 0.005543263912200927, 0.005501599788665772, 0.005453887939453125, 0.005450719833374023, 0.005471199989318848, 0.005533696174621582, 0.005480415821075439, 0.005455904006958008, 0.005443456172943115, 0.005494912147521972, 0.005477536201477051, 0.005426015853881836, 0.005427296161651611, 0.005451807975769043, 0.005793312072753906, 0.005507199764251709, 0.005884352207183838, 0.005524735927581787, 0.0054852161407470705, 0.005485727787017823, 0.005534656047821045, 0.005462016105651855, 0.005483551979064942, 0.005454783916473389, 0.005410848140716552, 0.005449120044708252, 0.005425888061523438, 0.005453184127807618, 0.005576799869537354, 0.005458335876464843, 0.00542310380935669, 0.005586944103240967, 0.00545692777633667, 0.005448671817779541, 0.005484543800354004, 0.005484576225280762, 0.005585919857025146, 0.005456736087799072, 0.005429599761962891, 0.005433343887329102, 0.005418784141540527, 0.00543449592590332, 0.005393280029296875, 0.005426368236541748, 0.005961728096008301, 0.005475232124328613, 0.005431295871734619, 0.0054967360496521, 0.005464064121246338, 0.005463679790496826, 0.005469791889190674, 0.005517375946044922, 0.00547708797454834, 0.00545088005065918, 0.005428095817565918, 0.005554207801818848, 0.005421023845672607, 0.005441184043884277, 0.005482848167419434, 0.00561356782913208, 0.005517600059509277, 0.005466976165771484, 0.005462560176849365, 0.005517024040222168, 0.005470880031585693, 0.005422048091888428, 0.005448991775512695, 0.005493887901306152, 0.005489439964294434, 0.005421887874603271, 0.005420224189758301, 0.005447487831115722, 0.005428768157958984, 0.005410624027252197, 0.005478047847747803, 0.005343200206756592, 0.00547382402420044, 0.005461599826812744, 0.005431392192840576, 0.005401311874389648, 0.005435488224029541, 0.005435391902923584, 0.005389311790466309, 0.005430335998535156, 0.005426112174987793, 0.0054057598114013675, 0.005472224235534668, 0.005400544166564942, 0.005406527996063232, 0.005409247875213623, 0.00543123197555542, 0.005429024219512939, 0.005382239818572998, 0.00543120002746582, 0.005445600032806397, 0.005414976119995117, 0.0054412479400634765, 0.005418911933898925, 0.005417312145233154, 0.005407872200012207, 0.005421055793762207, 0.005404928207397461, 0.005404384136199951, 0.005423423767089844, 0.00545472002029419, 0.0054189438819885255, 0.005484511852264404, 0.005416704177856445, 0.005434463977813721, 0.005426144123077393, 0.005439616203308105, 0.005497759819030762, 0.005430240154266358, 0.005404704093933106, 0.005459104061126709, 0.005442368030548096, 0.005482367992401123, 0.005492479801177979, 0.005476672172546387, 0.005484000205993652, 0.005458528041839599, 0.005470208168029785, 0.005486368179321289, 0.005470431804656982, 0.005453824043273926, 0.005478400230407715, 0.005500959873199463, 0.005493984222412109, 0.005438496112823486, 0.005476064205169677, 0.005470208168029785, 0.005420832157135009, 0.005456096172332764, 0.005469823837280273, 0.005474688053131104, 0.005462016105651855, 0.0055227518081665036, 0.005433055877685547, 0.005400576114654541, 0.005502272129058838, 0.005450431823730469, 0.005407936096191406, 0.005468031883239746, 0.005445663928985596, 0.00548960018157959, 0.005506815910339355, 0.005437664031982422, 0.005525504112243652, 0.005443583965301513, 0.005443136215209961, 0.005458367824554443, 0.0054579200744628905, 0.005447360038757324, 0.0054787201881408695, 0.005431295871734619, 0.005513120174407959, 0.0054477758407592775, 0.005429247856140137, 0.00545088005065918, 0.0054362878799438475, 0.0054534077644348145, 0.005474016189575195, 0.005448512077331543, 0.005402495861053467, 0.005451776027679443, 0.00550435209274292, 0.005463712215423584, 0.005507455825805664, 0.0054585919380187985, 0.0054617600440979, 0.00544979190826416, 0.0054438719749450686, 0.005459551811218262, 0.005454048156738282, 0.005415264129638672, 0.005510879993438721, 0.0054694080352783205, 0.005444191932678223, 0.005687488079071045, 0.005631999969482422, 0.005439487934112549, 0.005412576198577881, 0.005427648067474365, 0.0054754562377929685, 0.00542793607711792, 0.005484416007995606, 0.005630080223083496, 0.005443583965301513, 0.005470208168029785, 0.005440927982330322, 0.005448224067687989, 0.00549894380569458, 0.0054884800910949706, 0.005508607864379883, 0.0054967041015625, 0.005552927970886231, 0.005525504112243652, 0.005511168003082275, 0.0054967679977417, 0.005547616004943847, 0.0054966721534729, 0.005396096229553222, 0.005486303806304931, 0.005485311985015869, 0.0055008320808410645, 0.005451551914215088, 0.0054765758514404295, 0.005472320079803467, 0.005470143795013428, 0.005441823959350586, 0.0054617600440979, 0.005433311939239502, 0.005458047866821289, 0.0054967041015625, 0.005502975940704345, 0.005437183856964112, 0.00548035192489624, 0.005448031902313233, 0.005443488121032715, 0.005429344177246094, 0.005428671836853027, 0.005445919990539551, 0.005410208225250244, 0.005417312145233154, 0.00549561595916748, 0.005436543941497803, 0.005397280216217041, 0.005451583862304687, 0.005442624092102051, 0.005431968212127686, 0.00541103982925415, 0.005414303779602051, 0.00543775987625122, 0.005451968193054199, 0.00545308780670166, 0.005426176071166992, 0.00543068790435791, 0.0054280319213867185, 0.005437088012695312, 0.005433343887329102, 0.005418784141540527, 0.005463903903961182, 0.005427872180938721, 0.005475584030151367, 0.005521759986877441, 0.005496960163116455, 0.005477407932281494, 0.005485439777374267, 0.005465727806091308, 0.005482975959777832, 0.005472320079803467, 0.005457119941711426, 0.005472352027893066, 0.005464191913604736, 0.005587647914886475, 0.005518655776977539, 0.005470719814300537, 0.005516960144042969, 0.005525568008422852, 0.0055548157691955565, 0.00552342414855957, 0.0055447998046875, 0.005474815845489502, 0.005455967903137207, 0.0053489279747009275, 0.0054605121612548825, 0.005434783935546875, 0.005430111885070801, 0.005463808059692383, 0.00547215986251831, 0.005399551868438721, 0.005449855804443359, 0.005455008029937744, 0.00548419189453125, 0.0054498882293701175, 0.005433440208435059, 0.005443391799926758, 0.0053944320678710935, 0.005435647964477539, 0.00547814416885376, 0.005447711944580078, 0.005418528079986573, 0.0054464321136474605, 0.005500703811645508, 0.005439360141754151, 0.0054577279090881345, 0.005449920177459717, 0.00540169620513916, 0.005460896015167236, 0.005435200214385987, 0.005437952041625976, 0.0054327998161315914, 0.005602943897247315, 0.005497439861297608, 0.005437632083892822, 0.005420928001403809, 0.00544982385635376, 0.00544547176361084, 0.005453855991363526, 0.005425119876861572, 0.005440703868865967, 0.005491615772247315, 0.005425055980682373, 0.005477727890014649, 0.005452576160430908, 0.005430399894714355, 0.005415552139282227, 0.005442944049835205, 0.005430016040802002, 0.006418432235717773, 0.00603875207901001, 0.006087456226348877, 0.00546611213684082, 0.005490111827850342, 0.005483168125152588, 0.005461343765258789, 0.005482272148132324, 0.005456672191619873, 0.005431295871734619, 0.005457791805267334, 0.0054432001113891605, 0.005435904026031494, 0.005453824043273926, 0.0054858880043029785, 0.005425568103790283, 0.005455455780029297, 0.00541971206665039, 0.005344255924224854, 0.005522047996520996, 0.005443967819213867, 0.005414912223815918, 0.005465919971466064, 0.005503168106079101, 0.005512256145477295, 0.005461152076721191, 0.005430431842803955, 0.0055558719635009765, 0.005455872058868408, 0.005423840045928955, 0.005450016021728516, 0.005464032173156739, 0.005472256183624268, 0.00551526403427124, 0.005453983783721924, 0.005481344223022461, 0.005749728202819824, 0.0054568958282470706, 0.005471231937408447, 0.005449376106262207, 0.005453216075897217, 0.005485055923461914, 0.005489088058471679, 0.005558176040649414, 0.005467775821685791, 0.005506815910339355, 0.005478271961212158, 0.005454559803009033, 0.005446976184844971, 0.005579584121704102, 0.00547430419921875, 0.0054778242111206054, 0.005460544109344483, 0.005479936122894287, 0.0054540162086486816, 0.00543503999710083, 0.005421440124511719, 0.005439008235931396, 0.005455967903137207, 0.00542300796508789, 0.005455935955047607, 0.005442463874816895, 0.005431072235107422, 0.005494175910949707, 0.005427807807922363, 0.005461056232452393, 0.005438399791717529, 0.005633312225341797, 0.005464799880981445, 0.005433343887329102, 0.005410816192626953, 0.005457471847534179, 0.005437727928161621, 0.005494944095611572, 0.005447679996490478, 0.005510335922241211, 0.005440320014953613, 0.005435679912567139, 0.005495808124542236, 0.00545251178741455, 0.005451551914215088, 0.005399648189544678, 0.005493120193481445, 0.005435935974121094, 0.005428575992584229, 0.005468832015991211, 0.005443583965301513, 0.0054287037849426265, 0.005478432178497314, 0.005425663948059082, 0.00545084810256958, 0.0054115839004516605, 0.005456319808959961, 0.005425983905792236, 0.005428127765655518, 0.005431295871734619, 0.005433152198791504, 0.00543775987625122, 0.005420928001403809, 0.005441120147705078, 0.005448448181152344, 0.005426623821258545, 0.005490528106689453, 0.005462399959564209, 0.005537792205810547, 0.005439487934112549, 0.00542310380935669, 0.005476352214813233, 0.005441088199615479, 0.005441760063171387, 0.005547743797302246, 0.0054297599792480465, 0.005479936122894287, 0.00541107177734375, 0.00544927978515625, 0.005446591854095459, 0.005426943778991699, 0.005539167881011963, 0.005458720207214355, 0.005440896034240723, 0.005474016189575195, 0.005466623783111572, 0.0054841599464416505, 0.005457632064819336, 0.005434304237365723, 0.005420896053314209, 0.005442016124725342, 0.005437119960784912, 0.005412864208221436, 0.005435455799102784, 0.005429376125335694, 0.005409759998321533, 0.005486976146697998, 0.005439807891845703, 0.005409088134765625, 0.005445312023162842, 0.005508319854736328, 0.005437952041625976, 0.00543174409866333, 0.005492735862731934, 0.005445631980895996, 0.005419007778167725, 0.005459968090057373, 0.005424479961395264, 0.005348351955413819, 0.005430016040802002, 0.005439712047576904, 0.0054570879936218265, 0.005426015853881836, 0.0054390721321105956, 0.00543555212020874, 0.005476448059082031, 0.005447840213775635, 0.005447679996490478, 0.0054700798988342285, 0.005496352195739746, 0.005465951919555664, 0.005501567840576172, 0.005445759773254394, 0.005425151824951172, 0.0054291200637817385, 0.005588575839996338, 0.005442080020904541, 0.005451744079589844, 0.005406752109527588, 0.005418272018432617, 0.0054134721755981444, 0.005418528079986573, 0.005459743976593018, 0.005417183876037597, 0.005397088050842285, 0.005486911773681641, 0.005451583862304687, 0.0054065918922424315, 0.005449728012084961, 0.005435391902923584, 0.005437439918518067, 0.005449728012084961, 0.005437439918518067, 0.005439487934112549, 0.005445312023162842, 0.005496831893920898, 0.005443071842193604, 0.005432127952575683, 0.005390336036682129, 0.005431295871734619, 0.00540067195892334, 0.005400479793548584, 0.005425151824951172, 0.005426591873168946, 0.005395008087158203, 0.005423136234283447, 0.005486495971679687, 0.0054068160057067875, 0.005436992168426514, 0.0055136637687683105, 0.005414463996887207, 0.005417568206787109, 0.005418303966522217, 0.0055280961990356444, 0.005400000095367431, 0.005412960052490234, 0.00546457576751709, 0.00543503999710083, 0.005414463996887207, 0.005444352149963379, 0.005447648048400879, 0.0053673601150512696, 0.005465792179107666, 0.005473023891448975, 0.005504096031188965, 0.005477248191833496, 0.00544700813293457, 0.005521312236785889, 0.005469183921813964, 0.005486368179321289, 0.005502975940704345, 0.005510528087615967, 0.005467008113861084, 0.005543680191040039, 0.0055316481590271, 0.005483935832977295, 0.005510047912597656, 0.005497856140136719, 0.005521312236785889, 0.0055368318557739256, 0.005479743957519531, 0.00549724817276001, 0.00558406400680542, 0.005535711765289306, 0.005538368225097656, 0.005473567962646485, 0.005487872123718262, 0.005488383769989014, 0.0054943361282348635, 0.005515711784362793, 0.0054653120040893555, 0.005491136074066162, 0.0055055999755859375, 0.005473120212554932, 0.005548992156982422, 0.005494080066680908, 0.005468863964080811, 0.005479743957519531, 0.005503680229187012, 0.0055066561698913575, 0.005468575954437256, 0.005471776008605957, 0.005491007804870605, 0.005482880115509034, 0.005500703811645508, 0.005482656002044678, 0.0054925761222839355, 0.005423359870910645, 0.005476096153259277, 0.005469696044921875, 0.0054683837890625, 0.005509151935577393, 0.005431136131286621, 0.005513631820678711, 0.00547814416885376, 0.00542464017868042, 0.005441664218902588, 0.005452767848968506, 0.005461664199829102, 0.005455840110778809, 0.005453311920166016, 0.0054496641159057614, 0.005450111865997314, 0.005423327922821045, 0.005339200019836426, 0.005488639831542969, 0.005459328174591064, 0.005484672069549561, 0.005458432197570801, 0.005445631980895996, 0.005490623950958252, 0.005463488101959228, 0.005450272083282471, 0.005540031909942627, 0.005481919765472412, 0.005468255996704101, 0.005466495990753174, 0.005535679817199707, 0.005459392070770263, 0.005435776233673095, 0.005440832138061524, 0.005444543838500977, 0.00545084810256958, 0.005605823993682861, 0.005470047950744629, 0.0054765758514404295, 0.005466784000396728, 0.005452640056610107, 0.005424032211303711, 0.005451104164123535, 0.005423776149749756, 0.0054254717826843265, 0.005465792179107666, 0.005480127811431885, 0.0054152321815490725, 0.005449151992797852, 0.005433792114257812, 0.005490816116333007, 0.005533696174621582, 0.005431295871734619, 0.005433343887329102, 0.00544758415222168, 0.005421152114868164, 0.005489664077758789, 0.005429503917694092, 0.005403391838073731, 0.005436448097229004, 0.00543228816986084, 0.005453343868255615, 0.005433087825775147, 0.005440224170684815, 0.00544595193862915, 0.005502719879150391, 0.005413887977600097, 0.005571135997772217, 0.005446208000183105, 0.005415775775909424, 0.00548144006729126, 0.005455455780029297, 0.005506976127624512, 0.005443071842193604, 0.005419648170471191, 0.005467552185058594, 0.005459936141967774, 0.005475327968597412, 0.005453824043273926, 0.005441664218902588, 0.005399295806884766, 0.005486303806304931, 0.005484064102172851, 0.0054583997726440426, 0.005480447769165039, 0.005500895977020264, 0.005496863842010498, 0.0054988799095153805, 0.0054617919921875, 0.005490592002868652, 0.005479072093963623, 0.0054505281448364255, 0.005481056213378906, 0.005466400146484375, 0.005449728012084961, 0.005496831893920898, 0.00546611213684082, 0.00547430419921875, 0.005464064121246338, 0.005462016105651855, 0.005451935768127442, 0.0054642558097839355, 0.005452703952789306, 0.005485055923461914, 0.005475808143615723, 0.005481535911560058, 0.005486239910125732, 0.0054618239402770994, 0.005482272148132324, 0.005478879928588867, 0.005435455799102784, 0.005486527919769287, 0.0054579200744628905, 0.005453824043273926, 0.005496831893920898, 0.005552127838134766, 0.005484672069549561, 0.005465439796447754, 0.005454559803009033, 0.005452864170074463, 0.005530303955078125, 0.005451488018035889, 0.005482848167419434, 0.005466239929199219, 0.005623807907104492, 0.005535615921020508, 0.0054633598327636715, 0.0056121921539306644, 0.005516384124755859, 0.005438399791717529, 0.005480447769165039, 0.005461535930633545, 0.005435872077941895, 0.005480447769165039, 0.005448991775512695, 0.005508031845092774, 0.005439455986022949, 0.005456960201263428, 0.00544982385635376, 0.005528223991394043, 0.005470208168029785, 0.00546611213684082, 0.005419007778167725, 0.005357664108276367, 0.005523007869720459, 0.00561023998260498, 0.005595136165618897, 0.005591040134429932, 0.005443583965301513, 0.005459743976593018, 0.005469855785369873, 0.005502912044525146, 0.005464704036712647, 0.005440544128417968, 0.0055775361061096195, 0.005485983848571777, 0.0055838079452514644, 0.005506527900695801, 0.00547430419921875, 0.005544288158416748, 0.005490623950958252, 0.005468224048614502, 0.0054579200744628905, 0.005472256183624268, 0.0055270719528198245, 0.0054380159378051756, 0.005463967800140381, 0.005468160152435303, 0.005447936058044433, 0.00647321605682373, 0.005478367805480957, 0.005509247779846191, 0.005677248001098633, 0.005510464191436768, 0.005556159973144531, 0.005501535892486573, 0.0054776320457458495, 0.00544652795791626, 0.005486591815948487, 0.0054579200744628905, 0.005453824043273926, 0.005455872058868408, 0.005449728012084961, 0.005543935775756836, 0.005502848148345947, 0.005441152095794678, 0.005544447898864746, 0.005494431972503662, 0.00545801591873169, 0.005467679977416992, 0.005481184005737305, 0.005449728012084961, 0.0054620480537414555, 0.005476128101348877, 0.005601056098937988, 0.005439871788024902, 0.005426911830902099, 0.005453887939453125, 0.005459296226501465, 0.005442463874816895, 0.0054774398803710935, 0.0054486398696899415, 0.00547430419921875, 0.005428991794586182, 0.005490719795227051, 0.005439712047576904, 0.005359327793121338, 0.005448671817779541, 0.005445631980895996, 0.005515391826629639, 0.005437439918518067, 0.00545088005065918, 0.005468992233276368, 0.0054783358573913575, 0.005512191772460938, 0.005529952049255371, 0.005470592021942139, 0.005493023872375488, 0.0055131521224975584, 0.005483712196350098, 0.005485439777374267, 0.005487711906433106, 0.005594048023223877, 0.0058715839385986325, 0.005502463817596436, 0.00547382402420044, 0.0054568958282470706, 0.005475808143615723, 0.005468671798706055, 0.005432576179504395, 0.005493631839752197, 0.005475391864776611, 0.0054915199279785155, 0.005567647933959961, 0.0054915518760681155, 0.005459328174591064, 0.005466752052307129, 0.005477503776550293, 0.005474239826202393, 0.005456768035888672, 0.00546124792098999, 0.00548528003692627, 0.005459743976593018, 0.00555244779586792, 0.005556032180786132, 0.0054540162086486816, 0.005471456050872803, 0.005538591861724854, 0.005455872058868408, 0.005486591815948487, 0.005457632064819336, 0.005482751846313477, 0.005482783794403076, 0.005515039920806885, 0.005480415821075439, 0.0054939198493957515, 0.005460576057434082, 0.005468416213989258, 0.005465888023376465, 0.005483007907867431, 0.005518496036529541, 0.005488927841186523, 0.005508512020111084, 0.005534592151641845, 0.0054704642295837404, 0.005463840007781983, 0.005459936141967774, 0.0054759359359741215, 0.005491104125976563, 0.00543503999710083, 0.005458720207214355, 0.005502111911773681, 0.005514080047607422, 0.005447679996490478, 0.0054988799095153805, 0.0054701118469238285, 0.005461311817169189, 0.005471007823944092, 0.005442751884460449, 0.005481215953826904, 0.005445536136627197, 0.005464064121246338, 0.00550108814239502, 0.005455743789672851, 0.005449855804443359, 0.005485663890838623, 0.005477280139923096, 0.005478432178497314, 0.0054906558990478515, 0.005428319931030274, 0.005485472202301026, 0.005441184043884277, 0.00549513578414917, 0.005461599826812744, 0.005472320079803467, 0.005457568168640137, 0.005855936050415039, 0.005588992118835449, 0.005609407901763916, 0.005930751800537109, 0.006117695808410644, 0.005545504093170166, 0.005538464069366455, 0.005476223945617676, 0.005437376022338868, 0.005467360019683838, 0.00555452823638916, 0.0054382081031799315, 0.0054616317749023435, 0.005435328006744385, 0.005419392108917237, 0.0055129599571228025, 0.005463935852050781, 0.0054860801696777345, 0.005466752052307129, 0.005433504104614258, 0.005440896034240723, 0.005423327922821045, 0.005414559841156006, 0.005455743789672851, 0.005588863849639893, 0.005507487773895264, 0.0054349441528320315, 0.005403520107269287, 0.005441760063171387, 0.005410624027252197, 0.0053976960182189945, 0.005438240051269532, 0.005433343887329102, 0.005422976016998291, 0.005451903820037842, 0.005537568092346191, 0.005482367992401123, 0.005471903800964356, 0.0054481601715087894, 0.005416959762573242, 0.005441120147705078, 0.0054230718612670896, 0.005435840129852295, 0.005445631980895996, 0.005468160152435303, 0.005449408054351806, 0.005427840232849121, 0.005406400203704834, 0.0054271998405456545, 0.005421055793762207, 0.005449056148529053, 0.0054217281341552735, 0.005430943965911865, 0.0054316802024841305, 0.005494688034057617, 0.005431359767913819, 0.005439680099487304, 0.005474080085754394, 0.005482528209686279, 0.005482143878936767, 0.005478752136230469, 0.005486368179321289, 0.005544159889221191, 0.00550707197189331, 0.005549280166625976, 0.005486847877502442, 0.0055342397689819335, 0.005482495784759522, 0.005459968090057373, 0.005432320117950439, 0.0054876160621643065, 0.00557260799407959, 0.005464000225067139, 0.005482463836669922, 0.005504640102386474, 0.005488448143005371, 0.005541632175445557, 0.005482592105865478, 0.005515647888183594, 0.005513631820678711, 0.005496128082275391, 0.005515999794006347, 0.005586944103240967, 0.005569600105285645, 0.005579520225524903, 0.005482687950134278, 0.005495071887969971, 0.005471968173980713, 0.005486591815948487, 0.005488383769989014, 0.005452032089233398, 0.005507199764251709, 0.005494495868682861, 0.005546271800994873, 0.00553107213973999, 0.0055136637687683105, 0.005492767810821533, 0.005499872207641601, 0.005555391788482666, 0.005423903942108154, 0.005564576148986817, 0.0055188159942626954, 0.005535583972930908, 0.005517343997955323, 0.005449920177459717, 0.005489280223846435, 0.005513055801391602, 0.005464064121246338, 0.00547430419921875, 0.005466400146484375, 0.005489471912384033, 0.005501215934753418, 0.0054462399482727054, 0.005561952114105225, 0.005482912063598633, 0.0054579200744628905, 0.005492479801177979, 0.005520703792572021, 0.005509856224060058, 0.005469888210296631, 0.00546668815612793, 0.005431327819824219, 0.0054514241218566895, 0.005490880012512207, 0.005470335960388184, 0.00542303991317749, 0.0055337600708007815, 0.005486720085144043, 0.00544163179397583, 0.00546995210647583, 0.005543807983398438, 0.0054457921981811526, 0.005484543800354004, 0.005512576103210449, 0.0054791679382324215, 0.0055409598350524905, 0.005537759780883789, 0.005505248069763184, 0.005494688034057617, 0.005470143795013428, 0.005503744125366211, 0.005476352214813233, 0.0054988799095153805, 0.005771423816680908, 0.005498720169067383, 0.005496831893920898, 0.005478464126586914, 0.005457856178283691, 0.0054925761222839355, 0.005470143795013428, 0.00543942403793335, 0.005472544193267822, 0.005443583965301513, 0.005523744106292724, 0.00544927978515625, 0.005451615810394287, 0.0054702720642089845, 0.005445695877075195, 0.005449920177459717, 0.0054330239295959475, 0.0054336638450622554, 0.005420735836029053, 0.005402463912963867, 0.005660831928253174, 0.005435391902923584, 0.005609471797943116, 0.005512479782104492, 0.005425888061523438, 0.005459968090057373, 0.005431295871734619, 0.005457376003265381, 0.005440192222595215, 0.0055195198059082035, 0.005482175827026367, 0.005491936206817627, 0.005409823894500732, 0.0054757437705993655, 0.005462272167205811, 0.00543939208984375, 0.005418848037719726, 0.005458271980285644, 0.005549088001251221, 0.005647295951843261, 0.005476384162902832, 0.00561356782913208, 0.005664896011352539, 0.005475903987884521, 0.005484864234924316, 0.005456128120422363, 0.005496895790100097, 0.005580480098724365, 0.005421216011047363, 0.005513055801391602, 0.005474368095397949, 0.005642176151275635, 0.005484543800354004, 0.005609471797943116, 0.005462016105651855, 0.0054579200744628905, 0.005458047866821289, 0.005469183921813964, 0.005487296104431152, 0.00553388786315918, 0.005461696147918701, 0.006893472194671631, 0.006023583889007568, 0.005584320068359375, 0.005554080009460449, 0.005498623847961426, 0.005490623950958252, 0.0054651198387145996, 0.005511104106903076, 0.005564159870147705, 0.005511712074279785, 0.005537824153900146, 0.005648223876953125, 0.0055560641288757324, 0.005566463947296142, 0.005536799907684326, 0.0055502080917358395, 0.005729055881500244, 0.005595200061798096, 0.0055316481590271, 0.0055092802047729495, 0.005547264099121094, 0.005478400230407715, 0.005512639999389648, 0.005472832202911377, 0.005492735862731934, 0.006676000118255615, 0.005593567848205566, 0.005681568145751953, 0.0055701441764831544, 0.005533472061157226, 0.0054991040229797365, 0.0055409278869628905, 0.005529568195343017, 0.005626431941986084, 0.005542304039001465, 0.005513504028320313, 0.005582560062408448, 0.005509119987487793, 0.005545983791351319, 0.005603328227996827, 0.005479712009429932, 0.005550816059112549, 0.005545983791351319, 0.005498112201690674, 0.005517951965332031, 0.005637631893157959, 0.005496543884277344, 0.005567488193511963, 0.005490592002868652, 0.005505119800567627, 0.005508831977844238, 0.005628032207489014, 0.005490943908691406, 0.005469503879547119, 0.005532159805297852, 0.005588704109191894, 0.005484543800354004, 0.0055298562049865725, 0.005533055782318115, 0.005495456218719483, 0.005455552101135254, 0.005443903923034668, 0.0054988799095153805, 0.005500895977020264, 0.005455552101135254, 0.00553772783279419, 0.0055689277648925785, 0.005474527835845947, 0.00545366382598877, 0.0055725440979003905, 0.005473696231842041, 0.005544352054595947, 0.005449920177459717, 0.005455872058868408, 0.005478400230407715, 0.005535744190216065, 0.005472256183624268, 0.005429247856140137, 0.00546611213684082, 0.00547430419921875, 0.005453824043273926, 0.005502143859863281, 0.00548144006729126, 0.005502016067504883, 0.0055680317878723145, 0.005552608013153076, 0.005683199882507324, 0.005518943786621094, 0.005501152038574219, 0.006435103893280029, 0.005646431922912597, 0.005559360027313233, 0.0054952001571655276, 0.005488992214202881, 0.005525279998779297, 0.005474815845489502, 0.0055025601387023925, 0.005476191997528076, 0.0054479680061340335, 0.005555712223052979, 0.005468671798706055, 0.005484255790710449, 0.005474815845489502, 0.005463903903961182, 0.005538943767547607, 0.00555075216293335, 0.0054570879936218265, 0.005478623867034912, 0.005484960079193115, 0.0054603199958801266, 0.005505375862121582, 0.0055149121284484865, 0.0055419840812683105, 0.005525407791137696, 0.005517024040222168, 0.005515647888183594, 0.005482399940490722, 0.005453536033630371, 0.005467840194702148, 0.005474912166595459, 0.005555359840393066, 0.005503392219543457, 0.005470367908477783, 0.005458208084106445, 0.005519360065460205, 0.005464159965515137, 0.005486688137054443, 0.005463871955871582, 0.005471776008605957, 0.005466591835021972, 0.005434624195098877, 0.005476384162902832, 0.005466847896575928, 0.0054551677703857425, 0.005509024143218994, 0.005471231937408447, 0.0054534077644348145, 0.005468639850616455, 0.005465824127197266, 0.0055095357894897464, 0.005493887901306152, 0.005518879890441895, 0.005655168056488037, 0.0055196800231933595, 0.005508543968200684, 0.0055055999755859375, 0.0054963197708129885, 0.005477888107299805, 0.00548854398727417, 0.005461664199829102, 0.005469120025634765, 0.005514976024627686, 0.005441215991973877, 0.005519328117370606, 0.005547967910766602, 0.005459712028503418, 0.005471168041229248, 0.005442624092102051, 0.005823423862457275, 0.005477536201477051, 0.005436255931854248, 0.005453824043273926, 0.005478400230407715, 0.005449312210083008, 0.005564832210540772, 0.005445343971252441, 0.005435232162475586, 0.005466559886932373, 0.005467584133148193, 0.00546070384979248, 0.005464032173156739, 0.005474175930023193, 0.005462016105651855, 0.005488639831542969, 0.005427008152008057, 0.005480639934539795, 0.005564159870147705, 0.005478655815124512, 0.005463263988494873, 0.005437503814697266, 0.005477248191833496, 0.005467423915863037, 0.005478047847747803, 0.005546688079833984, 0.005478879928588867, 0.0055064959526062015, 0.0055032958984375, 0.0054988799095153805, 0.005492735862731934, 0.005568511962890625, 0.005494783878326416, 0.005523776054382324, 0.005476223945617676, 0.005556128025054932, 0.005515391826629639, 0.005520959854125976, 0.005483903884887695, 0.005464928150177002, 0.0054581441879272465, 0.005469183921813964, 0.0054559998512268065, 0.005450399875640869, 0.005474016189575195, 0.005504767894744873, 0.0054973759651184085, 0.0054514241218566895, 0.00548089599609375, 0.005463967800140381, 0.00544326400756836, 0.005427072048187256, 0.005386847972869873, 0.005539423942565918, 0.0055433278083801265, 0.005487679958343506, 0.005463935852050781, 0.005408736228942871, 0.005464096069335938, 0.005515488147735595, 0.0054617919921875, 0.005486591815948487, 0.005449728012084961, 0.005484799861907959, 0.005447264194488525, 0.00542742395401001, 0.005457215785980225, 0.005497471809387207, 0.005443583965301513, 0.005522848129272461, 0.005534304141998291, 0.005467775821685791, 0.005456255912780762, 0.005482495784759522, 0.005451839923858642, 0.005455808162689209, 0.005439648151397705, 0.0054618558883666995, 0.005462016105651855, 0.005449984073638916, 0.005467520236968994, 0.0054316802024841305, 0.0054579200744628905, 0.005509312152862549, 0.005448991775512695, 0.0054605121612548825, 0.0055212478637695316, 0.00544159984588623, 0.005453919887542725, 0.005451935768127442, 0.005451615810394287, 0.00545356798171997, 0.005531904220581055, 0.005525279998779297, 0.0054683837890625, 0.005447679996490478, 0.005472256183624268, 0.005483871936798096, 0.00544220781326294, 0.00547430419921875, 0.005459199905395508, 0.005479296207427978, 0.005435264110565186, 0.005425151824951172, 0.0054941439628601076, 0.005442272186279297, 0.005617280006408692, 0.0054848318099975585, 0.005504255771636963, 0.006189824104309082, 0.005599232196807862, 0.0055353279113769535, 0.006015391826629639, 0.0058716158866882326, 0.005513055801391602, 0.00540499210357666, 0.005467743873596192, 0.005525279998779297, 0.005456768035888672, 0.005553055763244629, 0.0055, 0.005590176105499268, 0.005553855895996094, 0.005466239929199219, 0.005447904109954834, 0.005460095882415771, 0.005500448226928711, 0.005452896118164062, 0.005463808059692383, 0.005460031986236572, 0.005482495784759522, 0.005503200054168701, 0.0054802241325378415, 0.005461120128631592, 0.005470560073852539, 0.0054349441528320315, 0.005497824192047119, 0.005453536033630371, 0.005467872142791748, 0.005444159984588623, 0.005431456089019775, 0.005448768138885498, 0.005499584197998047, 0.005425248146057129, 0.005476352214813233, 0.005424479961395264, 0.005421472072601318, 0.005426976203918457, 0.005452479839324951, 0.0055677762031555176, 0.005425151824951172, 0.005431807994842529, 0.005514304161071778, 0.0054501757621765136, 0.005435935974121094, 0.005412831783294678, 0.005435391902923584, 0.005438560009002685, 0.005482751846313477, 0.005466720104217529, 0.005644351959228516, 0.005453824043273926, 0.005478400230407715, 0.005441535949707031, 0.005451776027679443, 0.005421055793762207, 0.005491936206817627, 0.005432096004486084, 0.005418240070343017, 0.0054382081031799315, 0.005456992149353027, 0.005437439918518067, 0.005464992046356201, 0.005437312126159668, 0.005470335960388184, 0.005449728012084961, 0.005484384059906006, 0.005469888210296631]",tokens/s,182.4515167831683,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: DeciLMForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 205, in run_text_generation_memory_tracking _ = backend.generate(self.inputs, self.config.generate_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 430, in generate return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 610, in forward hidden_states = self.mlp(hidden_states) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 168, in forward return self.down_proj(self.act_fn(self.gate_proj(hidden_state)) * self.up_proj(hidden_state)) RuntimeError: CUDA error: an illegal memory access was encountered CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. For debugging consider passing CUDA_LAUNCH_BLOCKING=1. Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1012.629504,867.106816,0.0,488.636416,482.553856,s,1,8.0609853515625,8.0609853515625,0.0,8.0609853515625,8.0609853515625,8.0609853515625,8.0609853515625,[8.0609853515625],,kWh,2.8416175170870397e-05,3.1269823558455027e-06,8.546673504000735e-06,4.0089831030716635e-05,,MB,1239.527424,1024.393216,0.0,616.562688,582.974464,s,10,0.27054739379882814,0.027054739379882814,0.000286009729749811,0.026956192016601564,0.027177914237976077,0.027537452983856202,0.027825083980560304,"[0.02789699172973633, 0.02697020721435547, 0.026962175369262695, 0.026941280364990234, 0.026929088592529297, 0.02695020866394043, 0.02691299247741699, 0.027000160217285157, 0.027098016738891603, 0.02688627243041992]",tokens/s,9462.29776622261,kWh,8.612702605701969e-07,9.498283774394124e-08,5.70487182053024e-07,1.5267402803671622e-06,tokens/kWh,167677504.3483068,MB,1250.967552,1036.976128,0.0,629.1456,597.192192,s,10,11.528859130859376,1.1528859130859375,0.013783837027398216,1.1535343017578126,1.1710545288085938,1.1736548278808594,1.175735067138672,"[1.1483973388671875, 1.157601318359375, 1.1445784912109376, 1.128986328125, 1.14946728515625, 1.1581435546875, 1.135988525390625, 1.176255126953125, 1.1589644775390624, 1.1704766845703125]",tokens/s,54.645476438659465,kWh,3.303405783776308e-05,3.64319500948955e-06,1.3625607508146755e-05,5.0302860355399364e-05,tokens/kWh,1252413.8698056713,,s,630,11.52320091629028,0.01829079510522267,0.0005419719436823987,0.018167872428894043,0.018848256111145018,0.01896842737197876,0.019749512386322023,"[0.017847360610961913, 0.017965568542480468, 0.018098560333251953, 0.018417472839355468, 0.018214399337768555, 0.017998592376708984, 0.01790777587890625, 0.017827775955200194, 0.0179652156829834, 0.01795055961608887, 0.01780121612548828, 0.017927391052246094, 0.01788934326171875, 0.017883871078491213, 0.0178155517578125, 0.0178985595703125, 0.017884096145629882, 0.01797324752807617, 0.017950016021728514, 0.01790979194641113, 0.01780784034729004, 0.017891519546508788, 0.018081792831420897, 0.017856224060058594, 0.017899808883666993, 0.01845043182373047, 0.01801366424560547, 0.017934879302978515, 0.01805516815185547, 0.01799782371520996, 0.018082847595214845, 0.018039775848388673, 0.01966044807434082, 0.017974815368652343, 0.017895584106445313, 0.017866432189941408, 0.017930944442749022, 0.01852035140991211, 0.01975663948059082, 0.01851798439025879, 0.018310815811157226, 0.01817884826660156, 0.018152671813964842, 0.018063455581665038, 0.01804342460632324, 0.01812224006652832, 0.018064031600952147, 0.018214752197265625, 0.01839302444458008, 0.018623903274536134, 0.01873107147216797, 0.01918617630004883, 0.01857356834411621, 0.019163135528564454, 0.01863270378112793, 0.018753856658935548, 0.018578111648559572, 0.01836953544616699, 0.018874368667602538, 0.018527456283569336, 0.01844918441772461, 0.01835212707519531, 0.01816166305541992, 0.01779408073425293, 0.018195423126220703, 0.018294143676757812, 0.01831590461730957, 0.01845452880859375, 0.018568864822387697, 0.018526559829711915, 0.01859993553161621, 0.019179519653320314, 0.01928544044494629, 0.019657279968261717, 0.027027456283569336, 0.018192384719848635, 0.018225151062011717, 0.01872822380065918, 0.0186265926361084, 0.018621280670166017, 0.01853775978088379, 0.01842848014831543, 0.01823744010925293, 0.018426912307739258, 0.01848828887939453, 0.01853545570373535, 0.01840224075317383, 0.018439359664916992, 0.018068319320678712, 0.017979391098022462, 0.017968544006347655, 0.018022815704345704, 0.017965248107910156, 0.018108415603637695, 0.018155263900756835, 0.018168064117431642, 0.018169183731079102, 0.018137760162353515, 0.018288639068603514, 0.01851372718811035, 0.017970815658569336, 0.01798201560974121, 0.017944063186645508, 0.017856096267700194, 0.017904544830322267, 0.01791328048706055, 0.01800454330444336, 0.017966495513916016, 0.018004575729370118, 0.017900800704956053, 0.01806822395324707, 0.017909311294555665, 0.017951391220092774, 0.017894720077514647, 0.017844703674316405, 0.01778700828552246, 0.018076992034912108, 0.017996288299560546, 0.017791040420532228, 0.017952768325805665, 0.01793132781982422, 0.017882047653198244, 0.01782956886291504, 0.0178035831451416, 0.019271072387695314, 0.018202495574951173, 0.017990976333618163, 0.01796780776977539, 0.01807974433898926, 0.017747488021850586, 0.01781705665588379, 0.018241600036621095, 0.017855424880981446, 0.01778188705444336, 0.017891616821289064, 0.017903936386108397, 0.01791209602355957, 0.01796089553833008, 0.017844287872314454, 0.01784377670288086, 0.01777299118041992, 0.01791391944885254, 0.01780726432800293, 0.017915136337280275, 0.017904415130615234, 0.017864704132080078, 0.01780735969543457, 0.01807360076904297, 0.01973206329345703, 0.01810883140563965, 0.01802444839477539, 0.017922048568725587, 0.018001920700073244, 0.01811849594116211, 0.01831056022644043, 0.017926496505737306, 0.017848735809326173, 0.017967103958129883, 0.0180097599029541, 0.01923072052001953, 0.018368864059448244, 0.018095327377319337, 0.01816655921936035, 0.019165184020996092, 0.018245567321777345, 0.018174016952514648, 0.01821615982055664, 0.018201248168945312, 0.018282623291015626, 0.018214527130126952, 0.01828006362915039, 0.01837727928161621, 0.018311359405517577, 0.018215999603271485, 0.01814358329772949, 0.01814588737487793, 0.018251232147216797, 0.018322175979614257, 0.018707584381103516, 0.018612672805786133, 0.01845065689086914, 0.018374656677246092, 0.018312671661376952, 0.018589792251586915, 0.0183089599609375, 0.018165567398071288, 0.018274879455566405, 0.018030624389648437, 0.017982751846313476, 0.017631263732910157, 0.018114559173583983, 0.017987712860107422, 0.017986719131469726, 0.017875295639038086, 0.017837984085083008, 0.017864511489868163, 0.017746591567993165, 0.017846271514892577, 0.017750015258789064, 0.017909696578979492, 0.017872575759887696, 0.017852800369262695, 0.01794867134094238, 0.018079456329345704, 0.018014047622680666, 0.01780928039550781, 0.017920576095581054, 0.017790048599243165, 0.017822240829467772, 0.017856895446777345, 0.0178155517578125, 0.01784796714782715, 0.017878816604614257, 0.017840703964233397, 0.01779302406311035, 0.017887231826782226, 0.017923807144165037, 0.01791209602355957, 0.01782748794555664, 0.017881439208984374, 0.01780940818786621, 0.01780748748779297, 0.01786662483215332, 0.01779302406311035, 0.01781068801879883, 0.017801984786987305, 0.0178155517578125, 0.01781372833251953, 0.017810783386230468, 0.017942975997924805, 0.017960704803466798, 0.017930496215820314, 0.01803468894958496, 0.01803376007080078, 0.018080671310424803, 0.018059263229370116, 0.018062528610229493, 0.018297664642333983, 0.01805516815185547, 0.017989599227905273, 0.01802025604248047, 0.017859872817993165, 0.01798659133911133, 0.017884992599487306, 0.017903615951538086, 0.017889280319213868, 0.018126623153686523, 0.017938655853271486, 0.017945632934570313, 0.017970144271850588, 0.018077695846557617, 0.017988927841186525, 0.017742656707763673, 0.018051103591918947, 0.018124799728393554, 0.017960960388183594, 0.01790959930419922, 0.017903776168823243, 0.0178767032623291, 0.01802217674255371, 0.017947071075439452, 0.018091903686523438, 0.01807993507385254, 0.01834815979003906, 0.018573183059692383, 0.01877180862426758, 0.018969823837280273, 0.019212928771972657, 0.018620832443237305, 0.01856121635437012, 0.018681568145751955, 0.018653472900390624, 0.01852182388305664, 0.01857535934448242, 0.018933855056762695, 0.018614015579223632, 0.018431264877319335, 0.01828134346008301, 0.01890265655517578, 0.018216703414916994, 0.017981407165527343, 0.017869152069091798, 0.017961280822753906, 0.017862752914428712, 0.018077407836914063, 0.01820857620239258, 0.018133472442626954, 0.01807513618469238, 0.01807606315612793, 0.01815353584289551, 0.018077280044555662, 0.018405567169189452, 0.01795430374145508, 0.018092544555664062, 0.01833590316772461, 0.01806540870666504, 0.01794867134094238, 0.017969152450561524, 0.017860639572143556, 0.017983455657958985, 0.01801817512512207, 0.018135072708129883, 0.01819251251220703, 0.018352256774902344, 0.01823535919189453, 0.018014080047607423, 0.018019519805908202, 0.01818502426147461, 0.018704383850097657, 0.018280031204223633, 0.018313631057739258, 0.018201728820800782, 0.018191232681274414, 0.018130239486694337, 0.018250240325927734, 0.017895040512084962, 0.018139520645141603, 0.018149375915527344, 0.018091264724731444, 0.018579519271850586, 0.01864569664001465, 0.018605215072631836, 0.01853036880493164, 0.01865603256225586, 0.018605728149414063, 0.018774143218994142, 0.01872630310058594, 0.01864089584350586, 0.01865715217590332, 0.018502592086791992, 0.018467967987060546, 0.01826700782775879, 0.018110368728637697, 0.018014303207397463, 0.018208383560180664, 0.018530176162719725, 0.018700639724731447, 0.01865513610839844, 0.01886832046508789, 0.018845855712890627, 0.01875948715209961, 0.01873871994018555, 0.018651359558105467, 0.01852569580078125, 0.018545600891113283, 0.018413568496704103, 0.018370559692382812, 0.018452192306518556, 0.018827552795410155, 0.018204864501953126, 0.018177183151245117, 0.018178720474243164, 0.01827599906921387, 0.018310752868652344, 0.018363136291503906, 0.01823744010925293, 0.018147327423095702, 0.018008064270019532, 0.01806278419494629, 0.018116575241088867, 0.018268768310546874, 0.01821696090698242, 0.018316896438598632, 0.018288543701171875, 0.01829052734375, 0.018329599380493163, 0.018292512893676758, 0.018185184478759765, 0.01821878433227539, 0.018216543197631836, 0.018342079162597655, 0.01813711929321289, 0.018206943511962892, 0.018097248077392578, 0.01813811111450195, 0.018569215774536133, 0.018210912704467775, 0.018083295822143554, 0.017761568069458007, 0.018022239685058592, 0.018148223876953126, 0.018013248443603514, 0.018005056381225584, 0.017940351486206055, 0.01801158332824707, 0.017914527893066405, 0.017978303909301756, 0.017871488571166994, 0.017955167770385742, 0.017845504760742186, 0.01787571144104004, 0.017794912338256835, 0.0178832950592041, 0.017758367538452148, 0.017870687484741212, 0.01778819274902344, 0.017893407821655275, 0.017856800079345703, 0.017801631927490236, 0.017811456680297853, 0.017819648742675782, 0.018138303756713867, 0.0181276798248291, 0.017788703918457032, 0.017926048278808594, 0.017818111419677735, 0.0178175048828125, 0.017885087966918945, 0.017841440200805664, 0.017838016510009765, 0.017836639404296875, 0.017834175109863282, 0.017908927917480468, 0.01787593650817871, 0.017882591247558595, 0.017869440078735352, 0.0180031681060791, 0.01787059211730957, 0.018051872253417967, 0.017983488082885742, 0.01807360076904297, 0.01803468894958496, 0.018167680740356445, 0.018030208587646486, 0.018162176132202147, 0.0182108154296875, 0.018097375869750975, 0.01805392074584961, 0.018116416931152343, 0.018124544143676757, 0.018297279357910156, 0.018188608169555663, 0.018069183349609375, 0.0182476806640625, 0.018274303436279296, 0.01851408004760742, 0.018390815734863283, 0.018491104125976564, 0.018698591232299805, 0.018628543853759765, 0.018549951553344726, 0.01841379165649414, 0.018676959991455078, 0.018732959747314454, 0.01862723159790039, 0.01875289535522461, 0.018737279891967773, 0.01876838493347168, 0.01877363204956055, 0.018750112533569337, 0.018716384887695312, 0.01902387237548828, 0.018593311309814453, 0.018416095733642578, 0.01841391944885254, 0.018680767059326173, 0.01831923294067383, 0.01841375923156738, 0.018487136840820314, 0.018619199752807618, 0.01859993553161621, 0.01855404853820801, 0.018333631515502928, 0.018309696197509766, 0.018161088943481445, 0.018082687377929688, 0.01805267143249512, 0.018177824020385744, 0.018102880477905273, 0.01814352035522461, 0.0182205753326416, 0.018330015182495118, 0.018433055877685546, 0.018299936294555664, 0.01845840072631836, 0.01828976058959961, 0.018436704635620117, 0.01864896011352539, 0.018886560440063475, 0.01880873680114746, 0.01873369598388672, 0.019148704528808593, 0.01936191940307617, 0.018957311630249024, 0.01887945556640625, 0.01884774398803711, 0.018966720581054686, 0.019066688537597656, 0.018999551773071287, 0.01892076873779297, 0.01883795166015625, 0.01885968017578125, 0.01884156799316406, 0.018962175369262695, 0.018983552932739258, 0.01885753631591797, 0.018927743911743164, 0.01887433624267578, 0.018902944564819335, 0.01884819221496582, 0.018890752792358398, 0.01884172821044922, 0.01893987274169922, 0.01887753677368164, 0.018735200881958007, 0.018895296096801757, 0.01890883255004883, 0.019278175354003908, 0.01883135986328125, 0.018630847930908204, 0.018636255264282225, 0.018491519927978515, 0.018380224227905275, 0.018254623413085938, 0.018124799728393554, 0.018321407318115233, 0.018394176483154296, 0.018676671981811523, 0.02060006332397461, 0.018762527465820314, 0.01835935974121094, 0.01820355224609375, 0.01824492835998535, 0.018021055221557617, 0.01797324752807617, 0.01799123191833496, 0.017961408615112303, 0.018250751495361327, 0.01817523193359375, 0.01816694450378418, 0.018086496353149413, 0.018104320526123048, 0.018155519485473632, 0.018067455291748045, 0.01803264045715332, 0.0181014404296875, 0.01826464080810547, 0.018415456771850587, 0.01834239959716797, 0.018528160095214845, 0.019941120147705077, 0.018479360580444335, 0.018435840606689454, 0.018409727096557617, 0.018386943817138672, 0.018308832168579103, 0.01834217643737793, 0.01822719955444336, 0.01819011116027832, 0.018178272247314452, 0.018233407974243165, 0.01822015953063965, 0.018954111099243165, 0.0185599365234375, 0.01820163154602051, 0.01804617691040039, 0.01811164855957031, 0.01818009567260742, 0.018444992065429686, 0.018081792831420897, 0.018343103408813476, 0.018559711456298828, 0.018257919311523436, 0.01816307258605957, 0.018051712036132813, 0.017911808013916015, 0.017897472381591797, 0.01784182357788086, 0.01811008071899414, 0.01802511978149414, 0.017948352813720703, 0.017948991775512697, 0.018045984268188476, 0.019989503860473632, 0.020513856887817383, 0.0179866886138916, 0.01797088050842285, 0.017952863693237304, 0.018054399490356445, 0.017857280731201172, 0.01798748779296875, 0.01801955223083496, 0.01791200065612793, 0.018180192947387694, 0.01808211135864258, 0.018082080841064455, 0.01839308738708496, 0.01838719940185547, 0.01801116752624512, 0.017988319396972655, 0.01820467185974121, 0.01803878402709961, 0.017936384201049805, 0.0180633602142334, 0.018329599380493163, 0.018397184371948243, 0.018448383331298827, 0.018386175155639648, 0.018914047241210936, 0.0184333438873291, 0.01854316711425781, 0.018796672821044923, 0.01878790473937988, 0.018891199111938477, 0.018832895278930666, 0.018921215057373045, 0.019004383087158204, 0.018935840606689455, 0.018824960708618162, 0.01885593605041504, 0.018804767608642577, 0.019825664520263672, 0.018900192260742188, 0.018848831176757813, 0.019108543395996092, 0.018970624923706055, 0.01886591911315918, 0.018827520370483398, 0.018851839065551757, 0.018843839645385742, 0.01869923210144043, 0.019450719833374024, 0.01880473518371582, 0.018903039932250978, 0.018827104568481447, 0.018771711349487304, 0.018823583602905272, 0.01901468849182129, 0.019045343399047853, 0.018797760009765626]",tokens/s,54.67230889894254,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1550.25408,1546.584064,0.0,1168.113664,1154.613248,s,1,8.2976953125,8.2976953125,0.0,8.2976953125,8.2976953125,8.2976953125,8.2976953125,[8.2976953125],,kWh,3.979363800413769e-05,4.382259203332224e-06,1.239528769397813e-05,5.6571184901448046e-05,,MB,1622.872064,1777.270784,0.0,1369.440256,1323.44832,s,10,0.7577682571411132,0.07577682571411133,0.00044258388422384527,0.07567547225952148,0.07648710784912109,0.07660779266357422,0.07670434051513672,"[0.07672847747802734, 0.07563311767578125, 0.07556864166259766, 0.0753691864013672, 0.07646028900146484, 0.075830078125, 0.07573881530761718, 0.075328125, 0.07571782684326171, 0.0753936996459961]",tokens/s,3378.3415653465026,kWh,2.332164941666785e-06,2.5719568844260757e-07,1.5518420703968655e-06,4.141202700506258e-06,tokens/kWh,61817790.26868311,MB,1630.355456,1798.242304,0.0,1390.411776,1377.233408,s,10,11.643792358398438,1.1643792358398437,0.007841945048422793,1.1645668945312502,1.173194677734375,1.1750732299804687,1.1765760717773437,"[1.171073974609375, 1.1650872802734376, 1.166047607421875, 1.154470703125, 1.1769517822265625, 1.1624239501953124, 1.1640465087890626, 1.1727772216796875, 1.161658447265625, 1.1492548828125]",tokens/s,54.106083362573315,kWh,3.382460167167134e-05,3.7303772633406755e-06,1.7983645780203018e-05,5.553862471521503e-05,tokens/kWh,1134345.6976661666,,s,630,11.64137124061584,0.01847836704859658,0.00045703590184294283,0.018392287254333495,0.018807874298095703,0.018984730339050293,0.019883276405334476,"[0.01957548713684082, 0.018872224807739257, 0.018896896362304686, 0.018661376953125, 0.018702272415161134, 0.018657344818115235, 0.018571264266967775, 0.018425344467163086, 0.01833590316772461, 0.01831337547302246, 0.018380992889404296, 0.01836851119995117, 0.018375904083251952, 0.018444480895996093, 0.018460256576538086, 0.018538751602172852, 0.018739967346191405, 0.018561023712158203, 0.018550783157348632, 0.018461952209472655, 0.01827712059020996, 0.018309055328369142, 0.018442304611206054, 0.018290687561035156, 0.018288383483886717, 0.01825190353393555, 0.018683008193969727, 0.018422336578369142, 0.018305471420288086, 0.01843974494934082, 0.019503231048583983, 0.018897216796875, 0.018648672103881835, 0.019211999893188475, 0.018535104751586914, 0.01857535934448242, 0.018656448364257814, 0.018612543106079103, 0.0185533447265625, 0.01846886444091797, 0.018767551422119142, 0.018821151733398437, 0.01874358367919922, 0.018750783920288085, 0.019919551849365235, 0.01897635269165039, 0.018952608108520508, 0.018535903930664063, 0.018475168228149413, 0.018280672073364257, 0.01821238327026367, 0.019173311233520507, 0.018625120162963867, 0.01846886444091797, 0.01843142318725586, 0.018332319259643556, 0.018443647384643554, 0.0184117431640625, 0.018368223190307616, 0.018268863677978517, 0.018184192657470705, 0.01827020835876465, 0.018164800643920898, 0.01866422462463379, 0.01874684715270996, 0.018831296920776366, 0.018774143218994142, 0.01899158477783203, 0.01865894317626953, 0.018737056732177734, 0.01857174491882324, 0.018743167877197264, 0.01853766441345215, 0.01862499237060547, 0.018597375869750975, 0.018636959075927734, 0.018628896713256834, 0.018663423538208008, 0.018480863571166992, 0.01855135917663574, 0.018596096038818358, 0.018657024383544923, 0.018516223907470705, 0.01857695960998535, 0.018364864349365233, 0.018220800399780274, 0.018366176605224608, 0.018388864517211913, 0.018467487335205077, 0.01840127944946289, 0.018355775833129882, 0.01838470458984375, 0.01832940864562988, 0.01818841552734375, 0.01822175979614258, 0.018132095336914063, 0.01826425552368164, 0.01837468719482422, 0.01838310432434082, 0.01821072006225586, 0.018196863174438478, 0.019295808792114257, 0.018892351150512694, 0.01872719955444336, 0.018615135192871092, 0.0186243839263916, 0.018579456329345705, 0.018577215194702148, 0.018284416198730467, 0.018382144927978517, 0.01820569610595703, 0.018523263931274413, 0.018250335693359376, 0.01826617622375488, 0.018364639282226564, 0.018362207412719725, 0.018781856536865236, 0.01825836753845215, 0.018304191589355468, 0.01825676727294922, 0.018392799377441406, 0.01832579231262207, 0.018339359283447265, 0.018242015838623046, 0.01844121551513672, 0.018529279708862305, 0.018452608108520507, 0.01854863929748535, 0.01848099136352539, 0.01818649673461914, 0.018108415603637695, 0.018231039047241212, 0.018303199768066405, 0.018210464477539063, 0.01828652763366699, 0.01885638427734375, 0.018431232452392577, 0.018567775726318358, 0.018438047409057617, 0.01833171272277832, 0.01841097640991211, 0.018391775131225585, 0.01850729560852051, 0.01858793640136719, 0.018525407791137694, 0.018432992935180664, 0.018476480484008788, 0.018471488952636717, 0.01833344078063965, 0.018600128173828126, 0.018406784057617188, 0.01821766471862793, 0.01946214485168457, 0.018311168670654295, 0.018464767456054687, 0.018404544830322264, 0.018334527969360352, 0.01827020835876465, 0.018343040466308594, 0.018366783142089844, 0.018287168502807618, 0.01815283203125, 0.018267871856689454, 0.018275232315063478, 0.018325504302978517, 0.01827020835876465, 0.018169471740722656, 0.018214879989624025, 0.018229408264160155, 0.018485279083251954, 0.018350303649902342, 0.01859993553161621, 0.019056320190429688, 0.018877952575683594, 0.0189935359954834, 0.018839967727661132, 0.018914688110351564, 0.01880745506286621, 0.018817024230957033, 0.018726816177368166, 0.018781856536865236, 0.018849407196044922, 0.018614336013793944, 0.01872768020629883, 0.01845020866394043, 0.01866569519042969, 0.018610176086425782, 0.019066879272460938, 0.018609952926635743, 0.018722591400146486, 0.018574687957763673, 0.01846361541748047, 0.01827984046936035, 0.01847724723815918, 0.018407840728759766, 0.018336864471435548, 0.018369440078735352, 0.018415231704711914, 0.018368032455444334, 0.018314048767089842, 0.018212703704833983, 0.018249919891357422, 0.019539104461669923, 0.018850175857543946, 0.019057119369506836, 0.01835811233520508, 0.018206880569458007, 0.01823539161682129, 0.01824563217163086, 0.018413568496704103, 0.018453760147094725, 0.01850649642944336, 0.018532352447509767, 0.018413568496704103, 0.0183767032623291, 0.0182476806640625, 0.01821286392211914, 0.018165760040283203, 0.018151039123535155, 0.018338176727294923, 0.018288480758666993, 0.018311328887939453, 0.01863270378112793, 0.0182108154296875, 0.018268032073974608, 0.018352256774902344, 0.018251775741577148, 0.01818009567260742, 0.01813862419128418, 0.01811712074279785, 0.01820262336730957, 0.018324928283691408, 0.018143264770507813, 0.01814582443237305, 0.018060831069946288, 0.01816009521484375, 0.018245439529418945, 0.018171424865722655, 0.018170528411865235, 0.018113887786865235, 0.01810908889770508, 0.01824358367919922, 0.018164928436279298, 0.018182687759399414, 0.01814147186279297, 0.01820022392272949, 0.01825827217102051, 0.01828659248352051, 0.01826348876953125, 0.01807417678833008, 0.018143199920654298, 0.018188192367553712, 0.01923276710510254, 0.018382848739624022, 0.019967296600341796, 0.02612220764160156, 0.02220719909667969, 0.01830297660827637, 0.01825484848022461, 0.01828339195251465, 0.01819558334350586, 0.01826028823852539, 0.018399328231811524, 0.018410079956054686, 0.018651136398315428, 0.018536127090454102, 0.018559295654296874, 0.018507776260375978, 0.01845587158203125, 0.018897567749023438, 0.018535711288452147, 0.018467584609985353, 0.018863744735717773, 0.018434431076049803, 0.01844633674621582, 0.018968544006347655, 0.018418752670288085, 0.018129888534545897, 0.01817804718017578, 0.01821286392211914, 0.01827840042114258, 0.0183308162689209, 0.018268991470336914, 0.01820579147338867, 0.018164640426635743, 0.018270015716552734, 0.01827449607849121, 0.018292448043823243, 0.018189760208129884, 0.018166624069213867, 0.01823744010925293, 0.018200159072875977, 0.018330015182495118, 0.01820400047302246, 0.018209184646606445, 0.018135295867919923, 0.018258943557739257, 0.018278879165649416, 0.0185001277923584, 0.018683904647827147, 0.018817024230957033, 0.01841695976257324, 0.01842655944824219, 0.01857686424255371, 0.01848297691345215, 0.018512447357177733, 0.01997433662414551, 0.019256704330444335, 0.018835552215576173, 0.018782432556152345, 0.018566879272460937, 0.018581727981567382, 0.01850364875793457, 0.01867407989501953, 0.018524160385131837, 0.018459392547607423, 0.018520063400268554, 0.01841152000427246, 0.018898815155029298, 0.018411647796630858, 0.018354175567626953, 0.018290687561035156, 0.018339839935302735, 0.018448383331298827, 0.018253376007080078, 0.018387392044067384, 0.01847091293334961, 0.01826576042175293, 0.018759904861450197, 0.018310335159301756, 0.018348831176757813, 0.018386560440063475, 0.018221599578857422, 0.01817731285095215, 0.018352863311767578, 0.01840332794189453, 0.018570240020751954, 0.018490367889404297, 0.018323135375976563, 0.018386720657348633, 0.01835196876525879, 0.019326656341552735, 0.019794464111328125, 0.018814815521240234, 0.018608768463134764, 0.01869824028015137, 0.018610176086425782, 0.018652544021606446, 0.01849337577819824, 0.018568992614746094, 0.01844112014770508, 0.01847500801086426, 0.018321407318115233, 0.018372608184814454, 0.018199712753295898, 0.018227903366088868, 0.018311328887939453, 0.018292415618896486, 0.018358591079711915, 0.018333696365356447, 0.018278047561645507, 0.018286943435668945, 0.01817705535888672, 0.019112672805786133, 0.018256128311157225, 0.018176000595092775, 0.018184192657470705, 0.01818828773498535, 0.018241535186767577, 0.018307071685791015, 0.018297887802124022, 0.018205631256103517, 0.01831324768066406, 0.01824492835998535, 0.01838969612121582, 0.018585599899291993, 0.018857120513916015, 0.018569536209106445, 0.019322944641113282, 0.018934431076049803, 0.018890687942504883, 0.01895359992980957, 0.01861305618286133, 0.01862246322631836, 0.0186060791015625, 0.018531871795654298, 0.018629087448120116, 0.01872627258300781, 0.018496128082275392, 0.01827840042114258, 0.01819443130493164, 0.01820627212524414, 0.01823708724975586, 0.018292543411254882, 0.01826019287109375, 0.018227968215942383, 0.01824563217163086, 0.018364416122436524, 0.01849692726135254, 0.018221248626708986, 0.018227615356445313, 0.018179807662963867, 0.01819267272949219, 0.018374656677246092, 0.01816511917114258, 0.01817580795288086, 0.01925926399230957, 0.018795455932617187, 0.018366464614868162, 0.01824563217163086, 0.018222463607788085, 0.018395776748657226, 0.01863987159729004, 0.018626943588256836, 0.018446048736572265, 0.018248031616210938, 0.018293312072753906, 0.018266111373901366, 0.01821286392211914, 0.01823744010925293, 0.018206720352172853, 0.018245248794555663, 0.018243392944335936, 0.018255456924438477, 0.01824048042297363, 0.018238943099975587, 0.018432064056396483, 0.018684383392333984, 0.018626560211181642, 0.018699615478515626, 0.01858627128601074, 0.018740480422973632, 0.018811647415161132, 0.018423519134521484, 0.018470687866210936, 0.018878271102905273, 0.018670272827148438, 0.018571264266967775, 0.01866547203063965, 0.01860812759399414, 0.01856716728210449, 0.018688032150268555, 0.01849625587463379, 0.018296831130981444, 0.018396480560302735, 0.018326208114624022, 0.01844633674621582, 0.018440191268920898, 0.018563072204589845, 0.01859350395202637, 0.018796031951904296, 0.0187619514465332, 0.018778528213500977, 0.018679967880249025, 0.018585599899291993, 0.018710527420043945, 0.018700288772583007, 0.018696191787719727, 0.0187064323425293, 0.018702335357666015, 0.01863270378112793, 0.018550783157348632, 0.018366464614868162, 0.018298879623413086, 0.018351808547973632, 0.01828895950317383, 0.018378751754760742, 0.018386335372924806, 0.018390880584716798, 0.01827097511291504, 0.018264064788818358, 0.01821900749206543, 0.018364416122436524, 0.018173280715942382, 0.01822127914428711, 0.019513792037963867, 0.01923686408996582, 0.02043289566040039, 0.018341888427734376, 0.01844985580444336, 0.018244159698486327, 0.018259967803955078, 0.01830259132385254, 0.018448768615722658, 0.01860812759399414, 0.018612064361572266, 0.018614431381225587, 0.01847500801086426, 0.018556928634643553, 0.018413568496704103, 0.018534400939941405, 0.01861222457885742, 0.018572799682617186, 0.018601760864257813, 0.01863542366027832, 0.018602048873901367, 0.01862838363647461, 0.018539775848388673, 0.019479520797729494, 0.019959583282470703, 0.01963849639892578, 0.018615711212158204, 0.018618560791015624, 0.01846067237854004, 0.019082368850708006, 0.018760320663452148, 0.0186711368560791, 0.018461408615112303, 0.018311168670654295, 0.01812396812438965, 0.018264896392822267, 0.018091167449951172, 0.018124736785888673, 0.0181144962310791, 0.018161983489990235, 0.018117279052734376, 0.018317312240600587, 0.018159616470336915, 0.018226240158081056, 0.018422719955444335, 0.018406816482543945, 0.018415327072143554, 0.018403648376464844, 0.018301504135131836, 0.01835379219055176, 0.018450592041015627, 0.018595872879028322, 0.018626752853393554, 0.018593791961669923, 0.018625791549682618, 0.018851839065551757, 0.0187194881439209, 0.018753536224365236, 0.018695167541503906, 0.01877017593383789, 0.018688127517700194, 0.01867782402038574, 0.01848556709289551, 0.018323711395263672, 0.018176000595092775, 0.01820262336730957, 0.018225151062011717, 0.018356224060058594, 0.018593568801879883, 0.0185118408203125, 0.01846249580383301, 0.01842019271850586, 0.01847500801086426, 0.018372159957885742, 0.018358720779418945, 0.018382848739624022, 0.01836345672607422, 0.018424768447875977, 0.018388992309570314, 0.018183935165405275, 0.018139392852783202, 0.01815545654296875, 0.018116512298583985, 0.018157440185546873, 0.01858905601501465, 0.018858911514282227, 0.019281919479370118, 0.01862144088745117, 0.018465503692626953, 0.018333152770996095, 0.018385663986206054, 0.01827027130126953, 0.018289600372314453, 0.018214656829833985, 0.018172096252441407, 0.01815488052368164, 0.018143552780151367, 0.01816204833984375, 0.018145280838012694, 0.018114559173583983, 0.018114559173583983, 0.01839289665222168, 0.01820044708251953, 0.01809436798095703, 0.018225183486938478, 0.018241535186767577, 0.018241535186767577, 0.018122560501098634, 0.018122943878173828, 0.018157567977905274, 0.01817804718017578, 0.018157472610473634, 0.018139232635498048, 0.018184192657470705, 0.01804083251953125, 0.018124704360961915, 0.01807574462890625, 0.018128480911254883, 0.01808393669128418, 0.018114879608154298, 0.018120704650878908, 0.018085887908935547, 0.018092031478881835, 0.018106367111206053, 0.018075647354125975, 0.018116607666015624, 0.018124799728393554, 0.018130495071411134, 0.018135360717773438, 0.018090112686157227, 0.018126848220825196, 0.018138656616210936, 0.018115039825439452, 0.018141183853149414, 0.018210304260253905, 0.0181847038269043, 0.018495487213134765, 0.018826847076416017, 0.01834009552001953, 0.018265472412109376, 0.018320159912109377, 0.018357440948486327, 0.018331968307495117, 0.018343904495239257, 0.0184182071685791, 0.018777280807495116, 0.018401311874389648, 0.018483007431030273, 0.01852070426940918, 0.01862998390197754, 0.018477279663085936, 0.018423999786376953, 0.01845907211303711, 0.018311328887939453, 0.018307071685791015]",tokens/s,54.11733609198705,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,6725.931008,7525.564416,0.0,7147.094016,7138.9184,s,1,11.535142578125,11.535142578125,0.0,11.535142578125,11.535142578125,11.535142578125,11.535142578125,[11.535142578125],,kWh,0.00013089681240415605,1.4431528435576988e-05,4.277336755201255e-05,0.00018810170839174558,,MB,1598.083072,8236.498944,0.0,7828.668416,7715.649536,s,10,6.331598999023438,0.6331598999023437,0.000872366389456193,0.6334921264648437,0.6339573791503906,0.6340320037841797,0.6340917034912109,"[0.631711181640625, 0.63167041015625, 0.6332913208007812, 0.632321533203125, 0.6339407958984375, 0.6335877075195312, 0.6339322509765625, 0.633640625, 0.6333965454101562, 0.6341066284179687]",tokens/s,404.32124655949394,kWh,1.8516487303124526e-05,2.0407386927067532e-06,1.2314159156876295e-05,3.287138515270757e-05,tokens/kWh,7787928.583195517,MB,1602.039808,8383.299584,0.0,7975.469056,7906.446848,s,10,28.94725219726563,2.8947252197265625,0.00898014543589322,2.8938502197265628,2.905096484375,2.9061799560546877,2.9070467333984373,"[2.88181298828125, 2.882487060546875, 2.885023681640625, 2.903577392578125, 2.892619384765625, 2.89383740234375, 2.893863037109375, 2.907263427734375, 2.904855712890625, 2.901912109375]",tokens/s,21.763723745064485,kWh,8.467853091104378e-05,9.341712162362736e-06,5.626167348152209e-05,0.0001502819165549286,tokens/kWh,419212.11443276523,,s,630,28.943798686981182,0.04594253759838286,0.0008122710189837785,0.045904142379760746,0.046646790695190433,0.046840979194641114,0.04782046039581299,"[0.04779507064819336, 0.04495065689086914, 0.04474095916748047, 0.044648990631103516, 0.04461363220214844, 0.04544483184814453, 0.04571932983398438, 0.045623809814453124, 0.044635326385498046, 0.04497836685180664, 0.04530611038208008, 0.04524070358276367, 0.04518502426147461, 0.045764606475830076, 0.04552246475219727, 0.04531043243408203, 0.04498604965209961, 0.045447486877441406, 0.04543590545654297, 0.045308929443359375, 0.04561078262329102, 0.04629052734375, 0.04560345458984375, 0.04547100830078125, 0.045310558319091795, 0.0450847053527832, 0.04592601776123047, 0.04594313430786133, 0.0455865592956543, 0.04554265594482422, 0.04557635116577148, 0.04556246566772461, 0.04563731384277344, 0.045510433197021485, 0.04584694290161133, 0.045674625396728515, 0.045491294860839845, 0.04545820617675781, 0.04652864074707031, 0.046083648681640624, 0.04581577682495117, 0.04604095840454102, 0.04589014434814453, 0.04587478256225586, 0.04617871856689453, 0.04602265548706055, 0.04551804733276367, 0.045879329681396484, 0.0459881591796875, 0.046352832794189454, 0.04638105773925781, 0.046036991119384765, 0.04570111846923828, 0.04548591995239258, 0.04614364624023438, 0.04616396713256836, 0.04589126586914063, 0.04573215866088867, 0.04700899124145508, 0.04680374526977539, 0.04637900924682617, 0.04680499267578125, 0.04697660827636719, 0.04751283264160156, 0.04509552001953125, 0.044724384307861326, 0.04536099243164062, 0.04571123123168945, 0.04567068862915039, 0.04487782287597656, 0.044750846862792966, 0.04503551864624023, 0.045716545104980466, 0.04541718292236328, 0.045203678131103514, 0.04540620803833008, 0.045383678436279294, 0.04529520034790039, 0.04582345581054687, 0.04560550308227539, 0.045072193145751956, 0.04505241775512695, 0.04539801788330078, 0.045832191467285156, 0.04603257751464844, 0.04550073623657227, 0.045424320220947265, 0.04530176162719726, 0.04592572784423828, 0.04560688018798828, 0.04559564971923828, 0.04547174453735352, 0.04565375900268555, 0.045377792358398436, 0.04560281753540039, 0.04560591888427734, 0.04550118255615234, 0.045760353088378905, 0.04569331359863281, 0.04564761734008789, 0.04583449554443359, 0.04576665496826172, 0.04562739181518555, 0.046280288696289064, 0.04588380813598633, 0.04629708862304688, 0.04630636978149414, 0.04594905471801758, 0.04570745468139648, 0.04611468887329102, 0.04590822219848633, 0.045789569854736326, 0.04634636688232422, 0.04609638214111328, 0.045954719543457034, 0.04582640075683594, 0.0462044792175293, 0.04603731155395508, 0.04564086532592773, 0.04605436706542969, 0.04653875350952148, 0.046241790771484374, 0.046102527618408204, 0.047126529693603515, 0.04661990356445313, 0.04629142379760742, 0.0469365119934082, 0.04492710494995117, 0.045400257110595706, 0.044738494873046875, 0.045190654754638675, 0.045195423126220706, 0.044863903045654296, 0.04542867279052734, 0.045219905853271486, 0.045385726928710936, 0.04572774505615235, 0.04555980682373047, 0.045279232025146485, 0.04565305709838867, 0.04584864044189453, 0.045523841857910155, 0.04516864013671875, 0.04538777542114258, 0.04541571044921875, 0.04523491287231445, 0.045946880340576174, 0.045889537811279295, 0.04571891021728516, 0.04554608154296875, 0.04526470565795898, 0.04549363327026367, 0.045257568359375, 0.045502464294433595, 0.04530585479736328, 0.04553932952880859, 0.04561305618286133, 0.04606771087646484, 0.045674015045166015, 0.04572208023071289, 0.04570019149780274, 0.04615852737426758, 0.04617647933959961, 0.0457400016784668, 0.04542367935180664, 0.04586979293823242, 0.046695777893066406, 0.04634425735473633, 0.046005088806152346, 0.046288257598876954, 0.04603500747680664, 0.04562182235717773, 0.046022529602050784, 0.04635865783691406, 0.04627046585083008, 0.04559872055053711, 0.04593164825439453, 0.046010398864746095, 0.04637900924682617, 0.04613820648193359, 0.046053375244140625, 0.04658560180664063, 0.04616729736328125, 0.04679782485961914, 0.04629043197631836, 0.04611695861816406, 0.045967105865478516, 0.04664582443237305, 0.0467147216796875, 0.0480371208190918, 0.045766494750976563, 0.04476515197753906, 0.04467193603515625, 0.0451945915222168, 0.04557244873046875, 0.04486076736450195, 0.044915679931640626, 0.045246463775634765, 0.045674495697021485, 0.0451932144165039, 0.045250144958496094, 0.04553548812866211, 0.045531295776367185, 0.04541644668579101, 0.0452935676574707, 0.045767807006835935, 0.04546780776977539, 0.04553945541381836, 0.045879295349121094, 0.046580318450927735, 0.04642566299438477, 0.04596371078491211, 0.04570217514038086, 0.045888481140136717, 0.04579641723632812, 0.045560768127441406, 0.04616518402099609, 0.045361408233642576, 0.04524499130249023, 0.04569702529907226, 0.045610847473144533, 0.04593679809570313, 0.04614080047607422, 0.0601032943725586, 0.043550559997558594, 0.046217376708984376, 0.04608204650878906, 0.045352958679199216, 0.04581990432739258, 0.04622732925415039, 0.04603635025024414, 0.0459496955871582, 0.04637081527709961, 0.04632524871826172, 0.04550092697143555, 0.0459791374206543, 0.04686489486694336, 0.04633139038085937, 0.045978111267089845, 0.045744129180908207, 0.045559551239013674, 0.045787391662597654, 0.04749027252197266, 0.04756969451904297, 0.046781566619873045, 0.0460890884399414, 0.04625727844238281, 0.04615647888183594, 0.045840576171875, 0.04602675247192383, 0.04712783813476563, 0.04647190475463867, 0.04791827011108398, 0.045677120208740235, 0.04518096160888672, 0.04485116958618164, 0.044507102966308595, 0.04514188766479492, 0.04514604949951172, 0.044902111053466795, 0.04547222518920899, 0.04569731140136719, 0.04545036697387695, 0.045276031494140626, 0.04511059188842773, 0.04592291259765625, 0.04616767883300781, 0.04549587249755859, 0.04497273635864258, 0.045338497161865235, 0.0458583984375, 0.045832286834716796, 0.04587382507324219, 0.0465797119140625, 0.04643132781982422, 0.04597443389892578, 0.045690879821777344, 0.04551065444946289, 0.045162494659423826, 0.04541753768920898, 0.04553414535522461, 0.045794559478759767, 0.045384449005126955, 0.04561056137084961, 0.04567599868774414, 0.04538876724243164, 0.04669356918334961, 0.04687545776367188, 0.04635846328735352, 0.045829566955566406, 0.04556864166259766, 0.04626432037353516, 0.046284801483154295, 0.04603903961181641, 0.04628416061401367, 0.0460865592956543, 0.04594710540771484, 0.046057441711425784, 0.04607574462890625, 0.04594243240356445, 0.04610307312011719, 0.04622467041015625, 0.04594761657714844, 0.04595238494873047, 0.045695358276367185, 0.04625433731079102, 0.046655487060546875, 0.04665958404541016, 0.0468927993774414, 0.04635263824462891, 0.04619216156005859, 0.046766368865966794, 0.04681894302368164, 0.04683740615844727, 0.04667596817016602, 0.047826816558837894, 0.045377662658691406, 0.04483071899414062, 0.04474060821533203, 0.04549155044555664, 0.04563011169433594, 0.04533225631713867, 0.0451519660949707, 0.045375999450683595, 0.04577062225341797, 0.04544729614257813, 0.045055999755859374, 0.045129726409912106, 0.045742080688476565, 0.04548198318481445, 0.04510515213012695, 0.04518064117431641, 0.045900062561035154, 0.046004222869873046, 0.045774078369140624, 0.04580537414550781, 0.046119873046875, 0.045723648071289064, 0.04525247955322265, 0.045936767578125, 0.046045185089111325, 0.045742080688476565, 0.04556083297729492, 0.04570412826538086, 0.04579904174804687, 0.04625174331665039, 0.045779678344726564, 0.046285888671875, 0.046177215576171875, 0.04584447860717773, 0.046601951599121096, 0.0464161605834961, 0.04607302474975586, 0.04582252883911133, 0.04626252746582031, 0.046295040130615236, 0.045886558532714845, 0.045894561767578126, 0.046300224304199215, 0.04595404815673828, 0.04647724914550781, 0.046473217010498044, 0.04614105606079102, 0.04609471893310547, 0.04596460723876953, 0.04615766525268555, 0.046041950225830075, 0.04597760009765625, 0.04656947326660156, 0.04622883224487305, 0.04616016006469727, 0.04670460891723633, 0.04663132858276367, 0.04628044891357422, 0.046323966979980466, 0.046534366607666015, 0.04656883239746094, 0.046277534484863284, 0.04782675170898437, 0.045475967407226564, 0.04513859176635742, 0.044875072479248046, 0.045173439025878906, 0.04520131301879883, 0.044918495178222655, 0.04545574569702148, 0.045271041870117185, 0.04579894256591797, 0.045725406646728514, 0.04535372924804688, 0.04532633590698242, 0.045164192199707034, 0.04607132720947266, 0.045857601165771485, 0.045606910705566404, 0.045426559448242185, 0.045553279876708985, 0.04565827178955078, 0.04685654449462891, 0.04646892929077148, 0.045611198425292966, 0.04548812866210938, 0.045369342803955076, 0.04565523147583008, 0.04548649597167969, 0.045311870574951174, 0.045992481231689454, 0.04616787338256836, 0.045752513885498045, 0.046077953338623044, 0.04580966567993164, 0.04545881652832031, 0.04613798522949219, 0.046176254272460936, 0.04595507049560547, 0.045970752716064454, 0.04593529510498047, 0.045870174407958986, 0.04681820678710937, 0.04683283233642578, 0.046401409149169924, 0.046040000915527346, 0.04615167999267578, 0.04639334487915039, 0.04592230224609375, 0.04557942581176758, 0.04618735885620117, 0.046120960235595705, 0.04602470397949219, 0.04584243011474609, 0.046470462799072264, 0.04632403182983398, 0.04638105773925781, 0.04637120056152344, 0.046677440643310544, 0.046489246368408205, 0.04617718505859375, 0.046415870666503906, 0.04664115142822266, 0.04647731018066406, 0.04635238265991211, 0.04909372711181641, 0.04595939254760742, 0.04487443161010742, 0.04487760162353516, 0.04535932922363281, 0.04547174453735352, 0.04543283081054687, 0.04562739181518555, 0.045879295349121094, 0.045743839263916015, 0.0453798713684082, 0.04621305465698242, 0.045975231170654295, 0.04574860763549805, 0.045303104400634765, 0.045440799713134764, 0.04556832122802734, 0.04560713577270508, 0.045695358276367185, 0.04601036834716797, 0.04650188827514649, 0.046458881378173826, 0.04613119888305664, 0.04611270523071289, 0.0460513916015625, 0.04578508758544922, 0.04567244720458984, 0.045559070587158204, 0.046245761871337894, 0.046067745208740234, 0.045873985290527344, 0.04598169708251953, 0.046532608032226565, 0.04698726272583008, 0.046405216217041016, 0.04554959869384766, 0.04633990478515625, 0.046037567138671874, 0.04574003219604492, 0.04636262512207031, 0.04674063873291016, 0.04660105514526367, 0.04617555236816406, 0.04656758499145508, 0.046684703826904296, 0.046118911743164064, 0.04612860870361328, 0.04597119903564453, 0.04627523040771484, 0.04663292694091797, 0.04637507247924805, 0.046243839263916016, 0.04593868637084961, 0.046663585662841796, 0.046889057159423826, 0.04638876724243164, 0.046449119567871094, 0.046843902587890625, 0.04636671829223633, 0.04634624099731445, 0.04698316955566406, 0.046851776123046876, 0.046940479278564456, 0.04780505752563476, 0.04603500747680664, 0.045502689361572264, 0.04520755386352539, 0.045146110534667966, 0.04553292846679687, 0.04529324722290039, 0.04535932922363281, 0.045467201232910155, 0.04618729782104492, 0.045746112823486326, 0.04572576141357422, 0.04526694488525391, 0.04597350311279297, 0.04572063827514648, 0.04534697723388672, 0.04557907104492188, 0.04553110504150391, 0.0456888313293457, 0.04611686325073242, 0.046427776336669925, 0.046524158477783205, 0.04600640106201172, 0.04602316665649414, 0.04599193572998047, 0.04566835021972656, 0.04549017715454102, 0.046071807861328126, 0.04581171035766601, 0.04573798370361328, 0.04595097732543945, 0.04645843124389649, 0.04629318237304687, 0.045937950134277344, 0.046240734100341796, 0.04601446533203125, 0.04586659240722656, 0.04574796676635742, 0.045951648712158205, 0.04639744186401367, 0.04642406463623047, 0.0466165771484375, 0.04716953659057617, 0.0469769287109375, 0.0461082878112793, 0.04616236877441406, 0.04612489700317383, 0.046740993499755856, 0.046451393127441405, 0.046266368865966793, 0.04608992004394531, 0.04666336059570313, 0.04665769577026367, 0.046251903533935546, 0.046400096893310545, 0.04676192092895508, 0.046370880126953125, 0.046005569458007815, 0.04671891021728516, 0.04682828903198242, 0.04637062454223633, 0.046766143798828125, 0.04664332962036133, 0.048355262756347654, 0.04548409652709961, 0.044856895446777345, 0.045179039001464846, 0.04527715301513672, 0.04579564666748047, 0.045202751159667966, 0.04543353652954102, 0.04569497680664063, 0.04551200103759766, 0.045808319091796876, 0.04577859115600586, 0.045529441833496095, 0.04597750473022461, 0.04562134552001953, 0.045848575592041016, 0.04558438491821289, 0.045271041870117185, 0.04589977645874024, 0.04594483184814453, 0.04648953628540039, 0.04611884689331055, 0.04582003021240234, 0.04572137451171875, 0.045725887298583984, 0.04607136154174805, 0.046136993408203125, 0.0458691520690918, 0.04565260696411133, 0.04606780624389648, 0.04612505722045898, 0.045623294830322264, 0.046102527618408204, 0.046682113647460936, 0.04630323028564453, 0.045444511413574216, 0.045945438385009765, 0.04623878479003906, 0.04595536041259766, 0.046137439727783204, 0.047042686462402346, 0.04753395080566406, 0.046740032196044924, 0.04611196899414063, 0.04574288177490234, 0.04599763107299805, 0.046419841766357425, 0.046072383880615235, 0.04630515289306641, 0.046408958435058593, 0.04662486267089844, 0.04611692810058594, 0.04612579345703125, 0.04639324951171875, 0.04668127822875977, 0.0459571533203125, 0.04610342407226563, 0.0468592643737793, 0.046682174682617185, 0.04637382507324219, 0.04614144134521484, 0.046244895935058594, 0.04663510513305664]",tokens/s,21.76632054462745,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,11031.482368,12225.282048,0.0,11846.811648,11814.785024,s,1,14.1742060546875,14.1742060546875,0.0,14.1742060546875,14.1742060546875,14.1742060546875,14.1742060546875,[14.1742060546875],,kWh,0.0002060095365125221,2.2711510691751467e-05,6.820699901001481e-05,0.0002969280462142884,,MB,2041.421824,13158.514688,0.0,12750.68416,12632.68864,s,10,11.269819213867187,1.126981921386719,0.0009837382519498324,1.1272222290039062,1.1280132080078127,1.1280138671875,1.12801439453125,"[1.124601806640625, 1.1260865478515625, 1.126607421875, 1.1271981201171875, 1.1273677978515626, 1.127246337890625, 1.126885498046875, 1.1280145263671875, 1.127798095703125, 1.1280130615234376]",tokens/s,227.1553741385659,kWh,3.29439294770835e-05,3.6311154198365797e-06,2.189140640200121e-05,5.846645129892129e-05,tokens/kWh,4378579.412852499,MB,2045.68576,13431.144448,0.0,13023.31392,12936.608256,s,10,48.449287109375,4.8449287109375,0.011159043985337812,4.8456611328125,4.8569484375,4.8600782226562504,4.86258205078125,"[4.82113916015625, 4.83734375, 4.8402119140625, 4.8375341796875, 4.85026513671875, 4.852009765625, 4.84683984375, 4.8562529296875, 4.844482421875, 4.8632080078125]",tokens/s,13.003287304883669,kWh,0.00014183703294833096,1.5647829244971095e-05,9.426238096539863e-05,0.00025174724315870067,tokens/kWh,250251.00259105916,,s,630,48.44516996765141,0.0768970951867482,0.0009903678534263585,0.07707817459106445,0.07788375549316406,0.07814800834655762,0.07881736396789552,"[0.07785676574707032, 0.07526579284667968, 0.0757882537841797, 0.07408258819580078, 0.0748359375, 0.07607504272460938, 0.07666191864013672, 0.07567622375488281, 0.0742996826171875, 0.07581849670410157, 0.07445555114746094, 0.075499267578125, 0.07802678680419922, 0.07563900756835938, 0.07463731384277343, 0.07814102172851563, 0.07496336364746094, 0.07424934387207031, 0.07650188446044921, 0.07589024353027343, 0.0773364486694336, 0.07680601501464844, 0.07670159912109376, 0.07657756805419921, 0.07660115051269531, 0.07485203552246093, 0.0754892807006836, 0.0772828140258789, 0.07673446655273437, 0.07661465454101563, 0.07576576232910157, 0.07504608154296875, 0.07724320220947266, 0.07693881225585937, 0.07721619415283203, 0.0769751968383789, 0.07723299407958985, 0.07545043182373047, 0.07555891418457031, 0.07767228698730469, 0.07710121917724609, 0.0766967010498047, 0.0776200942993164, 0.07671807861328125, 0.0770109405517578, 0.0765132827758789, 0.07711043548583985, 0.07820941162109375, 0.07761670684814453, 0.07686236572265626, 0.07683891296386719, 0.0775920639038086, 0.07747824096679687, 0.07700495910644531, 0.07720960235595703, 0.0770920639038086, 0.07698697662353515, 0.07749037170410156, 0.07751270294189454, 0.07655014038085937, 0.07779942321777343, 0.07731715393066406, 0.07791276550292969, 0.0786033935546875, 0.07609638214111328, 0.0756305923461914, 0.07455129241943359, 0.07676723480224609, 0.07450998687744141, 0.07465516662597656, 0.07625820922851563, 0.07547084808349609, 0.07696924591064454, 0.07560060882568359, 0.07573689270019532, 0.07680223846435547, 0.07652047729492187, 0.07694560241699219, 0.07668611145019531, 0.07704524993896485, 0.07631104278564453, 0.07454515075683593, 0.07735295867919922, 0.07654605102539062, 0.07673414611816406, 0.07509843444824219, 0.07729462432861328, 0.07688089752197266, 0.076676513671875, 0.07678009796142578, 0.07659254455566407, 0.07755427551269531, 0.07671356964111328, 0.0764072036743164, 0.0778885726928711, 0.07701395416259765, 0.07698614501953124, 0.07710720062255859, 0.07626569366455078, 0.07636544036865234, 0.07758006286621094, 0.07662448120117188, 0.07685939025878906, 0.0775308837890625, 0.07715046691894531, 0.07702937316894531, 0.07752671813964844, 0.07750028991699219, 0.07753363037109375, 0.077557373046875, 0.07737586975097656, 0.07689132690429687, 0.07641769409179687, 0.07771692657470704, 0.07712371063232422, 0.07690300750732422, 0.07663206481933593, 0.07772569274902344, 0.07736038208007813, 0.07743360137939453, 0.0775513916015625, 0.07758665466308594, 0.07787721252441407, 0.07761244964599609, 0.07663001251220704, 0.07716307067871093, 0.07869641876220704, 0.074655517578125, 0.07691494750976563, 0.07670098876953126, 0.0772656021118164, 0.0762265625, 0.07540735626220703, 0.07659724426269532, 0.07531903839111329, 0.0760732192993164, 0.07762092590332031, 0.07583980560302735, 0.07471923065185547, 0.0766402587890625, 0.07631027221679687, 0.07604454040527343, 0.07751065826416016, 0.07634099578857421, 0.07458019256591797, 0.07683650970458984, 0.07583952331542969, 0.07519017791748046, 0.07445343780517578, 0.07854003143310546, 0.0767452163696289, 0.07760297393798828, 0.07468450927734376, 0.07680588531494141, 0.07654153442382812, 0.07720793914794923, 0.07788368225097657, 0.07785167694091796, 0.07628211212158204, 0.07803977966308594, 0.07440998077392579, 0.07607872009277344, 0.07754790496826172, 0.0771256332397461, 0.07664435577392578, 0.07686492919921875, 0.07659990692138671, 0.07732963562011719, 0.07672029113769531, 0.07771405029296875, 0.07757004547119141, 0.07758233642578125, 0.07725670623779297, 0.07758643341064453, 0.0777330551147461, 0.07841260528564453, 0.07736319732666015, 0.07719526672363282, 0.07733193969726562, 0.07717731475830078, 0.0771502685546875, 0.07722755432128907, 0.07721212768554687, 0.07785881805419922, 0.07827426910400391, 0.0773532485961914, 0.07762739562988281, 0.07744108581542969, 0.07751264190673827, 0.07925007629394532, 0.07684095764160156, 0.07577315521240234, 0.07403542327880859, 0.07693778991699218, 0.07521485137939453, 0.07411917114257813, 0.07441522979736329, 0.07878131103515625, 0.07660953521728515, 0.07570432281494141, 0.07775027465820313, 0.07691590118408204, 0.07613113403320312, 0.0768736343383789, 0.07605872344970703, 0.07715360260009765, 0.07604022216796875, 0.07630095672607422, 0.07614364624023437, 0.07704061126708985, 0.07682457733154296, 0.07708057403564453, 0.07727104187011719, 0.07668326568603516, 0.07685517120361328, 0.07695168304443359, 0.0762449951171875, 0.07683891296386719, 0.07775437164306641, 0.07650918579101562, 0.0745832290649414, 0.07519519805908204, 0.07621577453613282, 0.07803919982910157, 0.07445337677001954, 0.0750223388671875, 0.07552931213378906, 0.07757711791992188, 0.07706829071044922, 0.07715606689453125, 0.07758057403564453, 0.07744306945800782, 0.07767040252685548, 0.07735238647460937, 0.07695795440673828, 0.07777699279785157, 0.07663398742675781, 0.07759436798095704, 0.07701155090332032, 0.07735887908935547, 0.07798089599609374, 0.07720857238769531, 0.07684262084960937, 0.07729564666748047, 0.07750211334228516, 0.07781961822509766, 0.07717577362060547, 0.07777587127685547, 0.07749120330810547, 0.07719647979736328, 0.07751763153076172, 0.0780505599975586, 0.07878246307373046, 0.0764587173461914, 0.07495254516601563, 0.076017822265625, 0.07713177490234376, 0.07462297821044922, 0.07767040252685548, 0.07696383666992188, 0.07652146911621094, 0.0753070068359375, 0.07718911743164063, 0.07734591674804688, 0.07665718078613282, 0.07711548614501954, 0.07725820922851563, 0.07634611511230469, 0.07687305450439454, 0.07635014343261719, 0.07636790466308593, 0.07749423980712891, 0.07670579528808594, 0.07464959716796875, 0.07715948486328125, 0.07751747131347657, 0.07706038665771485, 0.07659056091308594, 0.07613497924804688, 0.07696112060546875, 0.07727372741699219, 0.07772905731201171, 0.07576416015625, 0.07787494659423828, 0.0769664306640625, 0.07671807861328125, 0.07602381134033204, 0.07737753295898438, 0.07680818939208985, 0.07723763275146485, 0.07706585693359375, 0.07669452667236328, 0.07692902374267578, 0.07649075317382813, 0.07679385375976562, 0.07705776214599609, 0.07724674987792969, 0.07799398040771484, 0.07680732727050782, 0.07703638458251953, 0.0780546875, 0.07819951629638672, 0.07699251556396484, 0.07728476715087891, 0.07729417419433594, 0.07802674865722656, 0.07722354888916015, 0.07786943817138672, 0.07762739562988281, 0.07706829071044922, 0.07726416015625, 0.07718370819091797, 0.07761714935302734, 0.07793663787841797, 0.07815372467041015, 0.07822201538085938, 0.07638582611083984, 0.07641951751708985, 0.07677903747558594, 0.0764400634765625, 0.07453244781494141, 0.07653212738037109, 0.07511859130859375, 0.07612620544433593, 0.07707577514648438, 0.07667167663574219, 0.07467826843261718, 0.07851417541503906, 0.07683020782470704, 0.07688448333740235, 0.07638015747070312, 0.07708057403564453, 0.07680000305175781, 0.07677133178710938, 0.07669734191894531, 0.07671424102783203, 0.07683670043945312, 0.07716470336914062, 0.07534793853759765, 0.07585081481933593, 0.07744406127929687, 0.07660134124755859, 0.07698204803466797, 0.07632038116455078, 0.07743689727783203, 0.07684729766845703, 0.07713581085205078, 0.07724079895019531, 0.07901708984375, 0.07703839874267578, 0.07687177276611327, 0.07702275085449219, 0.07675504302978516, 0.07715264129638671, 0.07742463684082031, 0.07666185760498047, 0.07786589050292969, 0.07731404876708985, 0.07705779266357422, 0.07789183807373047, 0.07792176055908204, 0.07710979461669921, 0.07697612762451173, 0.07749964904785156, 0.07705062103271484, 0.07774185943603516, 0.07761328125, 0.07746351623535157, 0.07772278594970704, 0.07722473907470703, 0.07704790496826172, 0.07735056304931641, 0.07844694519042969, 0.0774878387451172, 0.07728729248046876, 0.0780005111694336, 0.07716361236572265, 0.07762992095947266, 0.07868777465820312, 0.07463884735107422, 0.07674976348876954, 0.07597875213623047, 0.07626751708984375, 0.07761920166015625, 0.07573296356201172, 0.07388572692871094, 0.07562649536132812, 0.07541350555419922, 0.07682048034667968, 0.0760627212524414, 0.07795478057861328, 0.07738396453857421, 0.07663929748535156, 0.07458265686035156, 0.07432556915283203, 0.0784268798828125, 0.07734476470947266, 0.07474121856689453, 0.07815593719482422, 0.07657305908203126, 0.07528982543945313, 0.07589558410644531, 0.07804108428955078, 0.0772116470336914, 0.0766976318359375, 0.0776396484375, 0.07667501068115234, 0.07691986846923828, 0.07769805145263672, 0.07686246490478515, 0.07673677062988281, 0.07715033721923828, 0.07756864166259765, 0.07607091522216797, 0.07791820526123047, 0.07557939147949219, 0.07883161926269532, 0.0771783676147461, 0.07769110107421875, 0.07741468811035156, 0.0776349105834961, 0.0775297622680664, 0.07637308502197265, 0.07794985961914062, 0.07712358093261719, 0.07705359649658203, 0.07745689392089844, 0.07706505584716797, 0.07772774505615235, 0.07788441467285157, 0.07722675323486328, 0.07766041564941406, 0.07817215728759766, 0.07750857543945312, 0.07726902770996094, 0.07788339233398438, 0.07720550537109375, 0.07744694519042969, 0.07702754974365235, 0.077559326171875, 0.07702960205078126, 0.07878041839599609, 0.07671807861328125, 0.07690035247802735, 0.07653939056396485, 0.07479106903076171, 0.07884371185302734, 0.07458870697021484, 0.0742850570678711, 0.07681436920166015, 0.07709487915039062, 0.07635763549804687, 0.07550351715087891, 0.0775332794189453, 0.07669760131835937, 0.07688396453857421, 0.07683830261230469, 0.07687596893310547, 0.07729132843017578, 0.07674285125732422, 0.07652754974365235, 0.07665827178955079, 0.07438365173339843, 0.07757884979248048, 0.0772833251953125, 0.07690444946289063, 0.0771396484375, 0.07730963134765625, 0.07695958709716796, 0.07720220947265626, 0.07817388916015625, 0.07613881683349609, 0.07578214263916015, 0.07691622161865234, 0.07792447662353516, 0.07729804992675782, 0.07712767791748047, 0.07690220642089844, 0.07772496032714844, 0.07786383819580078, 0.07814076995849609, 0.07719503784179688, 0.07709990692138671, 0.07713916778564453, 0.07766236877441406, 0.07706451416015625, 0.07729170989990235, 0.078010498046875, 0.07687075042724609, 0.07658553314208984, 0.07706588745117188, 0.07716320037841796, 0.07909171295166016, 0.07752703857421875, 0.07747174072265625, 0.07759613037109375, 0.0773616943359375, 0.0779653091430664, 0.07723417663574218, 0.07785001373291016, 0.0781145248413086, 0.0771629409790039, 0.07760931396484375, 0.07773808288574219, 0.0798538589477539, 0.07681455993652343, 0.07651155090332032, 0.07493017578125, 0.07427276611328125, 0.07538893127441407, 0.0744120330810547, 0.07764991760253906, 0.0762429428100586, 0.07546265411376953, 0.07448371124267578, 0.07621017456054688, 0.0773939208984375, 0.07683891296386719, 0.07456563568115235, 0.07678361511230469, 0.07679180908203125, 0.07448751831054687, 0.07650947570800781, 0.07458937835693359, 0.0744865951538086, 0.07691011047363282, 0.07778556823730469, 0.07503814697265625, 0.0751212158203125, 0.077946044921875, 0.07851910400390626, 0.07610755157470703, 0.07744493103027343, 0.07744121551513672, 0.07714387512207031, 0.07747830200195313, 0.07756594848632813, 0.07625523376464843, 0.07684300994873047, 0.07740118408203125, 0.07712246704101562, 0.07687318420410157, 0.07808873748779296, 0.0777615966796875, 0.07747465515136719, 0.07694960021972656, 0.07742873382568359, 0.07797555541992188, 0.07748812866210937, 0.07691776275634765, 0.07771033477783203, 0.07739929962158203, 0.07777254486083984, 0.07729779052734374, 0.07773683166503906, 0.07766015625, 0.07825532531738282, 0.07744796752929688, 0.07757414245605469, 0.07752294158935547, 0.07762902069091797, 0.07822582244873047, 0.07762892913818359, 0.07705577850341797, 0.078115234375, 0.07751026916503906, 0.07758882904052734, 0.07943996429443359, 0.07385906982421875, 0.07786495971679687, 0.07657667541503907, 0.07699993896484375, 0.07492620849609374, 0.07848998260498047, 0.07709935760498046, 0.07669145965576171, 0.07688595581054687, 0.07626886749267578, 0.0770852508544922, 0.07454534149169922, 0.07774822235107422, 0.07693708801269532, 0.07696784210205078, 0.07469276428222656, 0.07707244873046876, 0.07777689361572265, 0.07666812896728516, 0.07705680084228515, 0.07629414367675781, 0.07714733123779297, 0.07667795562744141, 0.07740620422363281, 0.07673004913330078, 0.0770167999267578, 0.07725270080566406, 0.07702169799804688, 0.07718096160888673, 0.0775920639038086, 0.07710768127441406, 0.0771520004272461, 0.0776908187866211, 0.07769292449951172, 0.0774167709350586, 0.07798966217041016, 0.07731222534179688, 0.07706214141845703, 0.07735295867919922, 0.07738162994384766, 0.07771734619140624, 0.07840735626220703, 0.07716690826416016, 0.07732192230224609, 0.07730838775634766, 0.07738368225097657, 0.077246337890625, 0.07760086059570312, 0.07749977874755859, 0.07777961730957031, 0.07746543884277343, 0.07728144073486329, 0.07792144012451171, 0.07734937286376953, 0.07823168182373047, 0.07755558776855469, 0.07745980834960937, 0.07762124633789062, 0.07748780822753906, 0.07770758056640625, 0.07758233642578125, 0.07750399780273437]",tokens/s,13.004392397026873,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.747328,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3548.471296,4490.985472,0.0,4112.515072,3976.487424,s,1,9.674736328125,9.674736328125,0.0,9.674736328125,9.674736328125,9.674736328125,9.674736328125,[9.674736328125],,kWh,8.025016947500205e-05,8.844910268695005e-06,2.7366133003997595e-05,0.00011646121274769465,,MB,1480.708096,4690.214912,0.0,4282.384384,4102.199808,s,10,3.0004592285156244,0.3000459228515625,0.002382701810579694,0.30021560668945313,0.30208082580566403,0.3034429244995117,0.3045326034545899,"[0.2961402893066406, 0.2970372619628906, 0.3009500427246094, 0.3017781372070312, 0.3001174621582031, 0.29800216674804686, 0.29995126342773437, 0.30031375122070314, 0.3048050231933594, 0.30136383056640625]",tokens/s,853.2027283258478,kWh,8.738438462867693e-06,9.636931963324836e-07,5.819645178588244e-06,1.5521776837788418e-05,tokens/kWh,16492957.13212145,MB,1489.108992,4698.60352,0.0,4290.772992,4102.202368,s,10,17.7380908203125,1.77380908203125,0.009381673081748462,1.7759326782226563,1.7821324951171875,1.7833516479492186,1.7843269702148437,"[1.762597412109375, 1.7813135986328126, 1.78457080078125, 1.781861572265625, 1.7780919189453126, 1.751927734375, 1.7751317138671876, 1.7761566162109375, 1.775708740234375, 1.770730712890625]",tokens/s,35.516787369166316,kWh,5.201708206213239e-05,5.737347936559142e-06,3.4305164699011866e-05,9.20595946977034e-05,tokens/kWh,684339.315275865,,s,630,17.73558416748047,0.028151720900762647,0.0005233327945577362,0.02804528045654297,0.02865833911895752,0.02905292730331421,0.030192383632659914,"[0.029214719772338867, 0.028591264724731447, 0.02875859260559082, 0.02827008056640625, 0.0282061767578125, 0.027969215393066408, 0.027839807510375975, 0.028016544342041014, 0.02785305595397949, 0.027818527221679688, 0.028062816619873046, 0.02813225555419922, 0.028037120819091797, 0.027942399978637695, 0.02782249641418457, 0.027848512649536132, 0.027775264739990233, 0.027744255065917968, 0.027762208938598633, 0.028221887588500978, 0.027789024353027342, 0.028057920455932618, 0.027993471145629882, 0.028236415863037108, 0.028836000442504884, 0.02829523277282715, 0.028057376861572267, 0.028858112335205077, 0.02809881591796875, 0.028055551528930665, 0.027926528930664062, 0.027853023529052733, 0.02823468780517578, 0.027906911849975586, 0.028028511047363282, 0.027899520874023438, 0.0278492488861084, 0.027924736022949218, 0.028332191467285155, 0.027929920196533203, 0.027844192504882813, 0.027827135086059572, 0.028008447647094727, 0.02775644874572754, 0.027807136535644532, 0.027783615112304687, 0.027832351684570312, 0.027713760375976563, 0.02770969581604004, 0.027840351104736327, 0.02791823959350586, 0.028047359466552735, 0.02779654312133789, 0.027519935607910155, 0.027535360336303712, 0.027743871688842774, 0.02761356735229492, 0.027615232467651366, 0.0276778564453125, 0.027767648696899416, 0.02763795280456543, 0.02759561538696289, 0.027728607177734375, 0.028792640686035157, 0.028007999420166015, 0.027926591873168944, 0.02797654342651367, 0.027932640075683593, 0.027881631851196288, 0.027856767654418944, 0.02782579231262207, 0.027744640350341798, 0.02784876823425293, 0.027950975418090822, 0.027766847610473634, 0.027715583801269532, 0.02816204833984375, 0.028196863174438477, 0.02804118347167969, 0.02815939140319824, 0.02861712074279785, 0.028243967056274414, 0.028172063827514648, 0.02802227210998535, 0.02807699203491211, 0.02874153518676758, 0.028016544342041014, 0.0279615364074707, 0.028800928115844726, 0.028137567520141602, 0.027987680435180663, 0.028027168273925783, 0.02809004783630371, 0.028483903884887696, 0.028245920181274413, 0.028418144226074218, 0.028051456451416015, 0.02852355194091797, 0.028158943176269532, 0.028483392715454102, 0.02825337600708008, 0.028355424880981445, 0.028582048416137696, 0.02797398376464844, 0.030203519821166994, 0.02831772804260254, 0.028476448059082032, 0.028310272216796876, 0.028288608551025392, 0.02802547264099121, 0.027908063888549803, 0.02809654426574707, 0.028258304595947265, 0.02834444808959961, 0.028395296096801758, 0.02819081687927246, 0.02813091278076172, 0.02800271987915039, 0.028902463912963867, 0.029223871231079102, 0.028594175338745118, 0.0280798397064209, 0.028053247451782226, 0.030282272338867187, 0.028395423889160155, 0.02833404731750488, 0.02878870391845703, 0.02840985679626465, 0.028224607467651368, 0.02835139274597168, 0.028132991790771486, 0.02814300727844238, 0.027931615829467772, 0.028121088027954103, 0.027938816070556642, 0.028051456451416015, 0.028102655410766602, 0.028092639923095703, 0.028360479354858397, 0.02805936050415039, 0.028330272674560546, 0.02801840019226074, 0.027935007095336913, 0.0281180477142334, 0.028301376342773438, 0.028976032257080078, 0.028325887680053712, 0.0284279670715332, 0.02899193572998047, 0.028288063049316407, 0.02833024024963379, 0.028401695251464843, 0.028829856872558592, 0.028331647872924803, 0.028662687301635743, 0.02885001564025879, 0.02876374435424805, 0.028721887588500975, 0.02905465507507324, 0.02828463935852051, 0.0280599365234375, 0.028203008651733398, 0.02790969657897949, 0.028058048248291015, 0.02796080017089844, 0.028436992645263674, 0.02804672050476074, 0.028306079864501954, 0.02873958396911621, 0.02829212760925293, 0.02798863983154297, 0.02816160011291504, 0.028266656875610353, 0.028145503997802735, 0.0281976318359375, 0.028452224731445312, 0.028525184631347657, 0.028062976837158204, 0.028092447280883788, 0.02822844886779785, 0.02867340850830078, 0.028368864059448242, 0.028400352478027344, 0.029015871047973634, 0.028192031860351564, 0.027931360244750975, 0.028256256103515624, 0.028200960159301756, 0.02850201606750488, 0.029050815582275392, 0.028413856506347656, 0.028254304885864258, 0.028296192169189452, 0.02815692710876465, 0.028100608825683594, 0.028112895965576173, 0.02844268798828125, 0.028374975204467773, 0.028080127716064454, 0.028108800888061523, 0.02850931167602539, 0.028263296127319336, 0.028379264831542968, 0.028482879638671875, 0.02835308837890625, 0.02878463935852051, 0.02835251235961914, 0.0320552978515625, 0.028366975784301758, 0.02827244758605957, 0.02817184066772461, 0.028137279510498048, 0.02839360046386719, 0.028000831604003906, 0.029224960327148438, 0.029593599319458007, 0.02870230484008789, 0.028492191314697265, 0.0287825927734375, 0.02813542366027832, 0.028061695098876953, 0.028094463348388672, 0.028077856063842773, 0.028247520446777342, 0.028110944747924804, 0.027994783401489257, 0.02811712074279785, 0.027995424270629884, 0.02810736083984375, 0.028077856063842773, 0.027997472763061523, 0.027933216094970702, 0.02801091194152832, 0.02798204803466797, 0.027823904037475585, 0.028040576934814453, 0.027878015518188477, 0.027856895446777344, 0.027828224182128908, 0.028000255584716797, 0.028388639450073243, 0.02795939254760742, 0.027771520614624023, 0.02870044708251953, 0.027910367965698242, 0.02792448043823242, 0.027828224182128908, 0.02790809631347656, 0.02838528060913086, 0.02797750473022461, 0.027824159622192382, 0.027938976287841796, 0.028885536193847657, 0.02805718421936035, 0.027797632217407227, 0.027871519088745116, 0.028000255584716797, 0.028310880661010743, 0.028186975479125978, 0.02915900802612305, 0.028278495788574217, 0.028160255432128908, 0.027805471420288087, 0.02789798355102539, 0.027963775634765625, 0.028459487915039064, 0.0279552001953125, 0.027864992141723634, 0.028686431884765624, 0.02787942314147949, 0.02800783920288086, 0.02827120018005371, 0.028033088684082032, 0.02844460868835449, 0.028139328002929686, 0.028166015625, 0.0280883846282959, 0.02788915252685547, 0.02784332847595215, 0.028297216415405273, 0.02786012840270996, 0.028248512268066406, 0.028330400466918947, 0.028495872497558594, 0.028525600433349608, 0.02841289520263672, 0.028353952407836915, 0.028256864547729493, 0.028106752395629882, 0.028597408294677735, 0.028134239196777343, 0.02794905662536621, 0.027983583450317383, 0.02843836784362793, 0.028486080169677734, 0.028277984619140627, 0.02841379165649414, 0.028306240081787108, 0.028092544555664064, 0.02789900779724121, 0.028484352111816408, 0.027940927505493166, 0.030082719802856445, 0.029275808334350586, 0.028498687744140626, 0.02832383918762207, 0.028121088027954103, 0.028077280044555664, 0.02827939224243164, 0.027940383911132814, 0.027849376678466795, 0.027633119583129882, 0.027838207244873046, 0.027794208526611328, 0.02813132858276367, 0.029343679428100587, 0.028348447799682618, 0.02805097579956055, 0.02799804878234863, 0.027734111785888672, 0.02773027229309082, 0.02767401695251465, 0.027540128707885744, 0.02756211280822754, 0.027502592086791993, 0.028020736694335937, 0.027751615524291992, 0.02767545509338379, 0.028507295608520507, 0.027679328918457032, 0.027495967864990235, 0.02745622444152832, 0.027582496643066407, 0.027944927215576173, 0.02773324775695801, 0.027798271179199217, 0.027709056854248047, 0.027736352920532226, 0.02755593681335449, 0.02798591995239258, 0.027502592086791993, 0.02760028839111328, 0.02787164878845215, 0.027815135955810547, 0.02777350425720215, 0.02764227294921875, 0.02768076705932617, 0.027692512512207033, 0.02776323127746582, 0.027929664611816406, 0.02788652801513672, 0.027722848892211913, 0.0277106876373291, 0.027702592849731447, 0.02776019287109375, 0.027640607833862303, 0.027545631408691405, 0.027828447341918944, 0.02759657669067383, 0.02762704086303711, 0.02750716781616211, 0.027856895446777344, 0.027613183975219727, 0.02758246421813965, 0.029181663513183593, 0.027738399505615234, 0.027609151840209963, 0.028114879608154297, 0.02749849510192871, 0.028112895965576173, 0.02768076705932617, 0.02792448043823242, 0.02785865592956543, 0.027799840927124023, 0.027878847122192383, 0.028043840408325197, 0.027528255462646485, 0.027751136779785156, 0.029416479110717774, 0.02823779106140137, 0.027775199890136718, 0.027812639236450196, 0.027985279083251952, 0.027676416397094727, 0.027468704223632814, 0.02761724853515625, 0.027966976165771484, 0.02823129653930664, 0.027691072463989257, 0.02888172721862793, 0.029683328628540038, 0.027871616363525392, 0.02786479949951172, 0.027715871810913086, 0.028116992950439453, 0.027715583801269532, 0.02868342399597168, 0.02802569580078125, 0.02786623954772949, 0.028934944152832032, 0.028350048065185547, 0.027709951400756837, 0.027795616149902343, 0.02771504020690918, 0.028098047256469725, 0.02781683158874512, 0.0277708797454834, 0.028061695098876953, 0.02795929527282715, 0.027705408096313475, 0.027905471801757814, 0.027892480850219725, 0.0280184326171875, 0.02783014488220215, 0.02773334312438965, 0.027951904296875, 0.02772505569458008, 0.028066335678100587, 0.028117216110229493, 0.027668384552001952, 0.031035263061523436, 0.029231264114379884, 0.028401952743530273, 0.02870863914489746, 0.02872915267944336, 0.029957727432250978, 0.02830191993713379, 0.028047359466552735, 0.02813132858276367, 0.027960927963256835, 0.0279835205078125, 0.028330751419067383, 0.028212959289550782, 0.028125471115112304, 0.02806287956237793, 0.028126047134399413, 0.028250431060791014, 0.027799232482910156, 0.027989824295043944, 0.028274879455566407, 0.028086271286010742, 0.029573535919189452, 0.028387327194213868, 0.02862073516845703, 0.02803446388244629, 0.027937631607055664, 0.02798195266723633, 0.02855824089050293, 0.027867935180664063, 0.02853023910522461, 0.02806345558166504, 0.028030912399291993, 0.028058399200439454, 0.0279135684967041, 0.02863801574707031, 0.02808940887451172, 0.027871616363525392, 0.0278768310546875, 0.028234272003173827, 0.028436895370483398, 0.029132799148559572, 0.028430335998535155, 0.028274240493774413, 0.028456960678100586, 0.028267200469970704, 0.027911935806274414, 0.027850528717041016, 0.027584255218505858, 0.028383712768554687, 0.028039167404174805, 0.02783616065979004, 0.027914464950561522, 0.02796985626220703, 0.027935808181762695, 0.027919231414794923, 0.028141216278076173, 0.028104223251342774, 0.028102336883544923, 0.028363679885864256, 0.02797772789001465, 0.02840575981140137, 0.027850751876831056, 0.029393983840942384, 0.028009408950805663, 0.027846656799316406, 0.029296640396118165, 0.028173887252807617, 0.027938560485839845, 0.028368831634521485, 0.028224256515502928, 0.028639232635498047, 0.028023807525634766, 0.02804956817626953, 0.027988191604614257, 0.02785139274597168, 0.028108800888061523, 0.028560800552368162, 0.0282935676574707, 0.028094751358032227, 0.028011520385742186, 0.02812918472290039, 0.027906688690185547, 0.027709152221679686, 0.027736703872680665, 0.030134271621704102, 0.028178016662597657, 0.027710880279541016, 0.028185600280761718, 0.02772991943359375, 0.028263776779174805, 0.0319268798828125, 0.030612768173217772, 0.028657855987548827, 0.028160480499267577, 0.02813145637512207, 0.027729471206665038, 0.029638944625854494, 0.029280351638793944, 0.02852771186828613, 0.02785152053833008, 0.02829539108276367, 0.027936159133911134, 0.02789232063293457, 0.02817433547973633, 0.027797760009765624, 0.027991167068481447, 0.02766035270690918, 0.02772435188293457, 0.027688255310058595, 0.027643871307373048, 0.02767500877380371, 0.027786624908447265, 0.027922496795654297, 0.02789616012573242, 0.02793235206604004, 0.027744415283203126, 0.027976415634155274, 0.028063072204589843, 0.027822431564331056, 0.02777529525756836, 0.02755583953857422, 0.02770534324645996, 0.028194208145141602, 0.027848512649536132, 0.027699071884155272, 0.028509056091308594, 0.028118688583374022, 0.027817535400390624, 0.027559999465942384, 0.027988447189331054, 0.028172000885009766, 0.027695680618286134, 0.02807423973083496, 0.0277521915435791, 0.028430080413818358, 0.027940576553344726, 0.028269088745117188, 0.027922431945800782, 0.02776406478881836, 0.027830944061279297, 0.0314204158782959, 0.028023008346557618, 0.0278089599609375, 0.027728479385375978, 0.027830272674560546, 0.027944320678710936, 0.02774880027770996, 0.030165119171142576, 0.028558431625366212, 0.028283807754516603, 0.029252960205078126, 0.028064416885375976, 0.02800230407714844, 0.0276845760345459, 0.02768451118469238, 0.0280994873046875, 0.027713407516479494, 0.02782575988769531, 0.02770150375366211, 0.028141151428222655, 0.02781395149230957, 0.02820649528503418, 0.027691968917846678, 0.027979776382446288, 0.027584575653076173, 0.02762678337097168, 0.027896480560302736, 0.027697376251220703, 0.027703071594238283, 0.02767033576965332, 0.027949247360229492, 0.027733375549316406, 0.027902591705322267, 0.027952640533447266, 0.028445056915283203, 0.02834649658203125, 0.028201984405517577, 0.02775551986694336, 0.027809791564941407, 0.028102655410766602, 0.027828256607055665, 0.027844287872314452, 0.027746591567993164, 0.027877376556396483, 0.03007302474975586, 0.029748287200927735, 0.028857088088989256, 0.028288063049316407, 0.02814627265930176, 0.02789516830444336, 0.028232511520385743, 0.027945119857788085, 0.027805696487426756, 0.028079200744628906, 0.02820947265625, 0.028395391464233397, 0.028562143325805665, 0.028288576126098634, 0.02798851203918457, 0.027824159622192382, 0.027700799942016602, 0.02780342483520508, 0.028018655776977538, 0.027871231079101562, 0.027931327819824218, 0.028008319854736327, 0.028118207931518556, 0.027992895126342773, 0.02795110321044922, 0.028184576034545897]",tokens/s,35.52180712238126,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4925.566976,7227.768832,0.0,6849.298432,6445.09696,s,1,11.7717197265625,11.7717197265625,0.0,11.7717197265625,11.7717197265625,11.7717197265625,11.7717197265625,[11.7717197265625],,kWh,0.00012844015038757182,1.4160420944724027e-05,4.179503343604285e-05,0.0001843956047683387,,MB,2807.660544,7559.118848,0.0,7151.28832,6823.319552,s,10,3.5995702514648436,0.35995702514648437,0.002581422794684266,0.3598522186279297,0.3623224304199219,0.3633622039794922,0.36419402282714847,"[0.3542107238769531, 0.35824066162109375, 0.3620205078125, 0.3598674011230469, 0.35879513549804687, 0.3598370361328125, 0.35945697021484374, 0.3606484680175781, 0.3644019775390625, 0.36209136962890626]",tokens/s,711.1960098454,kWh,1.0511655587500475e-05,1.1586021996240848e-06,6.9803726477873126e-06,1.865063043491187e-05,tokens/kWh,13726077.565764047,MB,2807.660544,7561.216,0.0,7153.385472,6823.322112,s,10,20.433365112304685,2.0433365112304687,0.004517455761130949,2.0419090576171874,2.048548193359375,2.050531665039063,2.0521184423828127,"[2.05251513671875, 2.0403682861328125, 2.048107421875, 2.042965576171875, 2.0398515625, 2.0364364013671876, 2.0408525390625, 2.040412841796875, 2.0455999755859375, 2.04625537109375]",tokens/s,30.83192594746045,kWh,5.879838770042018e-05,6.4856176940007134e-06,3.89495252072087e-05,0.0001042335306016296,tokens/kWh,604412.0316789409,,s,630,20.430577440261853,0.032429488000415625,0.0006968669825549724,0.03233548927307128,0.03280437545776367,0.03304746971130371,0.03600647537231446,"[0.036607009887695316, 0.032174240112304686, 0.032224319458007813, 0.0319997444152832, 0.032126976013183595, 0.03224518585205078, 0.03227500915527344, 0.032299263000488285, 0.03261209487915039, 0.03238092803955078, 0.03284787368774414, 0.032435585021972656, 0.032479873657226564, 0.03234201431274414, 0.0335948486328125, 0.03505411148071289, 0.032606208801269534, 0.032538047790527345, 0.03242448043823242, 0.03245654296875, 0.032601310729980466, 0.03282796859741211, 0.03274383926391602, 0.03280665588378906, 0.03261670303344726, 0.03242969512939453, 0.03254105758666992, 0.03280486297607422, 0.03261027145385742, 0.03277622222900391, 0.03271456146240234, 0.032529983520507816, 0.032586368560791015, 0.03246694564819336, 0.03301990509033203, 0.03256892776489258, 0.03291795349121094, 0.03252774429321289, 0.032526878356933596, 0.03242348861694336, 0.03256304168701172, 0.03563996887207031, 0.03259801483154297, 0.032210945129394535, 0.032167808532714844, 0.0321393928527832, 0.03196108818054199, 0.032059391021728514, 0.03157606315612793, 0.031784032821655275, 0.031814559936523434, 0.03196425628662109, 0.03212179183959961, 0.03237638473510742, 0.0327632942199707, 0.03217737579345703, 0.032167713165283204, 0.03213926315307617, 0.03203430557250977, 0.03232172775268555, 0.03233209609985352, 0.032487422943115234, 0.03207376098632812, 0.036987392425537106, 0.0325937614440918, 0.03215734481811523, 0.032131454467773435, 0.03226134490966797, 0.03240419387817383, 0.03233222579956055, 0.03251747131347656, 0.0323260498046875, 0.03258556747436524, 0.03258179092407226, 0.032529983520507816, 0.03251859283447266, 0.03247264099121094, 0.03268652725219726, 0.03263651275634766, 0.03271424102783203, 0.032594528198242184, 0.032723167419433596, 0.03248751831054687, 0.03254447937011719, 0.03479782485961914, 0.034224159240722654, 0.032589824676513675, 0.03255855941772461, 0.032271968841552735, 0.032640254974365235, 0.03255612945556641, 0.03243215942382813, 0.03211452865600586, 0.03240419387817383, 0.032118209838867186, 0.03212131118774414, 0.03199180793762207, 0.03265897750854492, 0.03543443298339844, 0.03205593490600586, 0.03170236778259277, 0.03184255981445312, 0.03181404876708984, 0.03184134483337402, 0.03242899322509766, 0.03212083053588867, 0.0321798095703125, 0.03206796646118164, 0.03204716873168945, 0.03194630432128906, 0.03207004928588867, 0.03155267143249512, 0.03188617515563965, 0.03196723175048828, 0.03227852630615234, 0.03223062515258789, 0.031984415054321286, 0.03210649490356445, 0.03160838317871094, 0.03137580871582031, 0.0317706241607666, 0.03157196807861328, 0.031770368576049805, 0.03165376091003418, 0.03132259178161621, 0.031250591278076174, 0.035724864959716794, 0.032318241119384764, 0.032099937438964846, 0.03226416015625, 0.03232998275756836, 0.03233577728271484, 0.03233520126342773, 0.032188705444335934, 0.0323897933959961, 0.03301580810546875, 0.03221920013427734, 0.032359584808349606, 0.03235715103149414, 0.032456703186035156, 0.03257468795776367, 0.03249641418457031, 0.033127422332763674, 0.0324659194946289, 0.03246876907348633, 0.03241510391235351, 0.03249603271484375, 0.03256374359130859, 0.032492576599121095, 0.03243507385253906, 0.03240249633789063, 0.03238544082641601, 0.032482078552246094, 0.03241923141479492, 0.032796161651611325, 0.033172321319580075, 0.03261030578613281, 0.03243212890625, 0.032366592407226565, 0.03241984176635742, 0.03286220932006836, 0.03251609420776367, 0.032181472778320314, 0.03227676773071289, 0.0322503662109375, 0.03223455810546875, 0.03232966232299805, 0.03227305603027344, 0.03215126419067383, 0.03242377471923828, 0.03247183990478516, 0.03238399887084961, 0.03272192001342773, 0.03234406280517578, 0.03226009750366211, 0.032200702667236326, 0.03210015869140625, 0.03228076934814453, 0.03245439910888672, 0.03240371322631836, 0.03231948852539063, 0.032292385101318356, 0.03236912155151367, 0.0326383056640625, 0.03271673583984375, 0.032730880737304686, 0.032807903289794924, 0.032712223052978516, 0.032667327880859375, 0.03676774215698242, 0.03281510543823242, 0.03222220611572266, 0.03275027084350586, 0.03253894424438476, 0.03227987289428711, 0.03223948669433594, 0.03208643341064453, 0.03221750259399414, 0.03220076751708984, 0.03225798416137695, 0.03188908767700195, 0.032030593872070315, 0.032170433044433594, 0.03210150527954102, 0.03220060729980469, 0.03247612762451172, 0.032753089904785156, 0.03220742416381836, 0.03232150268554688, 0.03213929748535156, 0.032116127014160153, 0.03256553649902344, 0.032241695404052736, 0.032067359924316405, 0.03201484680175781, 0.032104545593261716, 0.03213302230834961, 0.03213459014892578, 0.032231136322021486, 0.03222380828857422, 0.03213545608520508, 0.032538623809814454, 0.032124832153320314, 0.03261449432373047, 0.0319522876739502, 0.03213155364990234, 0.03222748947143555, 0.0321715202331543, 0.032268768310546876, 0.032317440032958986, 0.032203903198242186, 0.03213526535034179, 0.032320159912109375, 0.03254079818725586, 0.03284108734130859, 0.03241024017333984, 0.03259811019897461, 0.03282294464111328, 0.03247091293334961, 0.0335582389831543, 0.03300579071044922, 0.03239904022216797, 0.03237964630126953, 0.03243632125854492, 0.032375934600830075, 0.03238579177856445, 0.03231545639038086, 0.032581600189208984, 0.03242598342895508, 0.03223497772216797, 0.0328135986328125, 0.032421886444091795, 0.035751968383789065, 0.032403839111328124, 0.03231363296508789, 0.03243622589111328, 0.03394924926757813, 0.032475582122802736, 0.032623905181884766, 0.032612735748291016, 0.032463455200195314, 0.03248463821411133, 0.03239984130859375, 0.03199948883056641, 0.03211663818359375, 0.03236105728149414, 0.032124736785888675, 0.03196944046020508, 0.03216521453857422, 0.03217641448974609, 0.03245235061645508, 0.03273513412475586, 0.03250252914428711, 0.03242393493652344, 0.032828800201416014, 0.032341793060302736, 0.032487808227539064, 0.032530654907226564, 0.03368489456176758, 0.032699230194091794, 0.0328043212890625, 0.032764190673828124, 0.032559326171875, 0.032542560577392576, 0.03262892913818359, 0.03257462310791016, 0.03243868637084961, 0.03269420623779297, 0.03260067367553711, 0.03276291275024414, 0.03251408004760742, 0.03236950302124023, 0.03230515289306641, 0.032194561004638675, 0.03214495849609375, 0.032457279205322265, 0.03216371154785156, 0.032000030517578125, 0.03230636978149414, 0.031531808853149414, 0.03158176040649414, 0.03158806419372558, 0.031916576385498045, 0.03203641510009766, 0.031561344146728516, 0.031579135894775394, 0.03486259078979492, 0.03154921531677246, 0.031482175827026365, 0.03224617767333984, 0.03136716842651367, 0.03158220863342285, 0.031393247604370116, 0.03144918441772461, 0.03151097679138184, 0.0361987190246582, 0.03243008041381836, 0.03190336036682129, 0.03179523277282715, 0.0333087043762207, 0.03174393653869629, 0.03176281547546387, 0.03237887954711914, 0.03234815979003906, 0.032323585510253904, 0.032464897155761716, 0.032306400299072266, 0.032492321014404295, 0.03237887954711914, 0.03223961639404297, 0.03232342529296875, 0.03221315383911133, 0.03227212905883789, 0.03229721450805664, 0.032382110595703124, 0.03240841674804688, 0.0320206413269043, 0.03204899215698242, 0.03207167816162109, 0.031954944610595705, 0.03198361587524414, 0.03209852981567383, 0.032077598571777347, 0.031858272552490234, 0.03168092727661133, 0.03163750457763672, 0.03173513603210449, 0.0323263053894043, 0.033039585113525394, 0.03301251220703125, 0.031990976333618165, 0.031679296493530275, 0.03147980880737305, 0.031682559967041016, 0.03158435249328613, 0.03158211135864258, 0.03200841522216797, 0.03176217651367187, 0.032038753509521484, 0.032147647857666016, 0.032194561004638675, 0.03227033615112305, 0.03413520050048828, 0.03287289428710938, 0.03247078323364258, 0.032422046661376956, 0.03259801483154297, 0.03235891342163086, 0.032595264434814454, 0.03325724792480469, 0.03258790588378906, 0.03260931015014648, 0.03252358245849609, 0.03233017730712891, 0.03239731216430664, 0.03222022247314453, 0.03239622497558594, 0.03243212890625, 0.037165599822998045, 0.032536033630371095, 0.03218486404418945, 0.03214054489135742, 0.03203968048095703, 0.03213516616821289, 0.03219574356079102, 0.03229782485961914, 0.03240959930419922, 0.03223551940917969, 0.03239023971557617, 0.03239004898071289, 0.03225145721435547, 0.032309696197509764, 0.03232767868041992, 0.03237887954711914, 0.032413471221923826, 0.03236476898193359, 0.03224576187133789, 0.032282623291015625, 0.03246899032592773, 0.03232582473754883, 0.032382942199707034, 0.03230908966064453, 0.032284671783447266, 0.03212492752075195, 0.0325591049194336, 0.03223952102661133, 0.03214755249023438, 0.03215315246582031, 0.032403839111328124, 0.03226828765869141, 0.03270867156982422, 0.03221206283569336, 0.03216476821899414, 0.032110591888427735, 0.032251998901367186, 0.03233331298828125, 0.032111167907714844, 0.03197727966308594, 0.03210243225097656, 0.03220275115966797, 0.032535552978515625, 0.03212502288818359, 0.03216025543212891, 0.0324202880859375, 0.032671710968017575, 0.032355712890625, 0.03236518478393555, 0.03237068939208984, 0.03239116668701172, 0.03228876876831055, 0.03227590560913086, 0.032263904571533206, 0.03222774505615234, 0.03226668930053711, 0.03230310440063477, 0.03254684829711914, 0.033028064727783205, 0.03236454391479492, 0.03240867233276367, 0.03229718399047852, 0.032342113494873044, 0.036028255462646486, 0.03249168014526367, 0.03265945434570312, 0.03268198394775391, 0.032497665405273435, 0.03233587265014649, 0.03221724700927735, 0.03249324798583984, 0.032132545471191404, 0.032168991088867185, 0.03206111907958984, 0.03187638473510742, 0.03164233589172363, 0.03197542381286621, 0.03220028686523437, 0.031918336868286136, 0.03221696090698242, 0.03189545631408691, 0.03190329551696777, 0.032191295623779294, 0.0317391357421875, 0.03164588737487793, 0.031703615188598634, 0.03206505584716797, 0.032663646697998046, 0.032247966766357425, 0.03215766525268555, 0.03214771270751953, 0.03220684814453125, 0.032942081451416014, 0.032313056945800785, 0.03218259048461914, 0.032221153259277345, 0.03236454391479492, 0.03230515289306641, 0.03242598342895508, 0.0324771842956543, 0.032395263671875, 0.03235430526733398, 0.0323766098022461, 0.03236656188964844, 0.03255526351928711, 0.032301151275634765, 0.03238886260986328, 0.03245072174072266, 0.03276950454711914, 0.03224348831176758, 0.032701183319091796, 0.03259392166137695, 0.032889854431152346, 0.03258169555664062, 0.03275462341308594, 0.03244371032714844, 0.03232223892211914, 0.03230515289306641, 0.0325733757019043, 0.032272449493408205, 0.03272294235229492, 0.032587295532226564, 0.03262307357788086, 0.03263488006591797, 0.032280384063720705, 0.032252128601074216, 0.036184383392333985, 0.03270774459838867, 0.03219468688964844, 0.03203763198852539, 0.032365631103515625, 0.03285084915161133, 0.03209625625610352, 0.032169151306152347, 0.03220297622680664, 0.032191070556640625, 0.03215359878540039, 0.03229695892333984, 0.03236044692993164, 0.032317440032958986, 0.032382240295410154, 0.032300926208496096, 0.03249238586425781, 0.03292979049682617, 0.03252396774291992, 0.03242351913452148, 0.03233219146728516, 0.032534847259521486, 0.03251609420776367, 0.03238291168212891, 0.032126495361328125, 0.03233824157714844, 0.03228819274902344, 0.03220121765136719, 0.032218528747558595, 0.032140159606933595, 0.03230428695678711, 0.032205825805664064, 0.032632095336914066, 0.0322545280456543, 0.03227852630615234, 0.03236774444580078, 0.032161792755126956, 0.03212992095947265, 0.03252787017822266, 0.03595315170288086, 0.03251200103759765, 0.03240345764160156, 0.032322975158691404, 0.03246147155761719, 0.03246384048461914, 0.03251670455932617, 0.03222566223144531, 0.03224313735961914, 0.03245523071289062, 0.032384544372558596, 0.03233039855957031, 0.03238278579711914, 0.032387134552001956, 0.03246659088134766, 0.03253071975708008, 0.03228793716430664, 0.03216073608398438, 0.03244972610473633, 0.03220137786865234, 0.032094207763671875, 0.032317440032958986, 0.032441600799560544, 0.03220761489868164, 0.03566227340698242, 0.032579265594482425, 0.032473793029785154, 0.032643070220947264, 0.03356467056274414, 0.032901119232177735, 0.032595966339111326, 0.03257753753662109, 0.03267379379272461, 0.03260588836669922, 0.03395372772216797, 0.03285452651977539, 0.032927616119384766, 0.03260540771484375, 0.03271350479125976, 0.03251366424560547, 0.03356038284301758, 0.03272867202758789, 0.03262108612060547, 0.03299296188354492, 0.03260083389282226, 0.032281982421875, 0.032528545379638674, 0.03241366577148438, 0.03217049789428711, 0.032118785858154295, 0.03213312149047851, 0.03305392074584961, 0.03297359848022461, 0.03223270416259766, 0.03159670448303223, 0.03190863990783691, 0.03282720184326172, 0.03209331130981445, 0.031533952713012695, 0.031624799728393556, 0.03149196815490723, 0.03147334480285645, 0.03201520156860352, 0.032032833099365235, 0.03196512031555176, 0.03154745674133301, 0.03175827217102051, 0.031948575973510744, 0.03187132835388184, 0.032853759765625, 0.03593423843383789, 0.03219055938720703, 0.03213926315307617, 0.03214131164550781, 0.03234761428833008, 0.0319879035949707, 0.0319780158996582, 0.03187078475952149, 0.03196723175048828, 0.03238447952270508, 0.03180521583557129, 0.03201433563232422, 0.032026782989501956, 0.03329289627075195, 0.032917407989501955, 0.032242015838623045, 0.031964223861694334]",tokens/s,30.83613284265184,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,888.6272,638.517248,0.0,260.046848,253.883392,s,1,8.1063505859375,8.1063505859375,0.0,8.1063505859375,8.1063505859375,8.1063505859375,8.1063505859375,[8.1063505859375],,kWh,2.2307581012470715e-05,2.453471619660418e-06,7.547506037985929e-06,3.230855867011706e-05,,MB,1197.21984,751.763456,0.0,343.932928,312.754176,s,16,0.18802291202545168,0.011751432001590728,0.0003932107752449667,0.011671087741851806,0.012107920169830321,0.012526600122451782,0.012900865983963013,"[0.01237065601348877, 0.011300288200378417, 0.011611424446105956, 0.011666367530822754, 0.01299443244934082, 0.01167580795288086, 0.011574015617370606, 0.011341152191162109, 0.01158505630493164, 0.011582271575927734, 0.011845184326171874, 0.011739456176757812, 0.011546303749084472, 0.011733183860778809, 0.011708959579467774, 0.01174835205078125]",tokens/s,21784.579101963627,kWh,3.4663537898855505e-07,3.822086240254721e-08,2.2844337357008875e-07,6.132996149611911e-07,tokens/kWh,417414251.9495947,MB,1208.389632,776.92928,0.0,369.098752,313.412096,s,16,10.237614807128907,0.6398509254455567,0.009020818156885704,0.6407603759765625,0.6492899475097655,0.6501135559082031,0.6516580139160156,"[0.6292039184570313, 0.6278231811523437, 0.6450326538085938, 0.6494700317382812, 0.6400562133789063, 0.619185546875, 0.63930419921875, 0.629085693359375, 0.63702001953125, 0.6465849609375, 0.648677001953125, 0.6401679077148438, 0.6413528442382812, 0.64910986328125, 0.6520441284179688, 0.6434966430664063]",tokens/s,98.46043428964379,kWh,1.8093102338200042e-05,1.9953449318071566e-06,7.187882003928834e-06,2.727632927393603e-05,tokens/kWh,2309694.9507864984,,s,1008,10.228204399108881,0.010147028173719133,0.0002855753306472145,0.010167295932769774,0.010352002906799317,0.010433611059188842,0.011110925121307368,"[0.011118975639343262, 0.01171878433227539, 0.010147520065307617, 0.010025600433349609, 0.010050559997558594, 0.010089471817016601, 0.010200768470764161, 0.010006848335266114, 0.010266143798828125, 0.010023072242736816, 0.01002291202545166, 0.01017843246459961, 0.010164095878601074, 0.00996787166595459, 0.010042624473571777, 0.009949888229370117, 0.00993062400817871, 0.009908384323120117, 0.009882847785949707, 0.009728832244873047, 0.009756671905517577, 0.010051584243774414, 0.010440671920776367, 0.00986348819732666, 0.009830112457275391, 0.009830400466918946, 0.00986019229888916, 0.009956128120422363, 0.00994655990600586, 0.009920224189758301, 0.009878527641296387, 0.009809344291687012, 0.009914239883422852, 0.009816991806030273, 0.009807616233825684, 0.00982425594329834, 0.0099104642868042, 0.009903039932250977, 0.009868351936340333, 0.010100480079650878, 0.01003551959991455, 0.00989356803894043, 0.009795519828796386, 0.009797760009765625, 0.009787391662597657, 0.009783488273620605, 0.009867072105407715, 0.010084511756896973, 0.009961312294006347, 0.0099104642868042, 0.00989574432373047, 0.009975872039794921, 0.009875391960144043, 0.009886719703674317, 0.00985587215423584, 0.009899328231811524, 0.009855808258056641, 0.009789312362670898, 0.009814399719238281, 0.010008255958557128, 0.009771072387695312, 0.009715744018554687, 0.009741920471191406, 0.009354911804199218, 0.009695167541503907, 0.009781408309936523, 0.010012672424316407, 0.0097958402633667, 0.009852383613586426, 0.009718303680419921, 0.009915807723999023, 0.009787936210632325, 0.009732159614562988, 0.009711615562438965, 0.010186335563659669, 0.010751392364501953, 0.012026880264282227, 0.010252287864685058, 0.009975711822509765, 0.009895168304443359, 0.0098985595703125, 0.01011571216583252, 0.01005123233795166, 0.010126912117004394, 0.010033599853515625, 0.01021945571899414, 0.00991648006439209, 0.009857024192810059, 0.009897983551025391, 0.009754624366760254, 0.009791487693786622, 0.009835583686828613, 0.009959808349609375, 0.010148415565490723, 0.010043392181396485, 0.009945088386535645, 0.009893088340759278, 0.009788064002990723, 0.009768671989440918, 0.009685407638549805, 0.009684672355651855, 0.009707200050354003, 0.009730688095092773, 0.009787263870239258, 0.009748160362243652, 0.009804544448852539, 0.009753984451293945, 0.009834879875183106, 0.009965503692626953, 0.009875552177429199, 0.010106368064880371, 0.009898688316345215, 0.009950943946838379, 0.00995695972442627, 0.009920255661010742, 0.00982857608795166, 0.009889344215393066, 0.009997119903564453, 0.010043456077575683, 0.01005731201171875, 0.010043807983398437, 0.010059712409973144, 0.010094592094421387, 0.010084223747253419, 0.010135904312133789, 0.01008777618408203, 0.009892448425292969, 0.010185279846191407, 0.010172032356262206, 0.010083711624145508, 0.01010086441040039, 0.010156800270080566, 0.010164128303527833, 0.010182144165039063, 0.0101725435256958, 0.010059712409973144, 0.01005568027496338, 0.01006345558166504, 0.010332384109497071, 0.010270624160766602, 0.01015830421447754, 0.010264415740966797, 0.010387455940246582, 0.010084704399108886, 0.01017625617980957, 0.010275360107421874, 0.010217375755310058, 0.0101725435256958, 0.010208895683288574, 0.010234304428100586, 0.010274368286132813, 0.010301183700561524, 0.010289792060852052, 0.010233247756958008, 0.010264287948608399, 0.010273632049560546, 0.010163392066955566, 0.01052348804473877, 0.010266752243041992, 0.010176095962524414, 0.010187040328979493, 0.010182656288146973, 0.010354911804199219, 0.010251520156860351, 0.010566176414489746, 0.010295552253723144, 0.010249983787536622, 0.01042841625213623, 0.010782367706298828, 0.010348896026611328, 0.010242143630981445, 0.010249152183532714, 0.010502847671508789, 0.01021571159362793, 0.010381312370300292, 0.010094592094421387, 0.010201087951660156, 0.010175871849060058, 0.010065983772277833, 0.0101396484375, 0.01010540771484375, 0.01011302375793457, 0.010153568267822266, 0.010140095710754395, 0.010108960151672364, 0.010264575958251953, 0.010340512275695801, 0.010264351844787597, 0.010206815719604492, 0.010054207801818848, 0.010325695991516113, 0.010261887550354004, 0.010317888259887695, 0.010333056449890137, 0.010249471664428711, 0.010320768356323243, 0.010201120376586914, 0.010204480171203614, 0.010160672187805175, 0.010131456375122071, 0.010188799858093261, 0.01017251205444336, 0.010108832359313966, 0.010164223670959472, 0.010084351539611817, 0.010202943801879883, 0.01027295970916748, 0.010198944091796875, 0.010244192123413086, 0.010145792007446289, 0.010259776115417481, 0.010187456130981445, 0.010168319702148437, 0.010096096038818359, 0.010375712394714355, 0.010208800315856933, 0.010196672439575196, 0.010725215911865234, 0.010185600280761719, 0.010266304016113281, 0.010312064170837403, 0.010196479797363281, 0.010226431846618653, 0.010210847854614258, 0.010256287574768066, 0.010248831748962402, 0.010272192001342774, 0.010170047760009765, 0.010232576370239258, 0.010149472236633301, 0.010142111778259277, 0.010210816383361816, 0.01043449592590332, 0.010228096008300781, 0.01026198387145996, 0.010297887802124024, 0.010630911827087402, 0.010293408393859864, 0.01026476764678955, 0.0102575044631958, 0.010301728248596192, 0.010322463989257813, 0.01033011245727539, 0.010283007621765136, 0.010222751617431641, 0.010233951568603515, 0.010195712089538574, 0.010179712295532226, 0.010215328216552735, 0.010259424209594727, 0.011761664390563965, 0.012142592430114747, 0.010226048469543458, 0.010266304016113281, 0.0102640962600708, 0.01020902442932129, 0.010154335975646973, 0.010291135787963868, 0.010217920303344726, 0.01065113639831543, 0.010248800277709962, 0.01019388771057129, 0.010267007827758788, 0.01022828769683838, 0.010102784156799317, 0.010122495651245117, 0.010417119979858398, 0.010708831787109374, 0.010276800155639648, 0.010168160438537598, 0.010106207847595215, 0.010119999885559082, 0.010096768379211425, 0.010090208053588868, 0.01011257553100586, 0.010101344108581543, 0.010057120323181153, 0.010205792427062989, 0.010031359672546387, 0.010022656440734863, 0.010000384330749512, 0.010087712287902832, 0.010031840324401856, 0.010053631782531738, 0.010057855606079101, 0.010076031684875489, 0.009973695755004883, 0.010143808364868164, 0.01071836757659912, 0.010038111686706543, 0.009936896324157715, 0.009943039894104003, 0.010030559539794922, 0.009943552017211914, 0.009899392127990722, 0.009919136047363282, 0.009991328239440917, 0.010127296447753906, 0.010074943542480468, 0.010117216110229492, 0.010123264312744141, 0.010088447570800782, 0.010676480293273926, 0.010055168151855469, 0.010422528266906738, 0.009965696334838868, 0.010133376121520996, 0.010092543601989747, 0.010059967994689941, 0.01013923168182373, 0.0101112003326416, 0.010160127639770507, 0.010136608123779298, 0.010132351875305175, 0.010063967704772948, 0.00985916805267334, 0.010191295623779297, 0.010116352081298827, 0.010086720466613769, 0.010076512336730957, 0.010090687751770019, 0.010087360382080078, 0.01009552001953125, 0.010200736045837402, 0.010103136062622071, 0.009981727600097657, 0.01026249599456787, 0.010178688049316407, 0.010071743965148925, 0.010011039733886718, 0.010083776473999023, 0.009988703727722169, 0.010011775970458985, 0.010209471702575683, 0.009923263549804688, 0.009852928161621094, 0.009910271644592286, 0.009861184120178222, 0.009893823623657226, 0.009783391952514648, 0.010086655616760254, 0.009871007919311524, 0.009862272262573243, 0.010031968116760255, 0.009787008285522461, 0.009804224014282227, 0.009788895606994629, 0.0097259521484375, 0.009629471778869629, 0.009720159530639648, 0.009707839965820312, 0.009648192405700683, 0.009602272033691406, 0.009667103767395019, 0.009545280456542968, 0.009560480117797851, 0.009551360130310058, 0.009644672393798829, 0.009547616004943848, 0.009504863739013672, 0.009511136054992675, 0.00951244831085205, 0.009677311897277833, 0.009565247535705566, 0.009574655532836915, 0.009590815544128418, 0.009476767539978028, 0.009543680191040039, 0.009550975799560548, 0.00956112003326416, 0.009594592094421386, 0.009619359970092773, 0.009668448448181152, 0.009604928016662597, 0.009621696472167969, 0.009752960205078124, 0.009938495635986328, 0.010043840408325195, 0.009901280403137208, 0.01015503978729248, 0.010150624275207519, 0.010129343986511231, 0.010153183937072754, 0.010107775688171387, 0.010158080101013184, 0.010147295951843261, 0.010187264442443847, 0.0101396484375, 0.010178720474243164, 0.010264255523681641, 0.010166048049926758, 0.010107071876525878, 0.010194399833679198, 0.01019753646850586, 0.010221088409423828, 0.010647999763488769, 0.010352800369262696, 0.010362943649291993, 0.010370623588562012, 0.01033033561706543, 0.010402112007141114, 0.010179807662963867, 0.010305983543395996, 0.010213631629943848, 0.010233375549316406, 0.01025500774383545, 0.010213215827941895, 0.010175552368164063, 0.010207712173461914, 0.01018899154663086, 0.010686176300048828, 0.010424863815307618, 0.01044825553894043, 0.010193535804748535, 0.010164031982421875, 0.010065759658813477, 0.009938272476196289, 0.01001369571685791, 0.010045696258544921, 0.010065664291381836, 0.01010912036895752, 0.010177696228027344, 0.010029727935791016, 0.009969440460205078, 0.010004704475402831, 0.00994428825378418, 0.009925439834594726, 0.00989692783355713, 0.009855999946594238, 0.009810111999511718, 0.009793279647827148, 0.009859135627746582, 0.009918560028076171, 0.010182559967041016, 0.010156031608581542, 0.010070015907287597, 0.010192895889282226, 0.010051584243774414, 0.010010623931884765, 0.00994099235534668, 0.009845855712890626, 0.009618464469909669, 0.00989897632598877, 0.009858624458312988, 0.009884320259094238, 0.009824031829833985, 0.009744671821594239, 0.009789152145385742, 0.009875040054321289, 0.00984438419342041, 0.009957792282104493, 0.010053024291992188, 0.010098976135253906, 0.010072575569152833, 0.010070176124572753, 0.01030947208404541, 0.010119327545166016, 0.010235903739929199, 0.010000384330749512, 0.010120736122131348, 0.010004704475402831, 0.009957920074462891, 0.009922271728515624, 0.009875231742858887, 0.009830623626708985, 0.009862688064575196, 0.009873888015747071, 0.009889375686645508, 0.00997212791442871, 0.009861120223999023, 0.009989184379577637, 0.009851327896118164, 0.009900544166564941, 0.009889792442321778, 0.00991641616821289, 0.009856032371520997, 0.009808671951293946, 0.00978172779083252, 0.009924351692199707, 0.00988595199584961, 0.009938655853271484, 0.009873632431030274, 0.00987331199645996, 0.009807071685791015, 0.009855648040771485, 0.00988684844970703, 0.010003328323364258, 0.010044416427612305, 0.01001471996307373, 0.010021056175231934, 0.010080960273742676, 0.010128928184509276, 0.010158687591552735, 0.010145440101623535, 0.010156512260437012, 0.01005072021484375, 0.010105567932128906, 0.010194720268249512, 0.010132863998413085, 0.010169183731079102, 0.010282336235046386, 0.01010256004333496, 0.010056575775146485, 0.010180607795715332, 0.009937088012695312, 0.010063936233520508, 0.010201215744018555, 0.010033568382263184, 0.010046815872192383, 0.010119839668273926, 0.010168416023254395, 0.010153056144714356, 0.01008249568939209, 0.01002575969696045, 0.010047327995300294, 0.009912320137023926, 0.009928031921386719, 0.010017248153686523, 0.010011072158813477, 0.010306336402893067, 0.010135968208312989, 0.01004310417175293, 0.010033568382263184, 0.009933247566223144, 0.010000415802001953, 0.01005123233795166, 0.010058112144470215, 0.010022848129272462, 0.010034367561340332, 0.010107775688171387, 0.010080191612243652, 0.010383359909057617, 0.01042950439453125, 0.010155200004577637, 0.010245887756347657, 0.010074111938476562, 0.01006991958618164, 0.009994336128234863, 0.009992192268371582, 0.010284735679626464, 0.010150208473205566, 0.010067935943603515, 0.009996128082275391, 0.010315967559814452, 0.010305536270141602, 0.010028863906860351, 0.009981599807739258, 0.010035743713378906, 0.010027008056640625, 0.010102463722229003, 0.010100319862365722, 0.010236512184143067, 0.010147968292236327, 0.01012940788269043, 0.010080256462097169, 0.010196224212646484, 0.010025728225708007, 0.010094335556030273, 0.010029120445251465, 0.010082143783569336, 0.010094688415527344, 0.010143775939941407, 0.010158399581909179, 0.01018051242828369, 0.010228927612304688, 0.010148320198059082, 0.010172767639160156, 0.010091296195983887, 0.01027455997467041, 0.010193056106567382, 0.01025385570526123, 0.010388031959533691, 0.010249759674072266, 0.01027939224243164, 0.01017958354949951, 0.0102488956451416, 0.010170687675476074, 0.010108511924743652, 0.010197407722473144, 0.010178751945495605, 0.010143487930297852, 0.0102740478515625, 0.010259103775024415, 0.010188960075378417, 0.010259936332702637, 0.010175007820129395, 0.01030348777770996, 0.010283007621765136, 0.010264127731323243, 0.01021174430847168, 0.010317855834960938, 0.010342399597167968, 0.010477760314941405, 0.01059615993499756, 0.010211487770080567, 0.010159968376159668, 0.010528736114501953, 0.010192928314208984, 0.010229567527770995, 0.010326208114624024, 0.010175840377807617, 0.010228384017944335, 0.01022976016998291, 0.010241984367370605, 0.01029081630706787, 0.010211775779724121, 0.010448896408081054, 0.01022156810760498, 0.010344127655029296, 0.010239487648010253, 0.010236736297607421, 0.010293439865112304, 0.010297151565551757, 0.01026460838317871, 0.010241888046264648, 0.0102741756439209, 0.0103156156539917, 0.010162176132202149, 0.010231007575988769, 0.010247872352600098, 0.010149248123168945, 0.010194944381713868, 0.010185343742370605, 0.010145792007446289, 0.01016972827911377, 0.010272416114807129, 0.010194944381713868, 0.0101693115234375, 0.01021132755279541, 0.010414112091064453, 0.01012940788269043, 0.01023196792602539, 0.010279775619506836, 0.010171456336975098, 0.010266559600830078, 0.010373120307922363, 0.010186752319335938, 0.010203136444091796, 0.010401344299316407, 0.010296832084655762, 0.010296256065368652, 0.010266655921936036, 0.010282976150512695, 0.010143872261047363, 0.010196864128112793, 0.01021491241455078, 0.01023846435546875, 0.010205120086669921, 0.010098431587219238, 0.010379424095153808, 0.01011513614654541, 0.010228128433227538, 0.01011292839050293, 0.010346495628356933, 0.01032585620880127, 0.010218879699707032, 0.01022169589996338, 0.010119903564453124, 0.01014140796661377, 0.010141823768615723, 0.010166144371032715, 0.010194944381713868, 0.010167455673217774, 0.010221920013427734, 0.010252799987792969, 0.010198847770690918, 0.010073951721191406, 0.010132991790771484, 0.010246720314025879, 0.01009721565246582, 0.010216320037841797, 0.010183520317077636, 0.010078207969665527, 0.010369024276733398, 0.011003968238830566, 0.013030752182006836, 0.01027244758605957, 0.010238495826721192, 0.010231776237487793, 0.01019331169128418, 0.010283007621765136, 0.010227871894836425, 0.010235424041748048, 0.010178144454956056, 0.010246975898742676, 0.010250144004821778, 0.010273792266845704, 0.010304512023925782, 0.01035260772705078, 0.010340383529663086, 0.010329248428344727, 0.01032588768005371, 0.010320480346679688, 0.010637311935424805, 0.010522624015808106, 0.01033011245727539, 0.010314047813415527, 0.01022492790222168, 0.010274880409240723, 0.010254688262939453, 0.010070015907287597, 0.01001587200164795, 0.010035296440124512, 0.010068767547607422, 0.009980159759521484, 0.01002847957611084, 0.010035743713378906, 0.010010399818420411, 0.00994099235534668, 0.010012672424316407, 0.009885312080383301, 0.009941375732421875, 0.009797632217407226, 0.00986348819732666, 0.009858752250671387, 0.009805824279785156, 0.00981606388092041, 0.009868512153625488, 0.009777952194213867, 0.009789664268493653, 0.009854751586914063, 0.009899711608886719, 0.010018752098083497, 0.01002943992614746, 0.010194944381713868, 0.010184672355651856, 0.010782848358154297, 0.010358688354492187, 0.010178560256958008, 0.010188608169555665, 0.010116352081298827, 0.010126272201538085, 0.010201184272766114, 0.010216896057128905, 0.010258367538452148, 0.010162976264953614, 0.010135616302490234, 0.010081695556640626, 0.010103072166442871, 0.010074048042297364, 0.010252351760864259, 0.01144380760192871, 0.010277152061462402, 0.010196895599365234, 0.01027280044555664, 0.010188032150268554, 0.010476415634155273, 0.010246047973632813, 0.010447135925292969, 0.010290528297424316, 0.01018899154663086, 0.010125663757324219, 0.010228063583374023, 0.010196415901184083, 0.010182880401611328, 0.010241920471191407, 0.009974047660827637, 0.01029145622253418, 0.010178751945495605, 0.01022156810760498, 0.010087648391723632, 0.01018070411682129, 0.01019155216217041, 0.010256383895874023, 0.010268863677978515, 0.01034348773956299, 0.010230527877807618, 0.010178560256958008, 0.010209280014038086, 0.01018611240386963, 0.010222335815429688, 0.01020902442932129, 0.010227840423583984, 0.010203136444091796, 0.01001471996307373, 0.010037247657775878, 0.010028032302856446, 0.01009990406036377, 0.010028863906860351, 0.010008576393127442, 0.009955327987670898, 0.010079936027526855, 0.010184384346008301, 0.010194815635681153, 0.010136128425598145, 0.010074560165405273, 0.010117183685302735, 0.010095840454101562, 0.010082079887390136, 0.010007455825805664, 0.010045215606689454, 0.010074111938476562, 0.010207008361816407, 0.010162431716918946, 0.010126463890075684, 0.010152640342712403, 0.010112735748291015, 0.01041862392425537, 0.010207008361816407, 0.010348768234252929, 0.010272768020629883, 0.01026358413696289, 0.010181599617004395, 0.010147839546203614, 0.010176511764526367, 0.010215295791625976, 0.010209407806396485, 0.010202336311340331, 0.010248991966247558, 0.010115072250366211, 0.010136927604675293, 0.010260128021240234, 0.010213472366333009, 0.010309920310974122, 0.010201087951660156, 0.010195263862609863, 0.0102357759475708, 0.01025830364227295, 0.010226240158081054, 0.010037280082702637, 0.010248127937316895, 0.010195648193359375, 0.010140735626220703, 0.010488767623901368, 0.010375167846679688, 0.01031123161315918, 0.010289312362670899, 0.010414560317993165, 0.010347455978393554, 0.010377728462219238, 0.010301823616027833, 0.010319744110107422, 0.010371199607849121, 0.010329567909240722, 0.010260128021240234, 0.010294143676757812, 0.010170368194580079, 0.010167360305786133, 0.010224448204040528, 0.010258655548095703, 0.010239904403686523, 0.010256192207336427, 0.010431967735290527, 0.01035654354095459, 0.0103056001663208, 0.010412896156311035, 0.010381312370300292, 0.01042851161956787, 0.010368415832519531, 0.010301952362060546, 0.010280960083007813, 0.010326047897338868, 0.010258399963378906, 0.010279104232788086, 0.010207200050354004, 0.010323488235473632, 0.010114432334899903, 0.010167231559753418, 0.010167360305786133, 0.010203424453735352, 0.01034716796875, 0.010335391998291015, 0.010458175659179687, 0.010312735557556152, 0.010304256439208985, 0.010248191833496094, 0.010171839714050293, 0.010307744026184082, 0.010277152061462402, 0.010549375534057617, 0.010293024063110352, 0.010315999984741211, 0.010286144256591797, 0.01028326416015625, 0.010321727752685547, 0.010310751914978027, 0.010289055824279785, 0.010233983993530274, 0.010214495658874511, 0.010279711723327636, 0.01025216007232666, 0.010364704132080078, 0.010141471862792968, 0.010319999694824219, 0.01035968017578125, 0.010362624168395996, 0.010475008010864258, 0.010324671745300293, 0.010363840103149414, 0.010379872322082519, 0.010402079582214355, 0.010223648071289063, 0.010207296371459961, 0.010258336067199707, 0.010266048431396484, 0.010261055946350097, 0.010273887634277343, 0.010339167594909668, 0.010351743698120117, 0.010372032165527343, 0.010339679718017579, 0.01036291217803955, 0.010285696029663087, 0.01042841625213623, 0.010244064331054688, 0.010301471710205078, 0.010324031829833984, 0.01027603244781494, 0.010246912002563476, 0.010223615646362304, 0.010249631881713867, 0.010234432220458984, 0.010528800010681152, 0.010483807563781738, 0.010330016136169434, 0.01032806396484375, 0.010270719528198241, 0.010302687644958496, 0.010349216461181641, 0.0103538236618042, 0.010711968421936035, 0.010450271606445313, 0.01037507152557373, 0.010285887718200683, 0.010233887672424316, 0.010774496078491212, 0.011421695709228515, 0.01023795223236084, 0.010586112022399903, 0.01074176025390625, 0.010184096336364747, 0.010195136070251465, 0.010217375755310058, 0.010214079856872558, 0.01024124813079834, 0.01017046356201172, 0.01016387176513672, 0.010150688171386719, 0.0102390718460083, 0.01040278434753418, 0.010397695541381835, 0.010231136322021484, 0.010221311569213867, 0.010234784126281739, 0.01023788833618164, 0.01025654411315918, 0.010371040344238282, 0.010332863807678222, 0.010373184204101563, 0.010409472465515136, 0.010481504440307618, 0.010326623916625977, 0.010354496002197266, 0.010274975776672364, 0.010219072341918945, 0.010039775848388673, 0.010088704109191894, 0.010286848068237304, 0.010170368194580079, 0.010171648025512695, 0.010141504287719727, 0.01014675235748291, 0.010106240272521973, 0.010067999839782715, 0.010082559585571289, 0.010172863960266114, 0.010286656379699707, 0.010797599792480468, 0.013413279533386231, 0.010668959617614746, 0.010178496360778808, 0.010169856071472168, 0.010119711875915528, 0.010139679908752442, 0.010196991920471191, 0.010176287651062012, 0.010172736167907715, 0.010020671844482422, 0.009940095901489258, 0.010011872291564942, 0.009977248191833496, 0.009939295768737793, 0.01002905559539795, 0.009879551887512206, 0.010013792037963867, 0.010007455825805664, 0.010195103645324707, 0.01019478416442871, 0.01029689598083496, 0.010332608222961425, 0.010191200256347656, 0.010164031982421875, 0.011173727989196778, 0.01134550380706787, 0.010014143943786621, 0.010004575729370118, 0.009947711944580079, 0.009881919860839843, 0.009842559814453125, 0.009889920234680176, 0.009834303855895996, 0.009941216468811034, 0.009879679679870605, 0.009856351852416992, 0.009857536315917969, 0.009772224426269531, 0.009671680450439453, 0.009697088241577148]",tokens/s,98.55102231705696,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1488.13824,1546.584064,0.0,1168.113664,1154.613248,s,1,8.2988203125,8.2988203125,0.0,8.2988203125,8.2988203125,8.2988203125,8.2988203125,[8.2988203125],,kWh,3.984401601249677e-05,4.387897293824016e-06,1.3327788440001265e-05,5.755970174632205e-05,,MB,1472.036864,1777.270784,0.0,1369.440256,1323.44832,s,10,0.7596723175048828,0.07596723175048828,0.00045580130739098163,0.07586155319213868,0.07628573913574219,0.07674357376098632,0.07710984146118163,"[0.07720140838623046, 0.07585977935791016, 0.07539791870117188, 0.0760009307861328, 0.0758256607055664, 0.07575039672851562, 0.07618399810791016, 0.0758633270263672, 0.07592291259765625, 0.07566598510742187]",tokens/s,3369.874011479358,kWh,2.338612180100002e-06,2.5782695515844666e-07,1.5509145740638663e-06,4.1473537093223155e-06,tokens/kWh,61726107.282475024,MB,1478.69696,1798.242304,0.0,1390.411776,1377.233408,s,10,11.5315078125,1.15315078125,0.009420732157273685,1.1547311401367186,1.1652214721679686,1.1661131774902342,1.1668265417480468,"[1.1529599609375, 1.1565023193359374, 1.15901416015625, 1.14597998046875, 1.160075439453125, 1.1650233154296874, 1.1398558349609376, 1.1445345458984375, 1.140557373046875, 1.1670048828125]",tokens/s,54.63292487363088,kWh,3.392155081198458e-05,3.7411917676915357e-06,1.8114184491337117e-05,5.5776927071013235e-05,tokens/kWh,1129499.2985144306,,s,630,11.52871917533875,0.01829955424656944,0.00045683104943365426,0.018227952003479004,0.018644780540466307,0.01881768684387207,0.01958206460952759,"[0.01893868827819824, 0.01852969551086426, 0.01835683250427246, 0.01823539161682129, 0.018175136566162108, 0.018379615783691405, 0.018335744857788085, 0.018201791763305664, 0.01855286407470703, 0.01823004722595215, 0.0183540153503418, 0.01809174346923828, 0.018112159729003905, 0.01813987159729004, 0.018210432052612305, 0.01826144027709961, 0.018387935638427735, 0.018468416213989258, 0.018586111068725587, 0.01854876708984375, 0.018660703659057618, 0.018496095657348634, 0.018388959884643556, 0.018207904815673828, 0.018201471328735352, 0.01817094421386719, 0.018080127716064452, 0.018104896545410157, 0.018210784912109375, 0.018325504302978517, 0.018268192291259765, 0.018257919311523436, 0.0183767032623291, 0.018130720138549803, 0.018241151809692383, 0.01851247978210449, 0.018593599319458008, 0.01833184051513672, 0.018118656158447266, 0.018137088775634767, 0.018144287109375, 0.01825606346130371, 0.018209823608398436, 0.018428831100463866, 0.018368480682373046, 0.018453535079956056, 0.0184006404876709, 0.01828713607788086, 0.018186176300048828, 0.018067455291748045, 0.018284000396728516, 0.018170400619506834, 0.018324480056762696, 0.018227359771728516, 0.018041343688964845, 0.01822329521179199, 0.018104736328125, 0.01863065528869629, 0.01828611183166504, 0.0182192325592041, 0.01818364715576172, 0.01808195114135742, 0.018226879119873047, 0.019005727767944337, 0.018481632232666016, 0.01836851119995117, 0.018163808822631834, 0.018136959075927734, 0.01804729652404785, 0.018044384002685546, 0.01805958366394043, 0.018010208129882813, 0.017913631439208984, 0.018114816665649413, 0.018016063690185546, 0.018046783447265624, 0.01803023910522461, 0.018024864196777343, 0.017958080291748047, 0.018052032470703125, 0.01816985511779785, 0.01806540870666504, 0.018166048049926758, 0.018392799377441406, 0.018157535552978516, 0.018128927230834962, 0.01826963233947754, 0.018468608856201173, 0.018565952301025392, 0.018550079345703126, 0.018444992065429686, 0.018362272262573243, 0.018472320556640626, 0.0184051513671875, 0.018367424011230468, 0.01843561553955078, 0.018448863983154297, 0.018548864364624024, 0.018544031143188477, 0.018531808853149413, 0.018412544250488282, 0.018390495300292967, 0.018288639068603514, 0.01833782386779785, 0.018622079849243165, 0.020173696517944335, 0.019198207855224608, 0.018476800918579103, 0.018378751754760742, 0.01834102439880371, 0.01818614387512207, 0.018111423492431642, 0.018073888778686525, 0.018521600723266602, 0.01811091232299805, 0.01834364891052246, 0.018353696823120116, 0.018184511184692383, 0.018311391830444335, 0.018529600143432617, 0.018982751846313477, 0.018254688262939453, 0.018228992462158203, 0.018249984741210937, 0.01846886444091797, 0.018679807662963867, 0.019154943466186524, 0.018548959732055663, 0.018499488830566405, 0.018371583938598633, 0.018312063217163086, 0.018302463531494142, 0.018051584243774413, 0.01840460777282715, 0.01813372802734375, 0.018313119888305664, 0.018493663787841796, 0.018943904876708984, 0.018504735946655273, 0.01836031913757324, 0.01839411163330078, 0.018332735061645507, 0.018389312744140626, 0.01848975944519043, 0.018333120346069334, 0.01854038429260254, 0.01877903938293457, 0.018659328460693358, 0.018716447830200194, 0.018723039627075194, 0.01857535934448242, 0.018485248565673826, 0.018443872451782226, 0.01832547187805176, 0.018269920349121095, 0.01915763282775879, 0.01824777603149414, 0.018369855880737303, 0.018362783432006837, 0.01834217643737793, 0.01819647979736328, 0.018095327377319337, 0.018115007400512695, 0.01837295913696289, 0.0182541446685791, 0.018261568069458008, 0.018092159271240235, 0.01807155227661133, 0.018106367111206053, 0.018288511276245117, 0.01811782455444336, 0.018234304428100586, 0.018374048233032226, 0.018306688308715822, 0.018318304061889647, 0.0185380802154541, 0.01846067237854004, 0.018338560104370117, 0.018158975601196288, 0.018124128341674806, 0.01818899154663086, 0.018298912048339843, 0.01833750343322754, 0.01852592086791992, 0.018377119064331055, 0.01849555206298828, 0.018307584762573242, 0.018675039291381836, 0.01838947105407715, 0.019040288925170897, 0.018362720489501952, 0.018059423446655273, 0.017892416000366212, 0.018026784896850587, 0.017877695083618163, 0.01802851104736328, 0.018190336227416993, 0.018253503799438478, 0.01834623908996582, 0.018190336227416993, 0.018413631439208985, 0.018094079971313477, 0.01805923271179199, 0.018167360305786133, 0.01812224006652832, 0.018082176208496094, 0.01909987258911133, 0.01821120071411133, 0.018724863052368163, 0.01818556785583496, 0.017995424270629883, 0.018092960357666017, 0.01805936050415039, 0.018335968017578124, 0.01843791961669922, 0.018440288543701173, 0.018239231109619142, 0.018012319564819336, 0.0179682559967041, 0.017965951919555665, 0.01800396728515625, 0.017971200942993162, 0.017960960388183594, 0.01804310417175293, 0.0179648323059082, 0.017919424057006837, 0.01816428756713867, 0.01819036865234375, 0.019322080612182616, 0.018115327835083007, 0.01809939193725586, 0.01809222412109375, 0.017968767166137694, 0.017947647094726564, 0.017953887939453125, 0.018060192108154297, 0.018118656158447266, 0.018051071166992186, 0.01839468765258789, 0.01802079963684082, 0.01800396728515625, 0.018415903091430662, 0.01857315254211426, 0.018267072677612305, 0.017988224029541016, 0.01804128074645996, 0.018056800842285156, 0.018015775680541992, 0.018123552322387694, 0.01812272071838379, 0.01828659248352051, 0.018501632690429686, 0.018251808166503906, 0.01806528091430664, 0.018141183853149414, 0.018040000915527345, 0.01809596824645996, 0.01894931221008301, 0.01799760055541992, 0.017888864517211913, 0.01794483184814453, 0.01795907211303711, 0.018077695846557617, 0.018177600860595704, 0.018041343688964845, 0.017995552062988283, 0.018223264694213866, 0.022200319290161134, 0.018237791061401366, 0.018072736740112304, 0.017926912307739257, 0.017958335876464844, 0.017944992065429686, 0.017932191848754882, 0.0180316162109375, 0.018151872634887694, 0.01791231918334961, 0.017846336364746095, 0.017964607238769532, 0.017932832717895506, 0.01800387191772461, 0.017983327865600585, 0.018106496810913086, 0.018120672225952147, 0.018210880279541014, 0.01826201629638672, 0.018265951156616212, 0.01876915168762207, 0.01852841567993164, 0.018698591232299805, 0.01859350395202637, 0.018702495574951173, 0.018819616317749022, 0.018661376953125, 0.019079168319702147, 0.018597055435180664, 0.018420543670654297, 0.01847500801086426, 0.01845039939880371, 0.01841155242919922, 0.018432096481323244, 0.01848054313659668, 0.018541055679321287, 0.01859993553161621, 0.018644895553588867, 0.018575456619262694, 0.018761375427246093, 0.018737503051757812, 0.018810911178588866, 0.01887228775024414, 0.01864089584350586, 0.01852796745300293, 0.01949929618835449, 0.018737087249755858, 0.018699424743652344, 0.01930259132385254, 0.018788223266601563, 0.018854719161987304, 0.018520063400268554, 0.018321407318115233, 0.018287872314453123, 0.018371328353881836, 0.01827020835876465, 0.01842585563659668, 0.018483072280883788, 0.01860544013977051, 0.018405567169189452, 0.018520639419555663, 0.018582559585571288, 0.018461824417114258, 0.018478431701660157, 0.018500095367431642, 0.01843587112426758, 0.019046144485473634, 0.01859452819824219, 0.018619327545166015, 0.018455360412597658, 0.018774015426635742, 0.018397184371948243, 0.018693536758422852, 0.01837526321411133, 0.018369951248168946, 0.018779903411865233, 0.018371423721313476, 0.018247615814208983, 0.0183255672454834, 0.0183287353515625, 0.018207584381103516, 0.01816166305541992, 0.018206720352172853, 0.018094079971313477, 0.018183456420898438, 0.018278367996215822, 0.01839142417907715, 0.018384544372558594, 0.018416128158569335, 0.01862403106689453, 0.01871241569519043, 0.018618431091308594, 0.01864169692993164, 0.01885593605041504, 0.01868185615539551, 0.01961587142944336, 0.01871164894104004, 0.018772768020629882, 0.018709728240966797, 0.018587520599365234, 0.018541471481323242, 0.018325504302978517, 0.018328832626342773, 0.018102815628051758, 0.018120576858520508, 0.01840982437133789, 0.018268032073974608, 0.01819251251220703, 0.018317087173461914, 0.01822854423522949, 0.018090911865234375, 0.018670272827148438, 0.018108415603637695, 0.017971200942993162, 0.018051071166992186, 0.018129087448120116, 0.018178911209106446, 0.018227264404296874, 0.01818707275390625, 0.01821295928955078, 0.018577407836914063, 0.01846886444091797, 0.01847871971130371, 0.01813542366027832, 0.018026399612426757, 0.017997919082641603, 0.018189632415771484, 0.01799616050720215, 0.017895744323730468, 0.017881088256835938, 0.017969152450561524, 0.01798761558532715, 0.01786851119995117, 0.017940511703491212, 0.0178874568939209, 0.017914112091064454, 0.018001951217651368, 0.018093439102172853, 0.017957216262817384, 0.017960031509399413, 0.018005983352661133, 0.01792095947265625, 0.01820057678222656, 0.01819647979736328, 0.017953855514526367, 0.018426111221313476, 0.01823577690124512, 0.017950687408447265, 0.017927967071533202, 0.017963680267333984, 0.01799567985534668, 0.017934335708618163, 0.017864736557006836, 0.018147296905517578, 0.017970815658569336, 0.01801411247253418, 0.017938911437988283, 0.018150943756103516, 0.01819593620300293, 0.018182687759399414, 0.018364896774291994, 0.018225151062011717, 0.018268159866333008, 0.018333696365356447, 0.018221120834350586, 0.01805427169799805, 0.018107200622558595, 0.01799577522277832, 0.018159551620483397, 0.017997888565063475, 0.017909759521484374, 0.01796505546569824, 0.017936384201049805, 0.017948320388793945, 0.01828518486022949, 0.018290815353393556, 0.018208864212036133, 0.018253599166870117, 0.017995744705200194, 0.017936416625976562, 0.018081792831420897, 0.018054847717285157, 0.018116928100585936, 0.018118207931518554, 0.018148895263671874, 0.01808259201049805, 0.018068960189819336, 0.018030399322509764, 0.0180163516998291, 0.01810470390319824, 0.018121055603027344, 0.019494943618774414, 0.01805459213256836, 0.01804083251953125, 0.017886816024780275, 0.017984447479248048, 0.017907039642333984, 0.018989759445190428, 0.01799750328063965, 0.01786092758178711, 0.01800934410095215, 0.017922815322875978, 0.017886783599853514, 0.017930688858032225, 0.018050912857055665, 0.018001279830932616, 0.018067583084106446, 0.017971872329711914, 0.017936384201049805, 0.017884767532348633, 0.01790332794189453, 0.017936256408691405, 0.01809676742553711, 0.01794272041320801, 0.017998912811279296, 0.018514879226684572, 0.02188083267211914, 0.01854080009460449, 0.018165599822998046, 0.018072864532470704, 0.01806604766845703, 0.017985536575317384, 0.0179704647064209, 0.017991487503051757, 0.018006656646728517, 0.01789571189880371, 0.017953792572021485, 0.01797427177429199, 0.01796816062927246, 0.01825276756286621, 0.01794047927856445, 0.01810963249206543, 0.018824064254760742, 0.018111616134643554, 0.018407712936401366, 0.017971744537353517, 0.01801625633239746, 0.018270751953125, 0.018038368225097655, 0.017971616744995117, 0.018164863586425783, 0.018176671981811523, 0.018462944030761718, 0.01807792091369629, 0.017938207626342774, 0.01790937614440918, 0.018008447647094725, 0.017926143646240233, 0.018950111389160158, 0.018065439224243165, 0.017858015060424805, 0.017951263427734374, 0.018188512802124024, 0.017895200729370116, 0.017979263305664062, 0.017967231750488283, 0.017977344512939454, 0.017896799087524413, 0.017998432159423827, 0.017825855255126952, 0.017911455154418946, 0.017947200775146485, 0.01786400032043457, 0.018033119201660158, 0.017930240631103517, 0.018210336685180663, 0.01796748733520508, 0.018079839706420898, 0.0179587516784668, 0.01791606330871582, 0.01804697608947754, 0.017967103958129883, 0.017920000076293945, 0.017974592208862303, 0.01801491165161133, 0.018031776428222655, 0.018084096908569335, 0.018031200408935546, 0.01804841613769531, 0.018045343399047852, 0.018411712646484377, 0.01804697608947754, 0.01815884780883789, 0.018417631149291992, 0.01819523239135742, 0.01806723213195801, 0.0181409912109375, 0.017942335128784178, 0.01797100830078125, 0.018002719879150392, 0.01799318313598633, 0.01809388732910156, 0.01815353584289551, 0.018179967880249025, 0.0184737606048584, 0.01834307289123535, 0.018435136795043945, 0.018644767761230467, 0.018521440505981444, 0.018516639709472656, 0.01913862419128418, 0.018613855361938478, 0.01867407989501953, 0.01864499282836914, 0.018528255462646484, 0.01851408004760742, 0.01835811233520508, 0.018355520248413085, 0.018325279235839844, 0.018205215454101562, 0.01829311943054199, 0.018305023193359374, 0.018321407318115233, 0.01827840042114258, 0.01825200080871582, 0.018302431106567384, 0.018333343505859374, 0.018328096389770506, 0.018335872650146485, 0.018225151062011717, 0.018159872055053712, 0.018091775894165038, 0.018118112564086915, 0.018188831329345703, 0.018247840881347656, 0.01825161552429199, 0.018261184692382814, 0.01832601547241211, 0.018323776245117187, 0.01838863945007324, 0.01881532859802246, 0.018427104949951173, 0.018387744903564453, 0.018472032546997072, 0.01837910461425781, 0.018436767578125, 0.018511968612670897, 0.018451295852661132, 0.01842438316345215, 0.018325632095336913, 0.01832374382019043, 0.01834003257751465, 0.0184718074798584, 0.01834284782409668, 0.018380800247192384, 0.018301183700561524, 0.018263935089111328, 0.018323200225830078, 0.018110015869140624, 0.018074432373046876, 0.018046720504760742, 0.019159040451049804, 0.02500351905822754, 0.019855871200561523, 0.01854640007019043, 0.01830735969543457, 0.0183438720703125, 0.018473024368286132, 0.018317312240600587, 0.018255903244018556, 0.01825702476501465, 0.01821504020690918, 0.01971887969970703]",tokens/s,54.646139819906644,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1852.98944,2718.892032,0.0,2340.421632,2285.568,s,1,8.9999931640625,8.9999931640625,0.0,8.9999931640625,8.9999931640625,8.9999931640625,8.9999931640625,[8.9999931640625],,kWh,6.0278640858329404e-05,6.639440169627309e-06,2.0128627214038852e-05,8.704670824199556e-05,,MB,1854.042112,3067.019264,0.0,2659.188736,2578.855936,s,10,0.8156142654418945,0.08156142654418945,0.0005569991454410323,0.08144235229492187,0.0822944351196289,0.08255667190551758,0.08276646133422852,"[0.08281890869140625, 0.08223616027832031, 0.08115347290039063, 0.08161580657958985, 0.08172006225585937, 0.0813210220336914, 0.08085462188720703, 0.08117372894287109, 0.08156368255615234, 0.08115679931640625]",tokens/s,3138.7386273988336,kWh,2.5467738779711973e-06,2.808641281398764e-07,1.6995375915130268e-06,4.527175597624101e-06,tokens/kWh,56547397.92606033,MB,1854.042112,3067.019264,0.0,2659.188736,2578.858496,s,10,14.036172607421873,1.4036172607421873,0.010475848935466391,1.3988833007812498,1.4191065185546874,1.4217946533203125,1.4239451611328124,"[1.4011680908203126, 1.3940382080078124, 1.39739501953125, 1.3950809326171876, 1.3964273681640624, 1.41350537109375, 1.4185091552734375, 1.4244827880859374, 1.395194091796875, 1.40037158203125]",tokens/s,44.88403054169314,kWh,4.119195426994443e-05,4.543112637616315e-06,2.2239787115484975e-05,6.797485402304572e-05,tokens/kWh,926813.3180343561,,s,630,14.03415205383302,0.022276431831480967,0.00047671132272992823,0.02214495944976807,0.022801987648010252,0.02292944688796997,0.023775007877349853,"[0.022978527069091797, 0.022815519332885743, 0.022460416793823244, 0.02232851219177246, 0.02240185546875, 0.022271999359130858, 0.022429088592529296, 0.0220446720123291, 0.022266016006469727, 0.02223967933654785, 0.022103071212768555, 0.022101184844970704, 0.022183712005615235, 0.02215692710876465, 0.022425344467163086, 0.02214259147644043, 0.022202495574951173, 0.022119295120239257, 0.02210201644897461, 0.02205427169799805, 0.022106752395629883, 0.022091775894165038, 0.022079488754272462, 0.022080928802490234, 0.022153791427612306, 0.022102048873901367, 0.022066719055175783, 0.02200828742980957, 0.022009855270385743, 0.021925888061523437, 0.022019807815551757, 0.02198739242553711, 0.022035743713378905, 0.022084352493286132, 0.02203887939453125, 0.022038368225097655, 0.022169599533081053, 0.024182559967041016, 0.0225784969329834, 0.022270879745483398, 0.02220796775817871, 0.022065696716308595, 0.022237184524536133, 0.0223191032409668, 0.022058784484863283, 0.022103296279907227, 0.022455263137817382, 0.022120704650878908, 0.022232704162597657, 0.022047008514404297, 0.022084543228149414, 0.02201433563232422, 0.021987871170043947, 0.021950464248657226, 0.021961856842041015, 0.021985279083251954, 0.021957632064819335, 0.024041343688964843, 0.022814720153808594, 0.02228223991394043, 0.022108160018920898, 0.021980352401733398, 0.02208799934387207, 0.02360540771484375, 0.022427648544311524, 0.022067455291748046, 0.02206185531616211, 0.022094911575317382, 0.021919647216796876, 0.021948415756225585, 0.02192595291137695, 0.02208291244506836, 0.021905664443969727, 0.021883167266845704, 0.022052127838134764, 0.021996320724487303, 0.021906591415405272, 0.02196156883239746, 0.02208358383178711, 0.02186649513244629, 0.021835775375366212, 0.021950239181518554, 0.022014272689819335, 0.022022048950195314, 0.02198102378845215, 0.02191360092163086, 0.02196291160583496, 0.022441568374633788, 0.023889728546142578, 0.022132383346557618, 0.022121280670166017, 0.022049951553344726, 0.022013952255249023, 0.022124576568603515, 0.022024192810058595, 0.02197599983215332, 0.022042367935180666, 0.021999103546142578, 0.022010623931884767, 0.022022144317626953, 0.02206003189086914, 0.022157951354980467, 0.02198361587524414, 0.02245792007446289, 0.0230830078125, 0.022022592544555665, 0.02202009582519531, 0.022011072158813476, 0.0219881591796875, 0.021970943450927736, 0.02198262405395508, 0.021915679931640626, 0.02203500747680664, 0.02194780731201172, 0.021989183425903322, 0.021990175247192382, 0.022239231109619142, 0.022123519897460937, 0.021963775634765623, 0.02189516830444336, 0.022195520401000975, 0.022223104476928712, 0.022208511352539064, 0.022458816528320314, 0.02223014450073242, 0.022346624374389647, 0.02329187202453613, 0.02265056037902832, 0.022506464004516603, 0.023739360809326173, 0.022272895812988282, 0.02214249610900879, 0.022079391479492186, 0.022249887466430664, 0.02240323257446289, 0.022210559844970702, 0.022009952545166016, 0.022169504165649414, 0.02198441505432129, 0.02194313621520996, 0.021942176818847657, 0.021997663497924806, 0.022396928787231447, 0.021961759567260743, 0.02242681694030762, 0.022103296279907227, 0.02215609550476074, 0.02236579132080078, 0.022354047775268556, 0.023201791763305665, 0.02264860725402832, 0.022644960403442382, 0.02215337562561035, 0.022083423614501954, 0.021952512741088868, 0.022132736206054687, 0.021929983139038087, 0.0224333438873291, 0.0218853759765625, 0.021940223693847655, 0.022052288055419922, 0.022032960891723633, 0.021905248641967773, 0.021978368759155275, 0.022180992126464842, 0.022054624557495118, 0.021978303909301757, 0.022059904098510743, 0.021952512741088868, 0.022073183059692383, 0.021968128204345704, 0.02206528091430664, 0.022053407669067382, 0.02207155227661133, 0.02196623992919922, 0.022046655654907227, 0.02209244728088379, 0.022108160018920898, 0.022075231552124024, 0.02203664016723633, 0.021963808059692384, 0.021941247940063476, 0.02196611213684082, 0.022060831069946288, 0.021969535827636718, 0.022040063858032227, 0.02202908706665039, 0.022134016036987305, 0.021971296310424805, 0.022476480484008788, 0.022200000762939452, 0.022075040817260742, 0.022130687713623046, 0.022262752532958983, 0.02209708786010742, 0.02211027145385742, 0.022113216400146483, 0.022112064361572266, 0.02208140754699707, 0.022130815505981446, 0.022038591384887694, 0.022140863418579102, 0.022145023345947267, 0.022013856887817384, 0.022320831298828125, 0.02235638427734375, 0.02229248046875, 0.02233344078063965, 0.022196224212646484, 0.022242399215698243, 0.022167648315429687, 0.022051647186279298, 0.022009471893310546, 0.02203094482421875, 0.022064224243164062, 0.022051519393920898, 0.021984832763671875, 0.022065216064453125, 0.022097663879394533, 0.022095680236816406, 0.02208675193786621, 0.022042335510253905, 0.021966848373413086, 0.02196672058105469, 0.02195686340332031, 0.022011775970458985, 0.021964351654052736, 0.021979583740234374, 0.022062591552734375, 0.022813215255737304, 0.02247235107421875, 0.022403392791748047, 0.02227737617492676, 0.022352640151977538, 0.022128480911254883, 0.022104223251342772, 0.0233123836517334, 0.022230016708374024, 0.02205183982849121, 0.022015615463256834, 0.022033887863159178, 0.02202716827392578, 0.021949920654296875, 0.021985439300537108, 0.021987712860107422, 0.02205414390563965, 0.022042560577392577, 0.021963647842407227, 0.022048704147338866, 0.02196403121948242, 0.022051584243774413, 0.022079488754272462, 0.022212831497192383, 0.021987327575683592, 0.021950464248657226, 0.02205900764465332, 0.021979135513305666, 0.022027679443359375, 0.021874784469604492, 0.02201036834716797, 0.022052608489990234, 0.022061311721801757, 0.02196611213684082, 0.021952991485595704, 0.022015615463256834, 0.021983903884887697, 0.021988704681396486, 0.021985919952392578, 0.02198089599609375, 0.02202448081970215, 0.022039968490600585, 0.022047359466552733, 0.022077407836914063, 0.022187807083129882, 0.021991296768188475, 0.02226211166381836, 0.021977088928222657, 0.022103168487548827, 0.022153823852539063, 0.022190559387207032, 0.02256057548522949, 0.02224332809448242, 0.022239231109619142, 0.022222848892211915, 0.02223414421081543, 0.022260704040527345, 0.022175743103027345, 0.022245376586914063, 0.02225152015686035, 0.022259775161743163, 0.02216441535949707, 0.02229574394226074, 0.022230335235595703, 0.022225055694580078, 0.022225248336791993, 0.02226585578918457, 0.022271999359130858, 0.02240675163269043, 0.022299039840698243, 0.022269760131835938, 0.022216096878051757, 0.02227280044555664, 0.022245376586914063, 0.022185983657836913, 0.02240208053588867, 0.02247923278808594, 0.022264415740966798, 0.022308095932006836, 0.02216783905029297, 0.02223369598388672, 0.022081472396850585, 0.022187711715698243, 0.022254911422729493, 0.022207424163818358, 0.022230239868164064, 0.022770240783691408, 0.022519712448120118, 0.022603071212768555, 0.022847999572753908, 0.023460351943969726, 0.022923040390014648, 0.022892576217651367, 0.02285385513305664, 0.022826080322265626, 0.022919231414794922, 0.022911584854125977, 0.02287001609802246, 0.022863712310791016, 0.022853504180908202, 0.02280415916442871, 0.02281942367553711, 0.022877792358398437, 0.023044416427612305, 0.022865280151367188, 0.022847679138183592, 0.02269059181213379, 0.022660928726196287, 0.022517696380615234, 0.022463680267333985, 0.022493120193481445, 0.022463455200195312, 0.022406143188476564, 0.022219680786132814, 0.022166624069213867, 0.022152095794677733, 0.02211020851135254, 0.022551904678344725, 0.0222740478515625, 0.02225833511352539, 0.022163551330566408, 0.02307619285583496, 0.023293920516967773, 0.02234569549560547, 0.022149152755737304, 0.02244207954406738, 0.022043136596679686, 0.02245631980895996, 0.022114303588867186, 0.022130687713623046, 0.02207855987548828, 0.022048831939697266, 0.022025056838989258, 0.021992576599121093, 0.022133216857910157, 0.02195702362060547, 0.02205900764465332, 0.021935136795043945, 0.022060096740722655, 0.021941728591918945, 0.02191404724121094, 0.02198739242553711, 0.021944255828857423, 0.022066303253173828, 0.022012191772460936, 0.02185647964477539, 0.022050336837768556, 0.022180704116821288, 0.022054367065429688, 0.022268352508544923, 0.022316991806030275, 0.023569952011108397, 0.022217632293701172, 0.022148895263671874, 0.022025951385498045, 0.022292543411254882, 0.022003936767578124, 0.022048063278198242, 0.0221624641418457, 0.02220796775817871, 0.022541791915893554, 0.02244384002685547, 0.022375328063964844, 0.0222873592376709, 0.02214137649536133, 0.02209030342102051, 0.02202150344848633, 0.022061504364013673, 0.021999807357788087, 0.022095296859741213, 0.02201398468017578, 0.0220677433013916, 0.022068256378173827, 0.0221778564453125, 0.022344608306884766, 0.02225971221923828, 0.02280182456970215, 0.022444831848144532, 0.02221036720275879, 0.022185888290405274, 0.02207139205932617, 0.022278144836425783, 0.02274070358276367, 0.022136287689208986, 0.022289215087890626, 0.02249497604370117, 0.022813983917236328, 0.02289731216430664, 0.022854015350341796, 0.022910783767700196, 0.02284783935546875, 0.02290870475769043, 0.022862911224365234, 0.02295235252380371, 0.022854175567626953, 0.023072639465332032, 0.022934688568115234, 0.023055328369140624, 0.022999040603637694, 0.02296124839782715, 0.022803455352783202, 0.02287171173095703, 0.022907295227050782, 0.0233654727935791, 0.023399871826171877, 0.02294822311401367, 0.022939424514770507, 0.02272275161743164, 0.022579519271850586, 0.022491039276123045, 0.02231817626953125, 0.02212895965576172, 0.022753471374511718, 0.022361600875854492, 0.02215372848510742, 0.022312576293945313, 0.022427040100097655, 0.022563327789306642, 0.022598112106323242, 0.022527999877929687, 0.02244764709472656, 0.02245884895324707, 0.022398464202880858, 0.022442495346069336, 0.022425600051879883, 0.0227061767578125, 0.0224783992767334, 0.022564416885375978, 0.022391679763793946, 0.022404991149902343, 0.022395008087158202, 0.02466534423828125, 0.0234747200012207, 0.023050079345703123, 0.02240105628967285, 0.022503583908081055, 0.02238387107849121, 0.022268512725830077, 0.022565343856811523, 0.022193119049072264, 0.022148000717163087, 0.022278144836425783, 0.022078527450561523, 0.02200876808166504, 0.022025760650634767, 0.02253811264038086, 0.02223369598388672, 0.02216985511779785, 0.0221909122467041, 0.022447296142578125, 0.022636287689208983, 0.023883775711059572, 0.023789567947387694, 0.02242745590209961, 0.02249951934814453, 0.022461759567260743, 0.022617887496948243, 0.022195104598999024, 0.02221897506713867, 0.02207107162475586, 0.022140928268432617, 0.022814111709594728, 0.02238070487976074, 0.022190656661987305, 0.02232268714904785, 0.022104448318481445, 0.02206105613708496, 0.0301711368560791, 0.02364739227294922, 0.022293344497680664, 0.02216160011291504, 0.02218943977355957, 0.022217151641845703, 0.022212608337402344, 0.022146207809448242, 0.022253568649291993, 0.02207539176940918, 0.022106016159057617, 0.022116031646728516, 0.022301279067993163, 0.022027711868286132, 0.02218227195739746, 0.022054912567138672, 0.0220437126159668, 0.02235078430175781, 0.022096960067749024, 0.022547487258911134, 0.022253471374511717, 0.022337535858154296, 0.022006847381591796, 0.022119359970092775, 0.022054880142211915, 0.022052896499633788, 0.02206924819946289, 0.02206515121459961, 0.022079488754272462, 0.022026239395141603, 0.021966400146484374, 0.02220691108703613, 0.022911136627197265, 0.022144895553588867, 0.022196191787719727, 0.022079488754272462, 0.022137855529785155, 0.022098943710327147, 0.022034431457519533, 0.02200371170043945, 0.0220897274017334, 0.022044927597045898, 0.022013696670532226, 0.022052160263061525, 0.022036767959594725, 0.022054912567138672, 0.022079904556274413, 0.021954559326171876, 0.02208768081665039, 0.022063104629516602, 0.0220897274017334, 0.022150304794311522, 0.022194656372070312, 0.022288063049316405, 0.022305471420288086, 0.02221993637084961, 0.022096736907958985, 0.022117855072021485, 0.02228278350830078, 0.02208780860900879, 0.02201900863647461, 0.02229088020324707, 0.022056543350219726, 0.022100448608398438, 0.02211680030822754, 0.022166784286499024, 0.02230928039550781, 0.022063455581665038, 0.022200319290161134, 0.02206835174560547, 0.022296768188476562, 0.022257919311523437, 0.02212441635131836, 0.022021984100341795, 0.02200214385986328, 0.022062047958374024, 0.022364799499511718, 0.02235811233520508, 0.022077312469482423, 0.022198400497436523, 0.022163455963134765, 0.022344831466674805, 0.022186368942260743, 0.022254079818725587, 0.022126399993896484, 0.02209401512145996, 0.02197427177429199, 0.02208345603942871, 0.02196361541748047, 0.022056991577148438, 0.022011808395385742, 0.022097312927246093, 0.02194486427307129, 0.022060863494873045, 0.021979488372802735, 0.022199935913085937, 0.022048255920410157, 0.022065183639526368, 0.022082399368286133, 0.022264896392822267, 0.022094688415527343, 0.0221463680267334, 0.022004512786865233, 0.02220460891723633, 0.02205398368835449, 0.022139616012573242, 0.0220361270904541, 0.02247100830078125, 0.02211008071899414, 0.022032800674438476, 0.02199216079711914, 0.022021120071411132, 0.02208153533935547, 0.02226380729675293, 0.022349151611328125, 0.022780576705932618, 0.022439968109130858, 0.0225382080078125, 0.022476383209228516, 0.022524320602416992, 0.022487039566040038, 0.022547584533691406, 0.02246950340270996, 0.022759424209594727, 0.022568832397460936, 0.022508960723876953, 0.022482688903808595, 0.02211292839050293, 0.022006080627441405, 0.022173696517944336, 0.022206047058105468, 0.022407455444335936, 0.022564096450805662, 0.02263324737548828]",tokens/s,44.89049267696471,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3792, in from_pretrained config = cls._autoset_attn_implementation( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1531, in _autoset_attn_implementation config = cls._check_and_enable_sdpa( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1697, in _check_and_enable_sdpa raise ValueError( ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4242.948096,5784.928256,0.0,5389.68064,5000.446464,s,1,11.290736328125,11.290736328125,0.0,11.290736328125,11.290736328125,11.290736328125,11.290736328125,[11.290736328125],,kWh,0.000129021958274969,1.4224751504962389e-05,5.5991711460007254e-05,0.00019923842123993865,,MB,1386.84416,5801.705472,0.0,5393.874944,4706.596864,s,10,29.3511875,2.93511875,0.0039120314747161915,2.9365336914062503,2.9385712646484374,2.938667834472656,2.938745090332031,"[2.9254248046875, 2.931824951171875, 2.933055419921875, 2.9367763671875, 2.93485791015625, 2.936291015625, 2.938044189453125, 2.9375986328125, 2.938764404296875, 2.9385498046875]",tokens/s,87.21963975052117,kWh,8.559857491208428e-05,9.441424095514069e-06,5.6893462181398234e-05,0.0001519334611889966,tokens/kWh,1684948.1213460315,MB,1390.985216,5801.705472,0.0,5393.874944,4876.091904,s,10,18.76966467285156,1.8769664672851563,0.0130080446991286,1.8743575439453126,1.8922148071289062,1.8949106872558594,1.8970673913574219,"[1.876090087890625, 1.8976065673828124, 1.872625, 1.886764404296875, 1.8672735595703125, 1.864760009765625, 1.8615469970703125, 1.89161572265625, 1.8606595458984374, 1.8907227783203124]",tokens/s,33.56479782567622,kWh,5.547119228875015e-05,6.118897369139303e-06,3.672366826780183e-05,9.831375792569128e-05,tokens/kWh,640805.5325035733,,s,630,18.766633272171003,0.029788306781223843,0.0004898927835957104,0.029722991943359374,0.030298616790771483,0.030615639686584473,0.03148992485046387,"[0.03096950340270996, 0.03009971237182617, 0.029877632141113282, 0.02973155212402344, 0.02972640037536621, 0.029688127517700197, 0.029478912353515626, 0.029432863235473634, 0.029373311996459962, 0.029357471466064454, 0.02940764808654785, 0.02943132781982422, 0.02942848014831543, 0.02953411293029785, 0.029824447631835938, 0.02971023941040039, 0.02960051155090332, 0.02959769630432129, 0.02936627197265625, 0.029566303253173828, 0.029614271163940428, 0.02994428825378418, 0.029580959320068358, 0.029720672607421873, 0.030179391860961913, 0.029755584716796873, 0.029554655075073242, 0.02964892768859863, 0.02953388786315918, 0.029495328903198243, 0.029458719253540037, 0.029462528228759766, 0.029461727142333985, 0.029479711532592774, 0.029829120635986327, 0.029886463165283202, 0.029970432281494142, 0.029914432525634766, 0.029776063919067383, 0.029704704284667968, 0.029783424377441407, 0.030356096267700194, 0.02977340888977051, 0.029850015640258788, 0.030011392593383788, 0.030130176544189452, 0.029877376556396485, 0.029818784713745116, 0.030100032806396483, 0.029915168762207033, 0.02983359909057617, 0.029857215881347657, 0.029904575347900392, 0.02983616065979004, 0.029847551345825195, 0.02993961524963379, 0.029698144912719725, 0.02967296028137207, 0.029788448333740235, 0.030008991241455077, 0.030212671279907226, 0.030191328048706053, 0.030187231063842773, 0.03135897636413574, 0.030502912521362304, 0.030570240020751954, 0.030398719787597655, 0.030327999114990234, 0.03054470443725586, 0.030469696044921876, 0.03066268730163574, 0.030642463684082032, 0.030802047729492188, 0.03066166305541992, 0.030614496231079102, 0.030616575241088868, 0.030299135208129883, 0.032126976013183595, 0.031229951858520507, 0.030052352905273437, 0.030100704193115235, 0.029686559677124025, 0.030021631240844726, 0.029566976547241212, 0.029665279388427734, 0.029405183792114258, 0.02957926368713379, 0.031498239517211916, 0.03035308837890625, 0.029557056427001953, 0.029846944808959962, 0.02959971237182617, 0.032943904876708986, 0.030569311141967773, 0.029965568542480468, 0.029805311203002928, 0.029853696823120116, 0.029566303253173828, 0.029633119583129884, 0.030142528533935547, 0.029800447463989257, 0.029472639083862304, 0.02965894317626953, 0.02961235237121582, 0.029681663513183593, 0.029406944274902345, 0.02962460708618164, 0.02956287956237793, 0.029652639389038084, 0.02941129684448242, 0.029808223724365233, 0.02966966438293457, 0.029827583312988282, 0.029548320770263672, 0.02968160057067871, 0.029566719055175782, 0.02982761573791504, 0.030015487670898438, 0.029894655227661132, 0.02970195198059082, 0.02978220748901367, 0.02991923141479492, 0.030111743927001954, 0.030238719940185548, 0.03034316825866699, 0.030159872055053712, 0.030890207290649414, 0.029956832885742187, 0.029879295349121093, 0.029702943801879884, 0.029702367782592772, 0.029800447463989257, 0.029861888885498046, 0.02952396774291992, 0.029296319961547853, 0.029714752197265625, 0.02963206481933594, 0.029530559539794922, 0.029636608123779298, 0.02962784004211426, 0.029520448684692384, 0.029980287551879883, 0.029688192367553712, 0.029493215560913087, 0.029382688522338867, 0.02958336067199707, 0.02956598472595215, 0.029539295196533203, 0.029499391555786132, 0.02962985610961914, 0.02927881622314453, 0.029560352325439455, 0.029409759521484374, 0.029534048080444335, 0.02934377670288086, 0.029237056732177736, 0.029284223556518555, 0.029366720199584962, 0.02939904022216797, 0.030125375747680663, 0.029455039978027345, 0.02959769630432129, 0.029564544677734374, 0.02948748779296875, 0.02968272018432617, 0.029496288299560545, 0.029392608642578123, 0.029499679565429687, 0.029429759979248047, 0.029572736740112304, 0.029713855743408204, 0.03208287811279297, 0.02960383987426758, 0.029594688415527343, 0.030145471572875976, 0.02990880012512207, 0.029892799377441406, 0.029881856918334962, 0.029942272186279296, 0.03026095962524414, 0.029951360702514647, 0.029805471420288086, 0.029875648498535155, 0.029905471801757812, 0.029892608642578124, 0.02984502410888672, 0.029669855117797853, 0.03002899169921875, 0.029940351486206055, 0.031337087631225585, 0.029970176696777345, 0.029931840896606447, 0.0302838077545166, 0.029593599319458007, 0.029812736511230467, 0.029747200012207032, 0.029900800704956054, 0.03031449508666992, 0.03041823959350586, 0.030325439453125, 0.030324735641479493, 0.030227807998657225, 0.030114463806152344, 0.030212095260620117, 0.030208000183105467, 0.030402559280395508, 0.030283008575439453, 0.03036851119995117, 0.030445568084716795, 0.030334047317504883, 0.03019254493713379, 0.030552064895629883, 0.03001747131347656, 0.02990246391296387, 0.02982956886291504, 0.0299233283996582, 0.03062579154968262, 0.029724672317504884, 0.029708127975463867, 0.029423776626586913, 0.02969375991821289, 0.029675264358520508, 0.030298559188842774, 0.02972876739501953, 0.029772863388061524, 0.02967033576965332, 0.0295731201171875, 0.0332861442565918, 0.029284351348876952, 0.029362112045288085, 0.030754144668579102, 0.02983919906616211, 0.029365119934082032, 0.02982707214355469, 0.029826879501342773, 0.029724159240722657, 0.029651647567749025, 0.029482656478881836, 0.02964240074157715, 0.02953696060180664, 0.029739007949829102, 0.02975849533081055, 0.02961712074279785, 0.02933964729309082, 0.029646848678588866, 0.029652992248535157, 0.029431264877319337, 0.02939753532409668, 0.029394336700439453, 0.029358688354492186, 0.02939289665222168, 0.029315071105957033, 0.03098467254638672, 0.029911327362060546, 0.029859840393066408, 0.029912927627563476, 0.029708063125610352, 0.029405120849609376, 0.02958585548400879, 0.02982707214355469, 0.029876224517822264, 0.02968288040161133, 0.02955961608886719, 0.029583200454711914, 0.029237407684326172, 0.029198335647583007, 0.029222911834716796, 0.029158687591552733, 0.02956723213195801, 0.029333984375, 0.029612031936645508, 0.02979840087890625, 0.02978201675415039, 0.02973695945739746, 0.030238719940185548, 0.029808639526367187, 0.02974070358276367, 0.02972003173828125, 0.02973731231689453, 0.029632896423339845, 0.029879680633544924, 0.029749343872070313, 0.0297192325592041, 0.030117887496948242, 0.029822975158691405, 0.029652639389038084, 0.029783456802368165, 0.029766592025756836, 0.029707263946533204, 0.029674495697021484, 0.02964409637451172, 0.029340351104736328, 0.02939084815979004, 0.029286399841308593, 0.02944000053405762, 0.029400127410888672, 0.02967238426208496, 0.029773344039916994, 0.029415903091430665, 0.029419519424438476, 0.029698047637939453, 0.029405183792114258, 0.029310976028442383, 0.029258783340454102, 0.02959459114074707, 0.02954854393005371, 0.029460479736328125, 0.02966691207885742, 0.029484735488891602, 0.029403871536254882, 0.029607936859130858, 0.029716480255126954, 0.029501440048217774, 0.0295280647277832, 0.029726303100585938, 0.030853151321411133, 0.029883615493774413, 0.029442655563354493, 0.029151424407958985, 0.029171712875366212, 0.029057024002075195, 0.029255327224731446, 0.02918844795227051, 0.0293621768951416, 0.029247488021850586, 0.02918796730041504, 0.02985487937927246, 0.02961302375793457, 0.02942060852050781, 0.029385663986206054, 0.029255327224731446, 0.029624671936035157, 0.029284351348876952, 0.02939289665222168, 0.029136159896850585, 0.0291232967376709, 0.029159423828125, 0.02931219291687012, 0.029299520492553712, 0.029270015716552734, 0.02919161605834961, 0.02924323272705078, 0.02916012763977051, 0.02932329559326172, 0.029521472930908205, 0.02920697593688965, 0.029454336166381836, 0.02927804756164551, 0.029296607971191407, 0.02933884811401367, 0.029554655075073242, 0.029312000274658204, 0.029400447845458984, 0.02935366439819336, 0.0294532470703125, 0.03081625556945801, 0.030511104583740234, 0.029421472549438478, 0.029513824462890626, 0.029353183746337892, 0.029487903594970704, 0.029321216583251954, 0.02960383987426758, 0.03000115203857422, 0.029868032455444334, 0.029829120635986327, 0.029673471450805664, 0.03219660949707031, 0.0295031681060791, 0.02977619171142578, 0.03146956825256347, 0.02997987174987793, 0.02957801628112793, 0.029691520690917968, 0.02970252799987793, 0.02981888008117676, 0.029999103546142578, 0.03035136032104492, 0.030580543518066407, 0.029860063552856444, 0.02958847999572754, 0.029158336639404297, 0.029320480346679688, 0.029205280303955077, 0.029267936706542968, 0.02910825538635254, 0.02918764877319336, 0.029170112609863283, 0.0293353271484375, 0.029296863555908204, 0.029435903549194335, 0.029288448333740235, 0.029367616653442383, 0.029307199478149415, 0.029347232818603516, 0.029346784591674804, 0.029419519424438476, 0.02940015983581543, 0.02955548858642578, 0.02935737609863281, 0.029406015396118163, 0.029179231643676758, 0.029278879165649415, 0.029410528182983398, 0.029233951568603516, 0.029152544021606445, 0.02926665687561035, 0.02934169578552246, 0.029290496826171877, 0.0294072322845459, 0.02935603141784668, 0.029582656860351563, 0.02929270362854004, 0.0306376953125, 0.02951875114440918, 0.029343360900878905, 0.029548095703125, 0.029406015396118163, 0.0295546875, 0.029812511444091798, 0.02999932861328125, 0.02982268714904785, 0.029860031127929686, 0.029743200302124025, 0.029814783096313476, 0.029812736511230467, 0.02979840087890625, 0.02980454444885254, 0.029782112121582032, 0.02979827117919922, 0.029855712890625, 0.030183488845825196, 0.030074880599975585, 0.029681663513183593, 0.029722623825073242, 0.02933884811401367, 0.02958620834350586, 0.029915136337280275, 0.029693952560424806, 0.0295133113861084, 0.029555103302001954, 0.031035104751586915, 0.029963808059692384, 0.029759328842163087, 0.029594112396240234, 0.029593088150024413, 0.030112672805786132, 0.030044160842895507, 0.029569023132324217, 0.02976563262939453, 0.029896703720092774, 0.029949888229370118, 0.029511743545532227, 0.029607648849487304, 0.02967296028137207, 0.029884992599487306, 0.029917407989501953, 0.029898752212524415, 0.029906784057617188, 0.029979808807373047, 0.030766080856323243, 0.02988627243041992, 0.029914688110351563, 0.02981747245788574, 0.029792255401611328, 0.030043872833251953, 0.02980892753601074, 0.02976972770690918, 0.03005606460571289, 0.029867935180664062, 0.029706592559814452, 0.02960358428955078, 0.029552928924560545, 0.03001968002319336, 0.030195711135864257, 0.029841407775878907, 0.0301527042388916, 0.029746431350708008, 0.029757183074951173, 0.03013734436035156, 0.029998815536499024, 0.02984783935546875, 0.029906944274902345, 0.02980601692199707, 0.02983788871765137, 0.02983907127380371, 0.029829248428344727, 0.029763744354248046, 0.029837312698364257, 0.030271488189697264, 0.029894655227661132, 0.029924352645874022, 0.030122047424316407, 0.031144800186157225, 0.03132953643798828, 0.030155616760253905, 0.030158016204833986, 0.030061344146728515, 0.030371871948242188, 0.032761856079101564, 0.030304256439208983, 0.03014441680908203, 0.02999510383605957, 0.0299233283996582, 0.030814207077026368, 0.03004185676574707, 0.02946668815612793, 0.029333696365356446, 0.029273216247558593, 0.029151647567749024, 0.029328927993774415, 0.029170591354370116, 0.029194271087646485, 0.02926710319519043, 0.029227872848510743, 0.02922822380065918, 0.029299135208129882, 0.029328927993774415, 0.029682527542114256, 0.02919628715515137, 0.02918400001525879, 0.029478143692016602, 0.029324031829833983, 0.029237056732177736, 0.029236480712890624, 0.02968057632446289, 0.02923520088195801, 0.02940889549255371, 0.029622623443603516, 0.0294932804107666, 0.029313024520874024, 0.029310976028442383, 0.02947068786621094, 0.029284032821655273, 0.029828672409057618, 0.02934864044189453, 0.029205503463745116, 0.029459455490112304, 0.029442047119140623, 0.02960588836669922, 0.029437952041625977, 0.02920243263244629, 0.029207775115966797, 0.029480991363525392, 0.029569311141967772, 0.029862079620361328, 0.029773536682128905, 0.029733440399169923, 0.02974515151977539, 0.029550592422485353, 0.029392192840576172, 0.029863744735717773, 0.030069631576538088, 0.029356000900268554, 0.029468704223632812, 0.029388799667358398, 0.029429279327392577, 0.030171615600585937, 0.0297574405670166, 0.029880319595336914, 0.029863935470581054, 0.030064640045166017, 0.029814783096313476, 0.029758655548095703, 0.02961827278137207, 0.02972336006164551, 0.02996019172668457, 0.03140940856933594, 0.03031260871887207, 0.03015875244140625, 0.030266048431396485, 0.030004480361938476, 0.02998944091796875, 0.029997247695922852, 0.029894655227661132, 0.029885824203491212, 0.02979193687438965, 0.030939680099487304, 0.02976358413696289, 0.02983772850036621, 0.02976083183288574, 0.029662912368774413, 0.029417695999145507, 0.029713184356689452, 0.029378559112548826, 0.02956185531616211, 0.029631488800048827, 0.031044736862182617, 0.029725536346435547, 0.02989468765258789, 0.030007295608520508, 0.029863935470581054, 0.029911039352416992, 0.030007295608520508, 0.03006780815124512, 0.03013929557800293, 0.029943231582641602, 0.02993414306640625, 0.030069791793823242, 0.030095935821533203, 0.0300097599029541, 0.029962240219116212, 0.029648895263671874, 0.030195711135864257, 0.029863199234008788, 0.030173023223876952, 0.030091968536376953, 0.03042729568481445, 0.029728031158447264, 0.030200351715087892, 0.030167167663574218, 0.029611200332641602, 0.029680543899536133, 0.02972876739501953, 0.029847423553466798, 0.029644927978515624, 0.03016499137878418, 0.0305930233001709, 0.030357503890991212, 0.029937440872192383, 0.03039254379272461, 0.030044160842895507, 0.030039968490600585, 0.030125343322753906, 0.02991391944885254, 0.030023679733276368, 0.030064640045166017, 0.030089216232299806, 0.02976959991455078, 0.029855871200561525]",tokens/s,33.57021959470082,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1302.87616,1127.153664,0.0,731.906048,703.86944,s,1,8.206771484375,8.206771484375,0.0,8.206771484375,8.206771484375,8.206771484375,8.206771484375,[8.206771484375],,kWh,3.93551164999811e-05,4.328630238657786e-06,1.4016677879979955e-05,5.7700424618618844e-05,,MB,1343.090688,1406.07488,0.0,998.244352,942.608384,s,10,1.392733917236328,0.1392733917236328,0.0007123218924973523,0.13926077270507814,0.14019971618652344,0.14021211700439454,0.14022203765869143,"[0.13982981872558595, 0.13877853393554687, 0.13936058044433594, 0.13832496643066405, 0.14019696044921875, 0.13818496704101563, 0.13876640319824218, 0.1391609649658203, 0.1399062042236328, 0.14022451782226564]",tokens/s,1838.11133506387,kWh,4.299435067341441e-06,4.7415109055924025e-07,2.863846245323007e-06,7.637432403223689e-06,tokens/kWh,33519118.26963533,MB,1371.758592,1416.56064,0.0,1006.63296,942.610944,s,10,14.617361083984376,1.4617361083984375,0.012876126630085818,1.4593582763671873,1.4756720092773439,1.4820754455566407,1.4871981945800783,"[1.4884788818359376, 1.441701416015625, 1.45390087890625, 1.4595833740234374, 1.4591331787109374, 1.452447998046875, 1.460952392578125, 1.473926025390625, 1.4529879150390625, 1.4742490234375]",tokens/s,43.099434732461006,kWh,4.2311381469324745e-05,4.66651367157255e-06,1.775817025228097e-05,6.473606539317828e-05,tokens/kWh,973182.407941629,,s,630,14.615057729721066,0.023198504332890585,0.0010747253268728483,0.0230548152923584,0.02352419910430908,0.023748496055603027,0.024971472358703634,"[0.02363670349121094, 0.0233123836517334, 0.023070783615112306, 0.023119487762451173, 0.023009599685668944, 0.023181312561035155, 0.023079999923706053, 0.02294166374206543, 0.04850377655029297, 0.02305174446105957, 0.02288483238220215, 0.022913087844848634, 0.022966272354125978, 0.023103296279907228, 0.02301971244812012, 0.02309119987487793, 0.023581727981567383, 0.023249504089355468, 0.023692928314208984, 0.023134975433349608, 0.0231112003326416, 0.023187360763549804, 0.023136831283569335, 0.023205888748168944, 0.023328800201416016, 0.023369056701660156, 0.023322271347045898, 0.02337481689453125, 0.023478271484375, 0.02348441505432129, 0.023517343521118166, 0.0235313606262207, 0.02349398422241211, 0.023423648834228514, 0.023410688400268553, 0.023393375396728516, 0.023516128540039063, 0.023551488876342775, 0.023534175872802734, 0.02339619255065918, 0.023410688400268553, 0.023279712677001952, 0.02335436820983887, 0.023339008331298827, 0.023380895614624024, 0.02327884864807129, 0.02335001564025879, 0.0233504638671875, 0.02338899230957031, 0.023314079284667968, 0.02327587127685547, 0.023146495819091797, 0.023029760360717775, 0.02293350410461426, 0.022916095733642578, 0.022803455352783202, 0.02289779281616211, 0.022886335372924806, 0.022965183258056642, 0.02289004707336426, 0.02288889694213867, 0.022990848541259764, 0.02289446449279785, 0.023520448684692382, 0.023073600769042968, 0.023182687759399415, 0.02286204719543457, 0.022997440338134764, 0.022939647674560547, 0.022837247848510742, 0.02286172866821289, 0.022797887802124023, 0.0227476806640625, 0.02266921615600586, 0.022767711639404296, 0.022735967636108398, 0.023396255493164063, 0.022614240646362305, 0.022860031127929687, 0.022872608184814455, 0.02274508857727051, 0.022898687362670898, 0.02269388771057129, 0.022777856826782225, 0.02291302490234375, 0.022872064590454103, 0.022743040084838868, 0.022645984649658203, 0.02284623908996582, 0.023092639923095702, 0.02277571105957031, 0.022710975646972657, 0.02292230415344238, 0.022946048736572265, 0.023136959075927735, 0.022756576538085938, 0.023046207427978516, 0.022895328521728514, 0.022865184783935545, 0.02280112075805664, 0.022918880462646483, 0.02282035255432129, 0.022752031326293946, 0.022839296340942384, 0.022834943771362304, 0.022888320922851563, 0.02285945510864258, 0.022909631729125978, 0.02285977554321289, 0.02312940788269043, 0.02301817512512207, 0.022873407363891603, 0.022870719909667967, 0.022822912216186524, 0.02285158348083496, 0.022884319305419922, 0.02287548828125, 0.022862367630004883, 0.02291859245300293, 0.022814815521240234, 0.022911392211914062, 0.022786272048950194, 0.02291097640991211, 0.02277891159057617, 0.022881248474121093, 0.02280243110656738, 0.023179264068603517, 0.023015071868896484, 0.022899040222167967, 0.022889503479003905, 0.022811616897583008, 0.022904224395751953, 0.02289254379272461, 0.02276620864868164, 0.022914016723632812, 0.022791263580322265, 0.02299075126647949, 0.02286796760559082, 0.022906015396118164, 0.022878400802612303, 0.02286467170715332, 0.022899871826171876, 0.022936288833618163, 0.023248512268066405, 0.02297894477844238, 0.022999040603637694, 0.022929407119750975, 0.023276992797851562, 0.023089727401733397, 0.022995071411132814, 0.022982528686523437, 0.022958335876464845, 0.022886016845703124, 0.022932960510253907, 0.02291801643371582, 0.023029535293579102, 0.023067840576171873, 0.022944576263427736, 0.022887744903564454, 0.02305913543701172, 0.0228121280670166, 0.022968864440917967, 0.022959295272827147, 0.023060287475585937, 0.02282352066040039, 0.023030176162719726, 0.022961631774902343, 0.023009824752807616, 0.02294806480407715, 0.022850656509399415, 0.02323891258239746, 0.02304979133605957, 0.0229116153717041, 0.022918527603149413, 0.02454207992553711, 0.025317375183105468, 0.02311577606201172, 0.023189504623413085, 0.022869983673095704, 0.02304115104675293, 0.023408607482910158, 0.02292627143859863, 0.02300444793701172, 0.025187040328979494, 0.02320128059387207, 0.022974111557006835, 0.022891359329223634, 0.02293286323547363, 0.02295257568359375, 0.02390572738647461, 0.02316739273071289, 0.023077024459838866, 0.02301568031311035, 0.02295577621459961, 0.02301468849182129, 0.023483104705810547, 0.022945791244506835, 0.02288640022277832, 0.022960128784179686, 0.025794240951538087, 0.023046623229980467, 0.02288800048828125, 0.02298918342590332, 0.02287401580810547, 0.02284873580932617, 0.02287081527709961, 0.022790143966674805, 0.022798336029052735, 0.022747135162353514, 0.022794240951538085, 0.02285772705078125, 0.022857599258422852, 0.022771232604980467, 0.022852191925048827, 0.022856767654418946, 0.0227706241607666, 0.022806079864501953, 0.02310188865661621, 0.023027711868286133, 0.0231014404296875, 0.02320115280151367, 0.023295679092407227, 0.02322528076171875, 0.024016895294189454, 0.02401424026489258, 0.02336582374572754, 0.02340496063232422, 0.023275199890136718, 0.023271615982055665, 0.023824512481689455, 0.02330419158935547, 0.023225759506225584, 0.023167520523071288, 0.023006271362304688, 0.023070816040039063, 0.0230449275970459, 0.023038047790527344, 0.023090463638305664, 0.02303657531738281, 0.022930559158325196, 0.02304300880432129, 0.02291302490234375, 0.02302764892578125, 0.023717952728271485, 0.023127487182617187, 0.023319103240966796, 0.023170463562011717, 0.023155296325683594, 0.023005184173583985, 0.023150592803955077, 0.023452735900878905, 0.023597375869750976, 0.024172128677368163, 0.023325056076049806, 0.024043071746826173, 0.023056896209716796, 0.02304198455810547, 0.02300851249694824, 0.02280729675292969, 0.022961151123046874, 0.022916095733642578, 0.022941919326782228, 0.022753055572509766, 0.022847488403320314, 0.022847488403320314, 0.022784000396728517, 0.022863775253295898, 0.02299679946899414, 0.02294416046142578, 0.023084640502929688, 0.022982944488525392, 0.02287001609802246, 0.022831104278564454, 0.022937599182128905, 0.02303385543823242, 0.02335273551940918, 0.023762752532958984, 0.02518671989440918, 0.023163135528564454, 0.023136383056640626, 0.02292736053466797, 0.022958208084106445, 0.023627647399902342, 0.02512227249145508, 0.023079072952270508, 0.023058816909790038, 0.022996992111206056, 0.02291916847229004, 0.022833152770996092, 0.022865087509155273, 0.022893375396728515, 0.02285772705078125, 0.022867136001586914, 0.022956863403320312, 0.023945215225219727, 0.024352767944335937, 0.02296944046020508, 0.023247776031494142, 0.023001087188720702, 0.02298454475402832, 0.023093408584594726, 0.022962175369262695, 0.022964223861694336, 0.022851232528686524, 0.02336944007873535, 0.02301804733276367, 0.02321824073791504, 0.02288640022277832, 0.022977823257446288, 0.023026399612426758, 0.023010623931884765, 0.023160736083984376, 0.02290505599975586, 0.02312169647216797, 0.023230432510375976, 0.023426143646240235, 0.022989055633544923, 0.02294646453857422, 0.02285935974121094, 0.02303545570373535, 0.022934560775756837, 0.022888256072998048, 0.022836767196655273, 0.022837728500366212, 0.02279347229003906, 0.022925567626953126, 0.022860288619995117, 0.022937599182128905, 0.022792127609252928, 0.023054079055786134, 0.022919488906860352, 0.022908607482910157, 0.022825279235839845, 0.023033504486083985, 0.022808927536010742, 0.023377344131469725, 0.023071296691894533, 0.023055551528930664, 0.022936159133911133, 0.022789567947387696, 0.02287696075439453, 0.022868223190307617, 0.022971744537353515, 0.022808031082153322, 0.022881216049194335, 0.0228351993560791, 0.02288844871520996, 0.022792192459106447, 0.02304777526855469, 0.02293507194519043, 0.022923744201660157, 0.022888864517211914, 0.02302329635620117, 0.022992607116699218, 0.022864448547363282, 0.022928991317749024, 0.02289299201965332, 0.022803808212280275, 0.022939712524414062, 0.022991455078125, 0.02309119987487793, 0.023033824920654297, 0.02301139259338379, 0.023295967102050782, 0.02307891273498535, 0.02302566337585449, 0.023126016616821288, 0.023045343399047853, 0.02351798439025879, 0.023557439804077148, 0.02349305534362793, 0.023537919998168944, 0.023452831268310548, 0.02347520065307617, 0.023521120071411133, 0.02354742431640625, 0.023775711059570312, 0.023623327255249023, 0.024602272033691405, 0.02423638343811035, 0.02342246437072754, 0.023298559188842775, 0.023205888748168944, 0.023231584548950194, 0.02315727996826172, 0.02322470474243164, 0.023162879943847657, 0.022929407119750975, 0.022894271850585936, 0.022782272338867187, 0.02282048034667969, 0.0228089599609375, 0.022819839477539062, 0.02277174377441406, 0.02274403190612793, 0.022925216674804686, 0.022997087478637695, 0.02288198471069336, 0.022884767532348634, 0.022903839111328126, 0.023134944915771484, 0.022976415634155273, 0.02291472053527832, 0.022959808349609374, 0.02293142318725586, 0.02297542381286621, 0.023090240478515624, 0.023148511886596678, 0.023038496017456056, 0.02293564796447754, 0.023219615936279296, 0.023286272048950195, 0.023151039123535156, 0.02313577651977539, 0.023190176010131836, 0.023365440368652343, 0.023171072006225587, 0.023037343978881835, 0.022952543258666993, 0.023114944458007814, 0.02303468894958496, 0.022992895126342772, 0.023044095993041993, 0.023227424621582032, 0.023094240188598632, 0.022984128952026367, 0.023058176040649414, 0.023105823516845703, 0.023269920349121093, 0.02329100799560547, 0.02349350357055664, 0.02363520050048828, 0.023523616790771484, 0.023548063278198243, 0.023605567932128906, 0.023406848907470704, 0.023518495559692383, 0.023424800872802735, 0.02352943992614746, 0.0237238712310791, 0.023731071472167967, 0.024048959732055664, 0.02337775993347168, 0.02346281623840332, 0.02342911911010742, 0.023513088226318358, 0.023330816268920897, 0.023471935272216797, 0.023332927703857424, 0.023444927215576172, 0.023372703552246094, 0.02349443244934082, 0.02346188735961914, 0.023627328872680663, 0.023643680572509766, 0.02364022445678711, 0.023479040145874024, 0.023497983932495116, 0.02366326332092285, 0.02378883171081543, 0.02348217582702637, 0.023369983673095705, 0.023345920562744142, 0.02356950378417969, 0.023409568786621093, 0.023353343963623048, 0.023811744689941405, 0.02333139228820801, 0.023190336227416994, 0.023231456756591797, 0.023391807556152344, 0.02336332893371582, 0.02338476753234863, 0.023195648193359376, 0.023289056777954103, 0.023480640411376954, 0.023343616485595704, 0.02350486373901367, 0.023307392120361328, 0.023331552505493163, 0.024008575439453124, 0.023488191604614257, 0.023230144500732422, 0.023358367919921876, 0.02330419158935547, 0.02336067199707031, 0.023184223175048826, 0.023084928512573242, 0.023023008346557617, 0.02288092803955078, 0.02299235153198242, 0.022891008377075195, 0.022881536483764647, 0.023142528533935548, 0.02305932807922363, 0.023315391540527343, 0.022950143814086915, 0.02342793655395508, 0.023104448318481446, 0.02330508804321289, 0.022975744247436522, 0.02314521598815918, 0.02356003189086914, 0.02525609588623047, 0.023888927459716797, 0.02334409523010254, 0.023220191955566405, 0.02300441551208496, 0.022983455657958986, 0.023157920837402344, 0.022805408477783205, 0.02285971260070801, 0.022877599716186522, 0.0227989444732666, 0.022812671661376953, 0.022889984130859374, 0.0227740478515625, 0.022851808547973633, 0.022847711563110353, 0.022927135467529298, 0.02284339141845703, 0.022826271057128908, 0.022884416580200195, 0.022821279525756837, 0.022774015426635742, 0.02287615966796875, 0.022975584030151368, 0.02291935920715332, 0.022989536285400392, 0.023025535583496095, 0.023122047424316405, 0.023257087707519532, 0.023080511093139647, 0.023457216262817382, 0.023617887496948244, 0.023017728805541992, 0.023253408432006836, 0.022930816650390626, 0.023232959747314454, 0.023957696914672852, 0.022982656478881838, 0.022998207092285155, 0.023019840240478515, 0.02302195167541504, 0.02291315269470215, 0.02314646339416504, 0.022958112716674806, 0.02297235107421875, 0.022829120635986327, 0.023100608825683593, 0.02288912010192871, 0.023010623931884765, 0.02362051200866699, 0.023019039154052734, 0.02293596839904785, 0.022982624053955077, 0.022972448348999024, 0.023035903930664063, 0.022960031509399414, 0.022927423477172852, 0.023130144119262695, 0.02314035224914551, 0.023234560012817384, 0.02325503921508789, 0.023321887969970704, 0.023200223922729492, 0.02324710464477539, 0.02442073631286621, 0.023675039291381837, 0.02340425682067871, 0.023546207427978517, 0.023469472885131838, 0.02328835105895996, 0.023420671463012695, 0.023372032165527343, 0.02339344024658203, 0.02316889572143555, 0.023120864868164063, 0.023152639389038086, 0.023998367309570313, 0.02312406349182129, 0.023179264068603517, 0.023138303756713868, 0.023242624282836914, 0.023146623611450194, 0.023225439071655272, 0.023247520446777345, 0.02325872039794922, 0.02329052734375, 0.023293407440185546, 0.023148927688598633, 0.02300739288330078, 0.023188831329345704, 0.023147167205810545, 0.023384063720703126, 0.02414182472229004, 0.0240762882232666, 0.023477632522583006, 0.023470592498779298, 0.023296127319335936, 0.023416831970214845, 0.023220096588134766, 0.023221792221069334, 0.02320240020751953, 0.02330419158935547, 0.02316867256164551, 0.023357791900634764, 0.023451648712158202, 0.023386112213134767, 0.023135648727416993, 0.02309087944030762, 0.023210912704467773, 0.023326719284057617, 0.023719072341918945, 0.023495359420776366, 0.023568544387817383, 0.02347337532043457, 0.02351798439025879, 0.024325408935546876, 0.02379225540161133, 0.023479455947875975, 0.023317440032958985, 0.023459583282470702, 0.023400703430175782, 0.023431167602539063, 0.02332876777648926, 0.023370752334594725, 0.023161855697631836, 0.02336329650878906, 0.023400672912597655]",tokens/s,43.10622726579019,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3808.09216,5229.182976,0.0,4833.93536,4546.659328,s,1,11.0329423828125,11.0329423828125,0.0,11.0329423828125,11.0329423828125,11.0329423828125,11.0329423828125,[11.0329423828125],,kWh,0.00011698755016667235,1.2897346935870295e-05,4.96389286000036e-05,0.00017952382570254625,,MB,2035.134464,5260.640256,0.0,4852.809728,4095.212032,s,10,25.266362792968753,2.526636279296875,0.005492580904158527,2.5275803222656252,2.5309092529296873,2.5315973266601564,2.5321477856445314,"[2.511630126953125, 2.524263671875, 2.52607373046875, 2.5275087890625, 2.526899658203125, 2.52872998046875, 2.53075634765625, 2.530563232421875, 2.52765185546875, 2.532285400390625]",tokens/s,101.32047976103667,kWh,7.364710218833332e-05,8.121952807567629e-06,4.891492802080004e-05,0.000130683983016701,tokens/kWh,1958924.070804331,MB,2044.633088,5260.640256,0.0,4852.809728,4197.763584,s,10,21.150250488281248,2.115025048828125,0.008539970125787092,2.114699951171875,2.124815576171875,2.127568310546875,2.129770498046875,"[2.120118896484375, 2.130321044921875, 2.113159423828125, 2.124203857421875, 2.116240478515625, 2.11951318359375, 2.10022412109375, 2.112948974609375, 2.10659716796875, 2.10692333984375]",tokens/s,29.786881264080776,kWh,6.259576008583317e-05,6.905332628904864e-06,4.141061646180007e-05,0.00011091170917653813,tokens/kWh,568019.3774646726,,s,630,21.147305030822746,0.0335671508425758,0.0006349428110996855,0.03345873641967774,0.03397428321838379,0.034314902687072754,0.035135231094360356,"[0.034490623474121095, 0.03379529571533203, 0.03387881469726563, 0.033453342437744144, 0.033535839080810544, 0.03355862426757812, 0.03369055938720703, 0.03366476821899414, 0.03350067138671875, 0.03330300903320312, 0.03340630340576172, 0.03346249771118164, 0.03342598342895508, 0.03328716659545899, 0.03358950424194336, 0.03359939193725586, 0.03370479965209961, 0.03390902328491211, 0.03356265640258789, 0.03363983917236328, 0.03332649612426758, 0.0340417594909668, 0.03332969665527344, 0.03318751907348633, 0.03332172775268555, 0.03328211212158203, 0.033275745391845704, 0.03336553573608399, 0.033317440032958986, 0.033250911712646485, 0.03344425582885742, 0.03344384002685547, 0.03427123260498047, 0.03370348739624023, 0.03368982315063476, 0.03338467025756836, 0.0336814079284668, 0.03350732803344727, 0.03427328109741211, 0.034197376251220706, 0.03367129516601562, 0.033914432525634766, 0.03360550308227539, 0.03353430557250976, 0.035514816284179684, 0.03366799926757812, 0.03403046417236328, 0.03344793701171875, 0.03352579116821289, 0.0337174072265625, 0.033498241424560544, 0.033425086975097655, 0.033544193267822264, 0.033495166778564456, 0.03348672103881836, 0.03349308776855469, 0.0335335693359375, 0.03349532699584961, 0.03353984069824219, 0.033945854187011716, 0.034681087493896486, 0.03374684906005859, 0.03355657577514649, 0.04509280014038086, 0.034170528411865235, 0.033417152404785155, 0.03335651016235352, 0.03499827194213867, 0.034350719451904294, 0.03334316635131836, 0.03509958267211914, 0.03372003173828125, 0.03326774215698242, 0.03320217514038086, 0.03323206329345703, 0.03536569595336914, 0.03677503967285156, 0.033418113708496094, 0.03345619201660156, 0.033626369476318356, 0.0332957763671875, 0.03360918426513672, 0.033893184661865236, 0.03337577438354492, 0.033376449584960936, 0.034545951843261716, 0.03330047988891602, 0.033436767578125, 0.03337104034423828, 0.03317504119873047, 0.03337094497680664, 0.03319692611694336, 0.033858497619628905, 0.03346540832519531, 0.033286048889160154, 0.033237247467041015, 0.03332777786254883, 0.033791519165039065, 0.03317193603515625, 0.03318076705932617, 0.03323961639404297, 0.03312870407104492, 0.0330830078125, 0.03320995330810547, 0.033260414123535156, 0.03340697479248047, 0.03349270248413086, 0.03368499374389648, 0.0336412467956543, 0.0335945930480957, 0.03375593566894531, 0.033658878326416015, 0.0337367057800293, 0.03363430404663086, 0.033849342346191406, 0.03361734390258789, 0.03354476928710937, 0.03353190231323242, 0.033560928344726564, 0.03361897659301758, 0.033610591888427736, 0.03343974304199219, 0.033405982971191406, 0.03335654449462891, 0.0334543342590332, 0.033365760803222656, 0.03506012725830078, 0.0334719352722168, 0.033178367614746095, 0.033259326934814454, 0.03305376052856445, 0.033842208862304685, 0.033556320190429687, 0.033427486419677736, 0.033355743408203124, 0.03363190460205078, 0.03371868896484375, 0.033753089904785157, 0.033726463317871096, 0.0335261116027832, 0.03377888107299805, 0.0341828498840332, 0.03359545516967773, 0.03363619232177734, 0.0336126708984375, 0.03340492630004883, 0.03357644653320312, 0.03369558334350586, 0.03343366241455078, 0.03335228729248047, 0.033492225646972656, 0.033436641693115235, 0.03344294357299805, 0.03368617630004883, 0.033670913696289065, 0.033708160400390624, 0.033974369049072264, 0.03372851181030274, 0.03680624008178711, 0.03341561508178711, 0.03349913787841797, 0.033486846923828126, 0.03350115203857422, 0.03348275375366211, 0.033484798431396484, 0.03336092758178711, 0.033360992431640625, 0.03313449478149414, 0.033974273681640625, 0.033394111633300784, 0.03319267272949219, 0.033896350860595705, 0.033306110382080076, 0.03321696090698242, 0.033072353363037106, 0.033108894348144534, 0.033113441467285155, 0.03314947128295898, 0.033040382385253905, 0.033056320190429686, 0.03326406478881836, 0.033277950286865234, 0.03331043243408203, 0.03328643035888672, 0.03316326522827148, 0.033454078674316406, 0.03361996841430664, 0.033253185272216795, 0.033237537384033206, 0.03418313598632813, 0.033468353271484376, 0.03337577438354492, 0.033300064086914063, 0.03321977615356445, 0.03333622360229492, 0.033463134765625, 0.03371993637084961, 0.033601055145263674, 0.033412158966064455, 0.03355215835571289, 0.03363670349121094, 0.03353152084350586, 0.03364156723022461, 0.03376460647583008, 0.033975105285644534, 0.03438876724243164, 0.03406038284301758, 0.03411558532714844, 0.03401651382446289, 0.03485897445678711, 0.03726416015625, 0.03417839813232422, 0.034200225830078125, 0.03442457580566406, 0.034181568145751955, 0.034496318817138674, 0.0336814079284668, 0.03373875045776367, 0.03360563278198242, 0.03336806488037109, 0.03337936019897461, 0.03336662292480469, 0.03479180908203125, 0.034914527893066406, 0.03362691116333008, 0.03374796676635742, 0.03362611389160156, 0.03361497497558594, 0.03343964767456055, 0.034859424591064454, 0.03327558517456055, 0.0331673583984375, 0.03305782318115234, 0.03308889770507813, 0.033173023223876955, 0.03317241668701172, 0.03318531036376953, 0.033185726165771486, 0.033186046600341794, 0.033128032684326174, 0.03332780838012695, 0.033029376983642576, 0.03312511825561523, 0.03316121673583984, 0.03365478515625, 0.034235454559326174, 0.03346089553833008, 0.03338412857055664, 0.033218143463134765, 0.03317452621459961, 0.033960254669189456, 0.03341881561279297, 0.034484222412109376, 0.03362748718261719, 0.03384592056274414, 0.033296607971191404, 0.033303520202636716, 0.033253921508789065, 0.03337801742553711, 0.03365126419067383, 0.03321036911010742, 0.033226593017578125, 0.033434814453125, 0.03325596618652344, 0.03323740768432617, 0.033176609039306644, 0.03334873580932617, 0.03310326385498047, 0.03343750381469727, 0.03369836807250977, 0.03369731140136719, 0.03376595306396484, 0.03384665679931641, 0.03397203063964844, 0.033761856079101565, 0.03354854583740234, 0.033589088439941406, 0.033683902740478514, 0.03338844680786133, 0.0333304328918457, 0.03340864181518555, 0.03358988952636719, 0.03360390472412109, 0.03349216079711914, 0.03366739273071289, 0.033476608276367184, 0.03348249435424805, 0.03349692916870117, 0.03404880142211914, 0.03347420883178711, 0.033462814331054684, 0.03395782470703125, 0.033724414825439454, 0.033726303100585935, 0.03368566513061524, 0.03352323150634766, 0.033685985565185546, 0.033581344604492185, 0.033672927856445316, 0.03372054290771485, 0.03374054336547851, 0.03376521682739258, 0.03363616180419922, 0.033441280364990236, 0.03359423828125, 0.03360153579711914, 0.03354243087768555, 0.0334681282043457, 0.03360153579711914, 0.033656639099121095, 0.03349113464355469, 0.03360943984985352, 0.034086910247802735, 0.03348099136352539, 0.034138111114501955, 0.03448627090454102, 0.03426873779296875, 0.034294368743896485, 0.033761470794677735, 0.03364828872680664, 0.03376537704467773, 0.03377280044555664, 0.03433865737915039, 0.03390262222290039, 0.0336864013671875, 0.03379404830932617, 0.03376265716552734, 0.03376224136352539, 0.033836769104003905, 0.03373897552490234, 0.0335617904663086, 0.0340711669921875, 0.03429782485961914, 0.03377686309814453, 0.033680160522460936, 0.03382681655883789, 0.03389132690429687, 0.033836158752441406, 0.033529472351074216, 0.03340723037719726, 0.03328121566772461, 0.033323104858398435, 0.033297119140625, 0.0334031982421875, 0.03344086456298828, 0.03343523025512695, 0.03369411087036133, 0.03347014236450195, 0.03334054565429687, 0.03325705718994141, 0.03315145492553711, 0.03316707229614258, 0.033067008972167966, 0.03307891082763672, 0.033795680999755856, 0.03431913757324219, 0.03451295852661133, 0.03430972671508789, 0.033804798126220705, 0.03387968063354492, 0.03354262542724609, 0.03336166381835937, 0.033142784118652346, 0.033212448120117186, 0.03312214279174805, 0.03318592071533203, 0.03320230484008789, 0.0330382080078125, 0.033029953002929685, 0.0339633903503418, 0.03471366500854492, 0.03357939147949219, 0.03336640167236328, 0.03324220657348633, 0.03321260833740235, 0.033259712219238284, 0.03360969543457031, 0.03369551849365234, 0.03444595336914062, 0.03356095886230469, 0.033224063873291014, 0.033134559631347656, 0.032978622436523435, 0.03305503845214844, 0.033922496795654296, 0.03317619323730469, 0.03310627365112305, 0.03330252838134766, 0.033181697845458984, 0.03322390365600586, 0.0331558723449707, 0.0331673583984375, 0.033140735626220705, 0.03315884780883789, 0.033056766510009765, 0.03318387222290039, 0.033140350341796875, 0.03318841552734375, 0.03314041519165039, 0.03339321517944336, 0.03350092697143554, 0.03336995315551758, 0.03380582427978516, 0.033530529022216794, 0.03337372970581055, 0.03355081558227539, 0.033435199737548826, 0.03338694381713867, 0.033287487030029296, 0.03336399841308594, 0.033266334533691405, 0.033261215209960934, 0.03331068801879883, 0.03316595077514648, 0.03319907379150391, 0.03321321487426758, 0.03314067077636719, 0.033255329132080076, 0.03334979248046875, 0.03332092666625976, 0.03470489501953125, 0.03354483032226562, 0.03327091217041016, 0.033898624420166015, 0.03350620651245117, 0.03321855926513672, 0.03326073455810547, 0.03330534362792969, 0.033111297607421875, 0.03314672088623047, 0.03313926315307617, 0.033118526458740236, 0.033748382568359374, 0.033382686614990234, 0.03320598220825195, 0.0332591667175293, 0.03327388763427734, 0.033169727325439456, 0.033199905395507816, 0.03315369415283203, 0.03321241760253906, 0.03433193588256836, 0.03380511856079101, 0.03379619216918945, 0.033950752258300784, 0.033397216796875, 0.03372905731201172, 0.03393430328369141, 0.03356979370117188, 0.03346361541748047, 0.033700576782226564, 0.03366652679443359, 0.03328870391845703, 0.033406814575195315, 0.03310950469970703, 0.03322332763671875, 0.033218719482421874, 0.033242721557617184, 0.033460384368896486, 0.0335975341796875, 0.03367660903930664, 0.03369209671020508, 0.0336096305847168, 0.03379260635375977, 0.03401894378662109, 0.03372969436645508, 0.033524703979492185, 0.03351523208618164, 0.03348918533325195, 0.03348652648925781, 0.03361724853515625, 0.03391791915893555, 0.03344179153442383, 0.03345388793945313, 0.03337648010253906, 0.033441375732421875, 0.03327638244628906, 0.03386934280395508, 0.033765151977539064, 0.03393727874755859, 0.03367599868774414, 0.033646209716796875, 0.03347081756591797, 0.03415657424926758, 0.03329766464233398, 0.033321697235107424, 0.03333491134643555, 0.03336214447021484, 0.03321615982055664, 0.03333766555786133, 0.03340630340576172, 0.03343193435668945, 0.03328287887573242, 0.03336956787109375, 0.03333065414428711, 0.03341769790649414, 0.03332089614868164, 0.03318790435791016, 0.033181343078613285, 0.03332979202270508, 0.033980766296386716, 0.03335452651977539, 0.033427326202392575, 0.03329280090332031, 0.0347061767578125, 0.03382454299926758, 0.033734142303466795, 0.03384396743774414, 0.03392086410522461, 0.03382195281982422, 0.03364064025878906, 0.03369039916992188, 0.033463199615478514, 0.03382374572753906, 0.03344384002685547, 0.03346022415161133, 0.033570816040039066, 0.03332620620727539, 0.03320646286010742, 0.03310768127441406, 0.03330572891235352, 0.03307475280761719, 0.03331660842895508, 0.03337884902954102, 0.03358428955078125, 0.03366588973999023, 0.03376054382324219, 0.03344854354858398, 0.03346444702148438, 0.03320585632324219, 0.03326809692382812, 0.033229183197021485, 0.03322403335571289, 0.03330694580078125, 0.033199264526367185, 0.033283103942871095, 0.03328307342529297, 0.03318767929077148, 0.033218849182128904, 0.033077953338623046, 0.03313792037963867, 0.03313945770263672, 0.03308755111694336, 0.03317343902587891, 0.033172767639160154, 0.03318851089477539, 0.033286209106445315, 0.03342742538452149, 0.03338582229614258, 0.03318035125732422, 0.03312553787231445, 0.03318364715576172, 0.03324409484863281, 0.033259136199951175, 0.03328201675415039, 0.03323945617675781, 0.03321343994140625, 0.03345331192016601, 0.03371366500854492, 0.03363587188720703, 0.03363913726806641, 0.03374396896362305, 0.03358972930908203, 0.03365548706054688, 0.03413891220092773, 0.033597888946533205, 0.03336422348022461, 0.03432662582397461, 0.03371465682983398, 0.03348223876953125, 0.0333540153503418, 0.03326793670654297, 0.03334966278076172, 0.0351497917175293, 0.03415407943725586, 0.03346713638305664, 0.03325609588623047, 0.0332193603515625, 0.03323926544189453, 0.033107967376708985, 0.03306694412231445, 0.03330227279663086, 0.03329609680175781, 0.03307171249389648, 0.03331651306152344, 0.03334924697875977, 0.03336163330078125, 0.03325417709350586, 0.033143009185791016, 0.0333138542175293, 0.03328710556030273, 0.03337968063354492, 0.03330319976806641, 0.033468414306640625, 0.03352892684936523, 0.03316009521484375, 0.033255199432373046, 0.0332224006652832, 0.033485279083251954, 0.03363401412963867, 0.033450271606445314, 0.033081344604492184, 0.033056766510009765, 0.03346031951904297, 0.03312332916259766, 0.033276927947998046, 0.03325247955322266, 0.03339699172973633, 0.03330841445922852, 0.03329513549804688, 0.0333496322631836, 0.0333375358581543, 0.03369753646850586, 0.03329977416992187, 0.03329455947875976, 0.03340671920776367, 0.03351126480102539, 0.03349708938598633, 0.03365488052368164, 0.03345724868774414, 0.0336176643371582, 0.03364601516723633, 0.03343008041381836, 0.033320735931396485, 0.033624542236328124, 0.0334354248046875, 0.03376534271240234, 0.03358415985107422, 0.034164928436279295, 0.0335153923034668]",tokens/s,29.791030066562072,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3798, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ self.model = DeciCoderModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ self.self_attn = DeciCoderAttention(config=config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ self._init_rope() File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1709, in __getattr__ raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2188.152832,2907.635712,0.0,2512.388096,2240.694784,s,1,9.295291015625,9.295291015625,0.0,9.295291015625,9.295291015625,9.295291015625,9.295291015625,[9.295291015625],,kWh,6.9199416970802e-05,7.625468309715327e-06,2.827668928798155e-05,0.00010510157456849888,,MB,2235.039744,2922.315776,0.0,2514.485248,2227.003904,s,10,11.443010864257813,1.1443010864257812,0.0009773683644739432,1.1445831298828124,1.1452647338867188,1.1453716857910157,1.145457247314453,"[1.1451075439453124, 1.143201904296875, 1.1424122314453125, 1.1437933349609375, 1.1447080078125, 1.1434659423828124, 1.144458251953125, 1.145240966796875, 1.1454786376953126, 1.14514404296875]",tokens/s,223.71734418221584,kWh,3.36507683370822e-05,3.7112161071116853e-06,2.2440184618799376e-05,5.980216906299326e-05,tokens/kWh,4280781.182541048,MB,2246.934528,2922.315776,0.0,2514.485248,2334.959104,s,10,18.52756018066406,1.852756018066406,0.012863735451607414,1.8540548706054687,1.8657800903320314,1.870597869873047,1.8744520935058595,"[1.8308460693359374, 1.85354541015625, 1.86470947265625, 1.84231201171875, 1.8545643310546875, 1.8624730224609376, 1.8384434814453126, 1.845342041015625, 1.85990869140625, 1.8754156494140626]",tokens/s,34.00339784930169,kWh,5.4726120434997937e-05,6.036620997139593e-06,2.994935729279946e-05,9.0712098724937e-05,tokens/kWh,694504.9324791019,,s,630,18.524415981292716,0.029403834890940832,0.00044869455187966965,0.02934132766723633,0.029830815505981444,0.030034862422943114,0.03134245605468751,"[0.029881599426269532, 0.02961894416809082, 0.029163040161132813, 0.029067264556884766, 0.028934623718261717, 0.028903423309326173, 0.028970783233642577, 0.028920032501220702, 0.028895231246948243, 0.028848127365112306, 0.029108224868774416, 0.029357088088989257, 0.0290416316986084, 0.029075456619262696, 0.028929855346679686, 0.02875366401672363, 0.028874208450317383, 0.028949472427368166, 0.02893529510498047, 0.028912511825561524, 0.028862464904785157, 0.028906496047973632, 0.02899046325683594, 0.028972063064575195, 0.028998239517211914, 0.02914067268371582, 0.02897990417480469, 0.029129728317260743, 0.02913382339477539, 0.029099679946899413, 0.02894883155822754, 0.029048351287841796, 0.02890595245361328, 0.029070463180541992, 0.028910463333129882, 0.02889904022216797, 0.02887299156188965, 0.028994592666625976, 0.02883683204650879, 0.0289434871673584, 0.0291910400390625, 0.02899660873413086, 0.029683935165405274, 0.02902889633178711, 0.02908185577392578, 0.029046783447265623, 0.029401023864746093, 0.02945849609375, 0.02899135971069336, 0.02897430419921875, 0.028849023818969727, 0.028860448837280273, 0.028850175857543944, 0.028887039184570314, 0.028794879913330077, 0.028987295150756837, 0.028885087966918944, 0.028882944107055664, 0.028878303527832033, 0.029157375335693358, 0.029600191116333007, 0.029636320114135743, 0.029606271743774414, 0.02998080062866211, 0.02951932716369629, 0.02926748847961426, 0.029428735733032226, 0.029128704071044922, 0.029319072723388673, 0.029057119369506838, 0.029265920639038087, 0.02879283142089844, 0.02904047966003418, 0.028951711654663086, 0.029320192337036134, 0.0290119686126709, 0.029648895263671874, 0.029017311096191406, 0.029143199920654297, 0.028895456314086913, 0.028938655853271485, 0.028839935302734376, 0.029190143585205077, 0.02936627197265625, 0.029509632110595704, 0.02943180847167969, 0.029546239852905273, 0.02951398468017578, 0.02962544059753418, 0.029682592391967775, 0.02948281669616699, 0.02950163269042969, 0.029560831069946288, 0.0299899845123291, 0.029729696273803712, 0.02950102424621582, 0.029599584579467774, 0.029329984664916993, 0.029490367889404297, 0.02937939262390137, 0.029616128921508788, 0.02939904022216797, 0.029638656616210936, 0.029403135299682616, 0.029670495986938477, 0.029338399887084962, 0.029367551803588868, 0.029427679061889648, 0.029813663482666015, 0.029437952041625977, 0.029895967483520507, 0.029359935760498047, 0.0294748477935791, 0.029158271789550782, 0.029198175430297853, 0.029232959747314453, 0.02954857635498047, 0.030172512054443358, 0.03141526412963867, 0.029382591247558595, 0.029267232894897462, 0.029168415069580077, 0.029288448333740235, 0.02914303970336914, 0.029261823654174804, 0.029163520812988283, 0.03021366310119629, 0.029829599380493163, 0.029708288192749024, 0.02925923156738281, 0.02929088020324707, 0.029615808486938476, 0.029513887405395508, 0.029457952499389647, 0.029205280303955077, 0.029304832458496095, 0.02926915168762207, 0.029696863174438478, 0.02941542434692383, 0.029624319076538085, 0.029913087844848633, 0.02955264091491699, 0.029703584671020508, 0.02969046401977539, 0.02953830337524414, 0.029492288589477538, 0.02925200080871582, 0.02937606430053711, 0.029516767501831055, 0.029814783096313476, 0.029693952560424806, 0.029877504348754882, 0.029906944274902345, 0.02986470413208008, 0.029689247131347657, 0.029698144912719725, 0.029530624389648437, 0.029509632110595704, 0.029330751419067384, 0.02942608070373535, 0.029077056884765626, 0.02907200050354004, 0.02914886474609375, 0.029816448211669924, 0.029181791305541993, 0.029178815841674803, 0.028983295440673826, 0.029187231063842772, 0.029807455062866212, 0.030068735122680663, 0.029122560501098634, 0.029620223999023438, 0.029153247833251954, 0.029779296875, 0.03011849594116211, 0.029694047927856446, 0.02971238327026367, 0.029615808486938476, 0.02945465660095215, 0.029749248504638674, 0.02972876739501953, 0.029718143463134766, 0.030136703491210937, 0.032372417449951174, 0.029878591537475584, 0.02953011131286621, 0.029155328750610353, 0.0293143367767334, 0.02924412727355957, 0.029981695175170898, 0.029518848419189454, 0.029174816131591796, 0.029034847259521483, 0.02892451286315918, 0.029091840744018556, 0.028886911392211913, 0.028971168518066408, 0.028802112579345705, 0.02891257667541504, 0.02872662353515625, 0.02902697563171387, 0.02873139190673828, 0.028964864730834962, 0.028807167053222657, 0.029390623092651367, 0.02896099281311035, 0.028833791732788085, 0.028892799377441405, 0.029329504013061523, 0.029242847442626955, 0.029124927520751954, 0.029190656661987304, 0.029378559112548826, 0.029190143585205077, 0.02936832046508789, 0.029485055923461914, 0.029722623825073242, 0.029618175506591796, 0.029964223861694336, 0.03068934440612793, 0.029757375717163085, 0.029335615158081054, 0.02928963279724121, 0.02964271926879883, 0.029426559448242188, 0.029708288192749024, 0.029515039443969725, 0.02924617576599121, 0.02919593620300293, 0.02938252830505371, 0.02950396728515625, 0.029422943115234374, 0.02936899185180664, 0.02927408027648926, 0.029620256423950195, 0.029120479583740234, 0.029222623825073242, 0.029347423553466798, 0.02929737663269043, 0.029069311141967775, 0.02906243133544922, 0.02895302391052246, 0.028983232498168945, 0.028809568405151368, 0.028929920196533204, 0.028899423599243163, 0.029099456787109373, 0.028916032791137695, 0.029147424697875977, 0.029042688369750977, 0.02920857620239258, 0.029255680084228516, 0.02991472053527832, 0.029235584259033203, 0.029681663513183593, 0.032481281280517575, 0.02920038414001465, 0.029087135314941406, 0.029091903686523438, 0.02911020851135254, 0.028957632064819334, 0.029066719055175782, 0.02918931198120117, 0.029682687759399414, 0.02931056022644043, 0.029289920806884764, 0.029460575103759764, 0.029528959274291992, 0.029609983444213867, 0.02957030487060547, 0.029383424758911134, 0.02923520088195801, 0.0293023681640625, 0.029615711212158204, 0.029610559463500975, 0.02969772720336914, 0.029696575164794924, 0.029726720809936522, 0.02958457565307617, 0.029743936538696288, 0.029517824172973633, 0.029701311111450194, 0.029341535568237306, 0.029506528854370117, 0.028915712356567383, 0.03224694442749024, 0.0290098876953125, 0.029055871963500977, 0.029046783447265623, 0.02917580795288086, 0.02895052719116211, 0.02913689613342285, 0.02917318344116211, 0.029344127655029296, 0.029122304916381837, 0.0293687686920166, 0.029411327362060546, 0.029464576721191408, 0.029493247985839844, 0.029790048599243165, 0.029455615997314454, 0.02929961585998535, 0.029210399627685547, 0.029540576934814454, 0.029059072494506837, 0.029327360153198243, 0.02914838409423828, 0.02919628715515137, 0.029145631790161133, 0.029230623245239257, 0.029059776306152342, 0.029128095626831055, 0.029522560119628907, 0.029122560501098634, 0.028990528106689454, 0.030030656814575195, 0.029372415542602538, 0.02897715187072754, 0.028983295440673826, 0.029009088516235352, 0.029397823333740233, 0.029248735427856446, 0.029240095138549804, 0.02917580795288086, 0.029306880950927733, 0.029121728897094728, 0.02926265525817871, 0.02936422348022461, 0.03140764808654785, 0.029757919311523436, 0.02952979278564453, 0.029278303146362306, 0.02965433692932129, 0.0295699520111084, 0.029646751403808593, 0.029562496185302736, 0.029556896209716795, 0.02923756790161133, 0.02929254341125488, 0.02921174430847168, 0.02934982490539551, 0.02919251251220703, 0.028989471435546876, 0.028990079879760742, 0.02931711959838867, 0.02962380790710449, 0.03012646484375, 0.02970841598510742, 0.02958131217956543, 0.029626176834106444, 0.02978160095214844, 0.02945289611816406, 0.0295546875, 0.029519872665405275, 0.029661184310913087, 0.029634559631347656, 0.02982022476196289, 0.029727231979370116, 0.029693151473999025, 0.029647167205810548, 0.029858463287353514, 0.029910240173339844, 0.029944351196289062, 0.02997052764892578, 0.02984351921081543, 0.029628511428833007, 0.02974515151977539, 0.030519296646118164, 0.030103008270263673, 0.029878271102905272, 0.0297938232421875, 0.02965376091003418, 0.029796607971191408, 0.02949734306335449, 0.029498464584350587, 0.02920745658874512, 0.029199487686157228, 0.02896143913269043, 0.029861663818359373, 0.0294017276763916, 0.028915712356567383, 0.029005727767944335, 0.028893280029296874, 0.029068288803100587, 0.02894745635986328, 0.029044351577758788, 0.029114751815795897, 0.029388639450073244, 0.029225120544433592, 0.029255519866943358, 0.02958051109313965, 0.029262624740600585, 0.029108383178710937, 0.029243392944335936, 0.028921184539794923, 0.029109952926635742, 0.02904707145690918, 0.029233407974243165, 0.029139392852783202, 0.029445152282714843, 0.029468992233276366, 0.02928236770629883, 0.028892799377441405, 0.029113311767578125, 0.0289849910736084, 0.0289836483001709, 0.02874163246154785, 0.029087648391723633, 0.028833663940429688, 0.02919036865234375, 0.02899558448791504, 0.029034496307373047, 0.028999679565429686, 0.02912665557861328, 0.02925763130187988, 0.029233247756958007, 0.029222911834716796, 0.029392127990722657, 0.029073312759399415, 0.029066080093383788, 0.028911615371704103, 0.028962560653686523, 0.02896691131591797, 0.029051136016845704, 0.029183263778686522, 0.029245983123779296, 0.029326976776123045, 0.02937094306945801, 0.029304832458496095, 0.02919424057006836, 0.028991487503051756, 0.029916255950927735, 0.02905900764465332, 0.02917875289916992, 0.02911852836608887, 0.03012816047668457, 0.028868064880371094, 0.0291190071105957, 0.02927110481262207, 0.029464832305908205, 0.029278911590576173, 0.03003830337524414, 0.029491327285766603, 0.029396991729736328, 0.029261823654174804, 0.029327360153198243, 0.029449504852294923, 0.029329408645629884, 0.029398752212524415, 0.02921779251098633, 0.029939647674560546, 0.029269184112548828, 0.029551488876342773, 0.029669376373291017, 0.029370111465454103, 0.02940883255004883, 0.029481664657592773, 0.029581375122070312, 0.029523872375488282, 0.029440256118774415, 0.029857568740844727, 0.03000934410095215, 0.02960095977783203, 0.0292109432220459, 0.02900223922729492, 0.028872127532958983, 0.02934432029724121, 0.029018112182617187, 0.029236223220825194, 0.029221887588500976, 0.029156383514404298, 0.02900432014465332, 0.029661312103271484, 0.0289401912689209, 0.029341119766235352, 0.02880988883972168, 0.028850496292114256, 0.028771360397338866, 0.02902524757385254, 0.028876800537109375, 0.03140809631347656, 0.028914976119995116, 0.028850944519042968, 0.02876620864868164, 0.02895871925354004, 0.02886649513244629, 0.02900982475280762, 0.028977312088012696, 0.02936627197265625, 0.029095935821533202, 0.0291059513092041, 0.02889686393737793, 0.029147775650024414, 0.029105375289916992, 0.029330175399780275, 0.029085727691650392, 0.029908992767333983, 0.029429119110107423, 0.029337631225585938, 0.029131359100341796, 0.02923107147216797, 0.02910972785949707, 0.029196832656860353, 0.02887887954711914, 0.030092191696166993, 0.029625663757324217, 0.02936467170715332, 0.029535839080810547, 0.029455007553100584, 0.02959769630432129, 0.029511615753173827, 0.02977987289428711, 0.029628576278686522, 0.02969536018371582, 0.02949184036254883, 0.029715776443481445, 0.029583072662353514, 0.030097759246826172, 0.02964339256286621, 0.029654624938964844, 0.029554912567138672, 0.029735103607177734, 0.03012403106689453, 0.029560831069946288, 0.029421567916870117, 0.03185183906555176, 0.030444095611572266, 0.029569055557250975, 0.02914041519165039, 0.029382720947265625, 0.029014623641967774, 0.029333311080932616, 0.0294169921875, 0.029301408767700196, 0.029159263610839845, 0.02930607986450195, 0.029108415603637694, 0.029258495330810548, 0.029244415283203123, 0.029166015625, 0.02931123161315918, 0.02929897689819336, 0.02916969680786133, 0.029421567916870117, 0.02935398483276367, 0.029308704376220702, 0.02922064018249512, 0.029229503631591797, 0.029195743560791014, 0.02946451187133789, 0.0295053768157959, 0.030706432342529295, 0.029685152053833007, 0.029583967208862305, 0.029313024520874024, 0.029418624877929688, 0.029147104263305666, 0.029442975997924805, 0.02918809509277344, 0.02920403289794922, 0.02917830467224121, 0.029198335647583007, 0.029003679275512697, 0.029112415313720705, 0.029287424087524414, 0.030271776199340822, 0.029743040084838867, 0.030091264724731445, 0.029640512466430666, 0.02954662322998047, 0.02949724769592285, 0.02972051239013672, 0.029634176254272462, 0.029620832443237304, 0.02971379280090332, 0.030816896438598633, 0.030748672485351562, 0.029726720809936522, 0.02995110321044922, 0.029808704376220702, 0.030386560440063475, 0.029825023651123047, 0.029872287750244142, 0.029628704071044922, 0.029713632583618164, 0.029725471496582032, 0.029778976440429688, 0.02966147232055664, 0.029747903823852537, 0.029515264511108398, 0.030722560882568358, 0.030390272140502928, 0.030029439926147462, 0.02972096061706543, 0.029672447204589843, 0.029420127868652345, 0.02959401512145996, 0.02956867218017578, 0.029841760635375976, 0.030611455917358397, 0.03118284797668457, 0.029718271255493166, 0.029823232650756835, 0.029599231719970705, 0.02988083267211914, 0.029603904724121093, 0.02976255989074707, 0.02970275115966797, 0.029820415496826173, 0.02963462448120117, 0.02958393669128418, 0.02936636734008789, 0.02956300735473633, 0.029413087844848633, 0.029681055068969727, 0.029444128036499023, 0.029508127212524413, 0.029263391494750976, 0.029502239227294922, 0.029351936340332032, 0.02967737579345703, 0.029332767486572264, 0.02951046371459961, 0.02935171127319336, 0.02958367919921875, 0.029620223999023438, 0.02986137580871582, 0.02965862464904785, 0.029637184143066406, 0.029490943908691405]",tokens/s,34.00916933825169,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,7303.335936,9457.041408,0.0,9061.793792,8463.626752,s,1,14.3428876953125,14.3428876953125,0.0,14.3428876953125,14.3428876953125,14.3428876953125,14.3428876953125,[14.3428876953125],,kWh,0.00021487623368749383,2.36943618241099e-05,9.731035562601797e-05,0.00033588095113762167,,MB,1703.931904,9469.62432,0.0,9061.793792,7991.22432,s,10,57.66108496093751,5.766108496093751,0.004118686907086784,5.766230224609375,5.77012763671875,5.7712974609375,5.7722333203125,"[5.75782275390625, 5.76150146484375, 5.76316064453125, 5.76605322265625, 5.76616748046875, 5.76629296875, 5.76811865234375, 5.7696328125, 5.76986767578125, 5.77246728515625]",tokens/s,44.39736091914107,kWh,0.00016823414614416645,1.8556737855335502e-05,0.00011176370052199957,0.00029855458452150154,tokens/kWh,857464.6422204352,MB,1710.641152,9469.62432,0.0,9061.793792,8265.583104,s,10,29.5501298828125,2.9550129882812497,0.008249503465217734,2.9533575439453124,2.9656091796875,2.965724755859375,2.965817216796875,"[2.946534912109375, 2.94372900390625, 2.94428515625, 2.952433349609375, 2.96433447265625, 2.952168701171875, 2.95428173828125, 2.96558349609375, 2.96584033203125, 2.960938720703125]",tokens/s,21.319703246598333,kWh,8.632863470291603e-05,9.522743306560679e-06,5.753785158580105e-05,0.00015338922959527777,tokens/kWh,410719.84106203186,,s,630,29.54630256652833,0.04689889296274336,0.00047923563625497655,0.04689219093322754,0.04741061515808105,0.0475846269607544,0.048471395149230954,"[0.04812003326416016, 0.046564510345458984, 0.046091262817382815, 0.04598806381225586, 0.045979038238525394, 0.04610086441040039, 0.04623769760131836, 0.046126625061035154, 0.04627487945556641, 0.0464119987487793, 0.04636870574951172, 0.04629491043090821, 0.04653004837036133, 0.04624604797363281, 0.0463037109375, 0.046439838409423825, 0.046236255645751956, 0.04641299057006836, 0.046534881591796876, 0.04644262313842774, 0.046895584106445315, 0.04693507385253906, 0.04678960037231445, 0.04663296127319336, 0.04636188888549805, 0.04636064147949219, 0.04637763214111328, 0.04654489517211914, 0.04665958404541016, 0.04658995056152344, 0.046532127380371095, 0.046698814392089845, 0.046695873260498046, 0.04654572677612305, 0.04671990585327149, 0.04714707183837891, 0.04687548828125, 0.04673750305175781, 0.04682160186767578, 0.04728582382202148, 0.04730492782592773, 0.04722687911987305, 0.04717926406860352, 0.04693664169311523, 0.046825408935546875, 0.04672726440429688, 0.046781696319580075, 0.0467435188293457, 0.046682815551757816, 0.04675481414794922, 0.047612480163574215, 0.047161792755126955, 0.04750851058959961, 0.046975967407226565, 0.04699750518798828, 0.047193248748779296, 0.047197025299072264, 0.04832470321655274, 0.04708467102050781, 0.04720038223266602, 0.04728416061401367, 0.047280864715576174, 0.04733488082885742, 0.04825859069824219, 0.046731712341308594, 0.04619222259521484, 0.04598188781738281, 0.04601036834716797, 0.046211296081542966, 0.04611072158813476, 0.046046367645263674, 0.04602352142333985, 0.046219264984130856, 0.04647129440307617, 0.046382881164550784, 0.04619843292236328, 0.046271968841552734, 0.04626124954223633, 0.04636463928222656, 0.04648339080810547, 0.04655459213256836, 0.04655728149414062, 0.0468218879699707, 0.04706086349487305, 0.046960769653320314, 0.046833438873291014, 0.04676425552368164, 0.04658784103393555, 0.04655724716186523, 0.04634975814819336, 0.046292640686035155, 0.046274848937988285, 0.04633235168457031, 0.04642015838623047, 0.046405406951904295, 0.04632947158813477, 0.046516544342041014, 0.04664144134521484, 0.046680065155029295, 0.04659097671508789, 0.04692070388793945, 0.046868480682373044, 0.046908447265625, 0.046994110107421876, 0.04712502288818359, 0.04843084716796875, 0.047249408721923826, 0.04719113540649414, 0.047292896270751957, 0.04710780715942383, 0.04702076721191406, 0.04681932830810547, 0.046726367950439454, 0.04676268768310547, 0.04696428680419922, 0.0470164794921875, 0.046911617279052735, 0.046884193420410156, 0.04677072143554688, 0.04708784103393555, 0.04701136016845703, 0.047035808563232424, 0.047032833099365234, 0.04721903991699219, 0.04713881683349609, 0.04717363357543945, 0.048448993682861326, 0.04681372833251953, 0.04632985687255859, 0.04605155181884766, 0.04613507080078125, 0.04598691177368164, 0.04597993469238281, 0.04617635345458984, 0.04636092758178711, 0.04602873611450195, 0.046094593048095704, 0.046214176177978517, 0.04620540618896484, 0.046414337158203124, 0.04656332778930664, 0.04659423828125, 0.04630886459350586, 0.04629052734375, 0.04627920150756836, 0.046602432250976565, 0.046956703186035155, 0.04686627197265625, 0.046723072052001956, 0.046835712432861325, 0.04675174331665039, 0.04644019317626953, 0.04631577682495117, 0.04640563201904297, 0.04634771347045898, 0.04648739242553711, 0.04634492874145508, 0.04656860733032227, 0.04652732849121094, 0.046635009765625, 0.04697494506835938, 0.04677846527099609, 0.046747230529785154, 0.0466778564453125, 0.0465986557006836, 0.046884864807128904, 0.047101470947265626, 0.04703004837036133, 0.04692243194580078, 0.04698124694824219, 0.04709977722167969, 0.047099903106689454, 0.046852096557617184, 0.04713260650634766, 0.047089729309082035, 0.04699916839599609, 0.04694668960571289, 0.04708681488037109, 0.046955295562744144, 0.04676403045654297, 0.04717158508300781, 0.04718153762817383, 0.047229217529296874, 0.04706860733032227, 0.047003520965576175, 0.047384990692138675, 0.04756099319458008, 0.0473287353515625, 0.047217182159423825, 0.04809164810180664, 0.046617855072021486, 0.046058238983154295, 0.04595711898803711, 0.04606140899658203, 0.045977760314941406, 0.04595711898803711, 0.04585660934448242, 0.04624326324462891, 0.04619241714477539, 0.04619116973876953, 0.04616230392456055, 0.04610793685913086, 0.04618723297119141, 0.04657875061035156, 0.04675171279907227, 0.04671401596069336, 0.046746688842773436, 0.04660019302368164, 0.04665420913696289, 0.046873950958251955, 0.047063713073730466, 0.04682057571411133, 0.04673193740844726, 0.046639232635498046, 0.04659142303466797, 0.04652492904663086, 0.046714847564697265, 0.0465280647277832, 0.0465843505859375, 0.04690668869018555, 0.0468007698059082, 0.04688924789428711, 0.04699107360839844, 0.04694208145141601, 0.046994369506835935, 0.04685996627807617, 0.047056385040283207, 0.04709177780151367, 0.04719488143920898, 0.04715267181396485, 0.04732566452026367, 0.04740425491333008, 0.04741340637207031, 0.047096256256103516, 0.04718963241577148, 0.04701590347290039, 0.04707183837890625, 0.046960639953613284, 0.04712457656860351, 0.04698233413696289, 0.04712499237060547, 0.04708988952636719, 0.04728543853759765, 0.047343807220458986, 0.04734336090087891, 0.047420478820800784, 0.04759142303466797, 0.047453727722167965, 0.04759958267211914, 0.047540542602539065, 0.0474002571105957, 0.04755321502685547, 0.048724063873291014, 0.04682227325439453, 0.046636959075927735, 0.046728862762451175, 0.046213569641113283, 0.04620080184936524, 0.046272705078125, 0.046229343414306644, 0.04642176055908203, 0.046558719635009765, 0.046414592742919925, 0.04636003112792969, 0.047006240844726564, 0.04674483108520508, 0.04697574234008789, 0.04689100646972656, 0.046952449798583984, 0.04684185409545898, 0.04677632141113281, 0.0469667854309082, 0.047221950531005856, 0.04711907196044922, 0.04715724945068359, 0.0469582405090332, 0.04715280151367188, 0.04689385604858398, 0.04697910308837891, 0.04666527938842773, 0.04676428985595703, 0.04676214218139649, 0.04694217681884766, 0.04680707168579101, 0.0466431999206543, 0.04682345581054687, 0.04698704147338867, 0.04714617538452148, 0.0470906867980957, 0.047230976104736325, 0.04712038421630859, 0.0470568962097168, 0.0473636474609375, 0.04744755172729492, 0.04756576156616211, 0.04741030502319336, 0.04722134399414062, 0.04740739059448242, 0.047405055999755856, 0.047669246673583986, 0.04712857437133789, 0.047257598876953126, 0.047144577026367186, 0.047333087921142575, 0.04726211166381836, 0.047143169403076175, 0.04737638473510742, 0.04762419128417969, 0.04753939056396484, 0.047360862731933594, 0.04735382461547852, 0.04731059265136719, 0.04736223983764649, 0.04748223876953125, 0.0475445442199707, 0.04853670501708984, 0.04701248168945313, 0.046185791015625, 0.04621027374267578, 0.04616486358642578, 0.046248798370361326, 0.046341312408447265, 0.046357311248779294, 0.04632342529296875, 0.04640377426147461, 0.04646102523803711, 0.04658499145507813, 0.04655331039428711, 0.04665817642211914, 0.046611457824707034, 0.04657664108276367, 0.046706687927246096, 0.04667801666259765, 0.046652992248535155, 0.04698361587524414, 0.04749107360839844, 0.04720841598510742, 0.047080543518066405, 0.046936286926269534, 0.04684483337402344, 0.046894401550292966, 0.04694188690185547, 0.04696271896362305, 0.0466885757446289, 0.04684233474731445, 0.04717772674560547, 0.04699135971069336, 0.046706687927246096, 0.04650188827514649, 0.04665139389038086, 0.04663641738891602, 0.04648819351196289, 0.046524417877197265, 0.04674764633178711, 0.047537311553955075, 0.04712326431274414, 0.047478816986083985, 0.04728416061401367, 0.04684601593017578, 0.047323135375976565, 0.046835712432861325, 0.04685593414306641, 0.046970497131347655, 0.04687731170654297, 0.0467949104309082, 0.046922752380371094, 0.04699836730957031, 0.04702764892578125, 0.04691001510620117, 0.0468513298034668, 0.046727935791015626, 0.04718735885620117, 0.04701203155517578, 0.04694803237915039, 0.047104286193847655, 0.04710044860839844, 0.047230880737304685, 0.0473436164855957, 0.048180767059326175, 0.047288352966308594, 0.04623619079589844, 0.04670022583007812, 0.04611718368530274, 0.04605718231201172, 0.046182689666748045, 0.045983200073242185, 0.046325408935546875, 0.046159870147705076, 0.046182689666748045, 0.04687472152709961, 0.04628879928588867, 0.046136161804199216, 0.0461923828125, 0.04648067092895508, 0.0463326416015625, 0.046430206298828124, 0.04658380889892578, 0.04667334365844727, 0.04673593521118164, 0.04682547378540039, 0.04678678512573242, 0.04673680114746094, 0.04646105575561523, 0.04631299209594727, 0.04646166229248047, 0.046462303161621095, 0.0463919677734375, 0.04666777420043945, 0.04651200103759766, 0.04669862365722656, 0.04664934539794922, 0.046413822174072264, 0.04664115142822266, 0.046796672821044924, 0.046755455017089845, 0.04684592056274414, 0.047129119873046875, 0.04713452911376953, 0.047496734619140626, 0.04745891189575195, 0.04741535949707031, 0.047328575134277344, 0.047200958251953126, 0.04694156646728516, 0.047075969696044925, 0.04699523162841797, 0.04706035232543945, 0.04746326446533203, 0.04729811096191406, 0.04711673736572266, 0.0470252799987793, 0.04721779251098633, 0.04720544052124023, 0.047306785583496096, 0.04720220947265625, 0.04948384094238281, 0.04874310302734375, 0.0472597770690918, 0.047451007843017576, 0.04776028823852539, 0.047513694763183595, 0.04848054504394531, 0.0467861442565918, 0.04626454544067383, 0.04634223937988281, 0.04628595352172851, 0.046324542999267575, 0.04638127899169922, 0.04650908660888672, 0.04653955078125, 0.04675971221923828, 0.04687689590454101, 0.04659827041625977, 0.04665315246582031, 0.04696092987060547, 0.04662681579589844, 0.046583553314208985, 0.04687692642211914, 0.046886913299560545, 0.04662243270874023, 0.046975265502929686, 0.04703228759765625, 0.047285633087158205, 0.0469466552734375, 0.046811168670654296, 0.046644542694091795, 0.046762977600097656, 0.04671897506713867, 0.04665958404541016, 0.04730470275878906, 0.046982177734375, 0.04720640182495117, 0.04689599990844726, 0.047887489318847655, 0.04728316879272461, 0.04692377471923828, 0.046970977783203124, 0.04709120178222656, 0.04726416015625, 0.04719404983520508, 0.04802975845336914, 0.04727340698242188, 0.04769055938720703, 0.047324928283691406, 0.04739465713500977, 0.04714672088623047, 0.04736454391479492, 0.04728422546386719, 0.04704051208496094, 0.04717772674560547, 0.0472591667175293, 0.04723555374145508, 0.04725964736938477, 0.047365856170654294, 0.047153438568115234, 0.04736115264892578, 0.04742015838623047, 0.0473289909362793, 0.04731536102294922, 0.04739670562744141, 0.04727004623413086, 0.047685630798339845, 0.047628288269042966, 0.0475074577331543, 0.048653087615966796, 0.046886974334716794, 0.04667862319946289, 0.04627289581298828, 0.04627862548828125, 0.04643129730224609, 0.04657984161376953, 0.046465854644775394, 0.04653609466552734, 0.04644720077514648, 0.04671043014526367, 0.04664169692993164, 0.04664118576049805, 0.046626750946044924, 0.046567264556884765, 0.04695241546630859, 0.04703171157836914, 0.04704288101196289, 0.04691366577148438, 0.04694009780883789, 0.047867904663085936, 0.04702848052978516, 0.046870529174804686, 0.04710400009155274, 0.04673535919189453, 0.04680908966064453, 0.0467720947265625, 0.04682281494140625, 0.046970752716064455, 0.04705094528198242, 0.046758560180664065, 0.04679884719848633, 0.04684799957275391, 0.047026176452636716, 0.04710995101928711, 0.046984512329101565, 0.04690995025634766, 0.047110111236572265, 0.046946529388427735, 0.04733910369873047, 0.04738227081298828, 0.047395679473876955, 0.04742694473266602, 0.04741353607177735, 0.04741360092163086, 0.04702207946777344, 0.04707084655761719, 0.04716787338256836, 0.04711734390258789, 0.047372768402099606, 0.04719209671020508, 0.04696931076049805, 0.04720025634765625, 0.04723507308959961, 0.047343902587890625, 0.047832992553710936, 0.0473741455078125, 0.047865375518798825, 0.04770588684082031, 0.04754713439941406, 0.04782688140869141, 0.04764672088623047, 0.047693824768066405, 0.04885590362548828, 0.04704585647583008, 0.046653247833251955, 0.04640252685546875, 0.04646841430664062, 0.04648006439208984, 0.046448638916015625, 0.04653875350952148, 0.04646092987060547, 0.04650137710571289, 0.046407455444335936, 0.046400062561035155, 0.046626976013183594, 0.04677836990356445, 0.046910911560058596, 0.04670659255981445, 0.04689337539672851, 0.04677785491943359, 0.046785377502441404, 0.04705887985229492, 0.04739692687988281, 0.04735795211791992, 0.04698521423339844, 0.04684524917602539, 0.04713062286376953, 0.046754142761230466, 0.04676406478881836, 0.046806785583496095, 0.04678915023803711, 0.04674256134033203, 0.046717952728271485, 0.046653438568115234, 0.04671680068969727, 0.046737537384033204, 0.04684492874145508, 0.04688115310668945, 0.04687936019897461, 0.046973953247070314, 0.04695462417602539, 0.046855072021484374, 0.0469257926940918, 0.04731891250610352, 0.047371681213378904, 0.047417278289794924, 0.04724531173706055, 0.04713347244262695, 0.04715660858154297, 0.04734835052490234, 0.047164737701416014, 0.04728070449829101, 0.04728435134887695, 0.04717772674560547, 0.047034366607666016, 0.04712963104248047, 0.04730569458007813, 0.047314239501953126, 0.04723987197875976, 0.047331329345703124, 0.0473436164855957, 0.04738662338256836, 0.04751564788818359, 0.04754022216796875, 0.04757632064819336]",tokens/s,21.322464920321327,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,8211.062784,12133.00736,0.0,11737.759744,11171.24352,s,1,16.47221875,16.47221875,0.0,16.47221875,16.47221875,16.47221875,16.47221875,[16.47221875],,kWh,0.0002681633353000355,2.9572969375535115e-05,0.00011389814667406295,0.0004116344513496336,,MB,4054.339584,12147.687424,0.0,11739.856896,10924.35968,s,10,58.20302490234375,5.820302490234375,0.0018394682126367818,5.820822265625,5.822356005859374,5.8225693603515625,5.822740043945312,"[5.817763671875, 5.8176748046875, 5.81830908203125, 5.81912646484375, 5.8204892578125, 5.82135888671875, 5.82205615234375, 5.8211552734375, 5.82278271484375, 5.82230859375]",tokens/s,43.98396826101924,kWh,0.00016975931377415842,1.8724993909382963e-05,0.00011283195137660185,0.0003013162590601432,tokens/kWh,849605.6628291737,MB,4058.60352,12151.881728,0.0,11741.954048,10924.36224,s,10,29.493633300781255,2.9493633300781257,0.0030624496666022006,2.9485103759765625,2.95195400390625,2.9544826171875,2.9565055078125,"[2.947757568359375, 2.945792724609375, 2.95139208984375, 2.947404052734375, 2.946928466796875, 2.949522705078125, 2.94754345703125, 2.94926318359375, 2.95701123046875, 2.951017822265625]",tokens/s,21.360542242291725,kWh,8.640699360418382e-05,9.531341060527186e-06,5.753701825179645e-05,0.0001534753529165074,tokens/kWh,410489.36394544615,,s,630,29.49085386276244,0.04681087914724198,0.0004593725109467762,0.04678553581237793,0.04721825866699219,0.047367992210388184,0.048674534416198734,"[0.04866867065429688, 0.046413822174072264, 0.04618345642089844, 0.04651516723632813, 0.04608572769165039, 0.04632783889770508, 0.046125438690185545, 0.04614339065551758, 0.04617942428588867, 0.04675888061523437, 0.04648361587524414, 0.0464087028503418, 0.046726016998291015, 0.04683327865600586, 0.04625241470336914, 0.04638505554199219, 0.04637868881225586, 0.04636016082763672, 0.04645151901245117, 0.04685004806518555, 0.04707328033447265, 0.046903297424316405, 0.046712833404541014, 0.046407550811767576, 0.04621939086914063, 0.047035808563232424, 0.04671139144897461, 0.04629199981689453, 0.046486526489257815, 0.04717772674560547, 0.04672918319702148, 0.04671897506713867, 0.04680089569091797, 0.046683872222900394, 0.046765567779541016, 0.04692863845825195, 0.04672707366943359, 0.046833599090576175, 0.04679478454589844, 0.04696035385131836, 0.04716527938842773, 0.04724787139892578, 0.04719011306762695, 0.04705254364013672, 0.046885120391845704, 0.04702412796020508, 0.04683980941772461, 0.04697907257080078, 0.04659811019897461, 0.04688598251342774, 0.047166015625, 0.04697238540649414, 0.0468427848815918, 0.04683331298828125, 0.04686617660522461, 0.04737699127197265, 0.04708563232421875, 0.04683564758300781, 0.046832958221435544, 0.04742828750610351, 0.04734956741333008, 0.04744790267944336, 0.047097408294677734, 0.04867692947387695, 0.04647366333007812, 0.04609417724609375, 0.04623987197875976, 0.04605132675170898, 0.04640768051147461, 0.04640476989746094, 0.04623846435546875, 0.04634838485717773, 0.04682060623168945, 0.04675417709350586, 0.046487136840820314, 0.046445343017578126, 0.04641996765136719, 0.046440448760986325, 0.04660960006713867, 0.046319873809814456, 0.04663929748535156, 0.046817569732666015, 0.04691567993164063, 0.046688255310058595, 0.046903297424316405, 0.04663100814819336, 0.046342048645019535, 0.0462479362487793, 0.04661372756958008, 0.04653100967407227, 0.04643203353881836, 0.04660076904296875, 0.04679679870605469, 0.046486751556396484, 0.0469587516784668, 0.046913505554199215, 0.04678118515014648, 0.046655391693115236, 0.0466429443359375, 0.04671100616455078, 0.0467108154296875, 0.047065086364746093, 0.04709523010253906, 0.04705046463012695, 0.046965599060058594, 0.04717977523803711, 0.047134719848632815, 0.04689715194702149, 0.046680065155029295, 0.04660224151611328, 0.04666163253784179, 0.04664476776123047, 0.04689750289916992, 0.04720243072509766, 0.04700495910644531, 0.04681596755981445, 0.04680089569091797, 0.046747711181640624, 0.047248481750488285, 0.04703228759765625, 0.0468645133972168, 0.04677299118041992, 0.04712262344360352, 0.04713862228393555, 0.04715068817138672, 0.047429759979248046, 0.04886294555664063, 0.0467092170715332, 0.04633411026000977, 0.046499839782714845, 0.04631350326538086, 0.046235614776611325, 0.04637865447998047, 0.04649324798583984, 0.0463056640625, 0.04674806213378906, 0.04651753616333008, 0.046367454528808597, 0.04650608062744141, 0.04646083068847656, 0.04641753768920898, 0.04683200073242187, 0.04684799957275391, 0.04668627166748047, 0.04635836791992187, 0.046589344024658204, 0.04654473495483399, 0.04640444946289062, 0.0465715217590332, 0.04648287963867188, 0.046405406951904295, 0.04699625778198242, 0.046721023559570314, 0.04642201614379883, 0.04657561492919922, 0.046992576599121094, 0.04655718231201172, 0.04666041564941406, 0.04822220611572266, 0.046723072052001956, 0.04652032089233398, 0.04722073745727539, 0.046960033416748044, 0.04709404754638672, 0.0467542724609375, 0.04675363159179687, 0.047177345275878906, 0.04733779144287109, 0.04723686218261719, 0.04687699127197266, 0.04723497772216797, 0.04703241729736328, 0.046843902587890625, 0.046680065155029295, 0.04712243270874023, 0.04709785461425781, 0.046677696228027345, 0.046765567779541016, 0.046885025024414065, 0.04709872055053711, 0.04678047943115234, 0.046849246978759765, 0.04732368087768555, 0.0474683837890625, 0.04727619171142578, 0.047675392150878904, 0.04728400039672852, 0.04720374298095703, 0.047075969696044925, 0.048652256011962894, 0.04674291229248047, 0.04641241455078125, 0.04623974227905273, 0.04607385635375977, 0.046415870666503906, 0.04646883010864258, 0.04625187301635742, 0.04642211151123047, 0.04685654449462891, 0.04661164855957031, 0.04655596923828125, 0.046784446716308596, 0.04651219177246094, 0.0463296012878418, 0.04679724884033203, 0.04680275344848633, 0.04639289474487305, 0.046338687896728514, 0.04690719985961914, 0.04664115142822266, 0.046632190704345704, 0.04668806457519531, 0.0465880012512207, 0.046500320434570315, 0.04649155044555664, 0.04653104019165039, 0.04674745559692383, 0.04694355010986328, 0.04675859069824219, 0.04644473648071289, 0.046769729614257814, 0.04692559814453125, 0.046502655029296874, 0.04638300704956055, 0.04673535919189453, 0.04670259094238281, 0.0466247673034668, 0.046524417877197265, 0.04655516815185547, 0.04703638458251953, 0.04714713668823242, 0.04692892837524414, 0.04720441436767578, 0.0471429443359375, 0.04685696029663086, 0.04677552032470703, 0.04713865661621094, 0.046934623718261716, 0.04724361419677734, 0.046960639953613284, 0.04677017593383789, 0.04685619354248047, 0.04731635284423828, 0.04694899368286133, 0.046884864807128904, 0.04680704116821289, 0.0473675537109375, 0.0471385612487793, 0.04702230453491211, 0.04691603088378906, 0.04737014389038086, 0.047129951477050784, 0.04873878479003906, 0.04679471969604492, 0.04648537445068359, 0.04627878570556641, 0.04614352035522461, 0.04615126419067383, 0.04634048080444336, 0.04620268630981445, 0.046018753051757816, 0.046208446502685546, 0.046524673461914065, 0.046303550720214845, 0.04640284729003906, 0.0467279052734375, 0.046565376281738284, 0.046780193328857425, 0.0467371826171875, 0.04668451309204102, 0.04640572738647461, 0.04663849639892578, 0.04688956832885742, 0.04665139389038086, 0.04647292709350586, 0.04689894485473633, 0.046373409271240236, 0.04635772705078125, 0.04659462356567383, 0.04681740951538086, 0.04655728149414062, 0.0467322883605957, 0.046437374114990236, 0.04658111953735351, 0.04693635177612305, 0.046790111541748045, 0.04639628982543945, 0.04705484771728516, 0.047031646728515626, 0.04678662490844727, 0.04703907012939453, 0.046857662200927734, 0.046641727447509766, 0.047239166259765625, 0.04693734359741211, 0.04687497711181641, 0.04672710418701172, 0.04677862548828125, 0.046618526458740234, 0.046839553833007815, 0.04721516799926758, 0.04703641510009766, 0.04683366394042969, 0.04681926345825195, 0.04695251083374023, 0.04719411087036133, 0.047115585327148435, 0.047000255584716794, 0.04730579376220703, 0.04722710418701172, 0.047185985565185544, 0.04742399978637695, 0.04728438568115234, 0.04703641510009766, 0.04700284957885742, 0.04905023956298828, 0.04683967971801758, 0.04607804870605469, 0.04619468688964844, 0.04604431915283203, 0.04626313781738281, 0.04627046585083008, 0.04629094314575195, 0.04619862365722656, 0.046440608978271486, 0.04656243133544922, 0.046519168853759764, 0.046176254272460936, 0.046368255615234374, 0.046793216705322264, 0.046774303436279294, 0.04680643081665039, 0.046647872924804684, 0.046516223907470705, 0.04653055953979492, 0.046935264587402346, 0.04692031860351562, 0.04660847854614258, 0.04704022216796875, 0.046920032501220704, 0.04698316955566406, 0.046639102935791016, 0.04653670501708984, 0.046433311462402344, 0.04691247940063477, 0.047034366607666016, 0.046475265502929686, 0.04665884780883789, 0.04693267059326172, 0.046569503784179685, 0.046852096557617184, 0.046925857543945314, 0.04678857421875, 0.04684358215332031, 0.04704412841796875, 0.047368350982666015, 0.04733737564086914, 0.046975711822509765, 0.04695654296875, 0.04716252899169922, 0.04703228759765625, 0.04678473663330078, 0.04681772613525391, 0.04659222412109375, 0.04653875350952148, 0.04697248077392578, 0.047292865753173825, 0.04692115020751953, 0.04692025756835937, 0.04679270553588867, 0.047484127044677735, 0.047266590118408204, 0.046956127166748046, 0.046975265502929686, 0.04735110473632813, 0.04731711959838867, 0.047049217224121094, 0.04698559951782227, 0.048882720947265625, 0.04644758224487305, 0.04625139236450195, 0.046276832580566404, 0.04634460830688476, 0.046432254791259765, 0.04643635177612305, 0.0461781120300293, 0.046259967803955075, 0.04688032150268555, 0.046501953125, 0.04623443222045898, 0.046383102416992186, 0.046620670318603515, 0.04625423812866211, 0.04638089752197266, 0.04673126220703125, 0.04660019302368164, 0.04645478439331055, 0.0468746223449707, 0.04690739059448242, 0.047124000549316404, 0.04655875015258789, 0.04657843017578125, 0.046260414123535154, 0.04657929611206055, 0.046940353393554686, 0.04646899032592773, 0.04632598495483398, 0.04671017456054687, 0.04702057647705078, 0.04691782379150391, 0.04654006576538086, 0.04677913665771485, 0.047011550903320314, 0.047026432037353516, 0.046701953887939456, 0.046731136322021485, 0.04654975891113281, 0.046755840301513675, 0.047002880096435544, 0.04715971374511719, 0.046774688720703124, 0.04688435363769531, 0.046712993621826175, 0.046997791290283204, 0.04714700698852539, 0.047110145568847656, 0.046773536682128906, 0.04689788818359375, 0.046795936584472654, 0.047008609771728514, 0.04683958435058594, 0.04685372924804688, 0.04699216079711914, 0.047056640625, 0.04725347137451172, 0.047083648681640625, 0.04695654296875, 0.047392704010009765, 0.047179840087890626, 0.04739276885986328, 0.04711196899414063, 0.048344127655029295, 0.04673529434204102, 0.046445568084716796, 0.046344192504882815, 0.0459417610168457, 0.046235904693603516, 0.04639411163330078, 0.04638671875, 0.04633430480957031, 0.04691571044921875, 0.0464257926940918, 0.04695199966430664, 0.04660857772827148, 0.04642387390136719, 0.04646915054321289, 0.04685491180419922, 0.046827167510986326, 0.04656159973144531, 0.0466063346862793, 0.04688076782226563, 0.046685791015625, 0.046534847259521485, 0.046567649841308595, 0.046637054443359374, 0.046465023040771485, 0.04649273681640625, 0.04666476821899414, 0.04668403244018555, 0.04670259094238281, 0.04687033462524414, 0.046700736999511716, 0.047050750732421875, 0.046565025329589844, 0.0465145263671875, 0.046704639434814454, 0.046919166564941404, 0.04685260772705078, 0.04714707183837891, 0.047052734375, 0.04679884719848633, 0.047156448364257815, 0.047012287139892577, 0.04668412780761719, 0.04719244766235352, 0.04690943908691406, 0.04670374298095703, 0.046636959075927735, 0.04671526336669922, 0.04672572708129883, 0.047324928283691406, 0.04713894271850586, 0.04669353485107422, 0.04694720077514648, 0.04696227264404297, 0.04673766326904297, 0.04735820770263672, 0.047185726165771484, 0.04712419128417969, 0.0473864631652832, 0.0474582405090332, 0.04705472183227539, 0.04748166275024414, 0.04711139297485351, 0.04863577651977539, 0.04684812927246094, 0.04640768051147461, 0.046176254272460936, 0.04609209442138672, 0.04646873474121094, 0.04643692779541016, 0.04636671829223633, 0.052022911071777346, 0.04558067321777344, 0.046456127166748046, 0.04674956893920899, 0.04658448028564453, 0.04659574508666992, 0.04642828750610352, 0.04681270217895508, 0.04681814575195312, 0.04677427291870117, 0.04638105773925781, 0.046617984771728516, 0.04688272094726562, 0.04701660919189453, 0.046609886169433595, 0.046639713287353515, 0.04668620681762695, 0.04647731018066406, 0.04665273666381836, 0.046792510986328126, 0.04674035263061523, 0.04675945663452148, 0.04683209609985352, 0.04652844619750977, 0.04671920013427734, 0.04699081420898438, 0.046537086486816405, 0.04678192138671875, 0.04714275360107422, 0.04692464065551758, 0.04711103820800781, 0.047158241271972656, 0.04685785675048828, 0.04689686584472656, 0.047186016082763675, 0.04699363327026367, 0.046903518676757815, 0.04690752029418945, 0.04694220733642578, 0.046849822998046874, 0.04672463989257813, 0.04695667266845703, 0.047343936920166016, 0.04715340805053711, 0.04701388931274414, 0.04707328033447265, 0.04721798324584961, 0.047130561828613284, 0.04701686477661133, 0.04744927978515625, 0.04709238433837891, 0.04749926376342774, 0.04741939163208008, 0.047351806640625, 0.04750140762329102, 0.04920111846923828, 0.04674150466918945, 0.046599937438964845, 0.046352577209472653, 0.046182464599609375, 0.046475265502929686, 0.046712833404541014, 0.04674355316162109, 0.04636569595336914, 0.0469486083984375, 0.04654265594482422, 0.04623455810546875, 0.046698593139648435, 0.04640143966674805, 0.04634406280517578, 0.04679212951660156, 0.04658844757080078, 0.046610591888427734, 0.046458881378173826, 0.04651827239990235, 0.046911102294921875, 0.04702777481079102, 0.046938209533691405, 0.046897598266601566, 0.04673974227905273, 0.046671424865722656, 0.046690750122070315, 0.04667596817016602, 0.0470384635925293, 0.04697907257080078, 0.04678656005859375, 0.04669440078735351, 0.04678633499145508, 0.04656560134887695, 0.04672512054443359, 0.047376415252685544, 0.046575584411621095, 0.047132671356201174, 0.04680019378662109, 0.046448543548583986, 0.04674828720092773, 0.046975135803222656, 0.04703619384765625, 0.047223007202148434, 0.04700364685058594, 0.046655487060546875, 0.04681932830810547, 0.04698128128051758, 0.04682735824584961, 0.047263904571533205, 0.04724671936035156, 0.04694803237915039, 0.04688566589355469, 0.047082687377929686, 0.0467869758605957, 0.046973342895507815, 0.04691353607177735, 0.04690124893188476, 0.04735145568847656, 0.04720793533325195, 0.04681404876708985, 0.04702572631835938, 0.04707980728149414]",tokens/s,21.362555419105355,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,837.541888,558.825472,0.0,163.577856,152.009216,s,1,7.91282421875,7.91282421875,0.0,7.91282421875,7.91282421875,7.91282421875,7.91282421875,[7.91282421875],,kWh,2.1531786091691175e-05,2.3678079476961558e-06,7.5216726840021675e-06,3.14212667233895e-05,,MB,1184.538624,609.15712,0.0,201.326592,184.525824,s,25,0.26236425590515133,0.010494570236206056,0.00017778441085430742,0.010455743789672851,0.010575654602050781,0.010639872360229492,0.011143333168029785,"[0.011298399925231933, 0.010406208038330079, 0.010385151863098144, 0.010464159965515137, 0.010456607818603516, 0.010431903839111328, 0.010458687782287597, 0.010485952377319335, 0.010497440338134765, 0.010553824424743652, 0.010455743789672851, 0.010590208053588868, 0.010537983894348145, 0.010449248313903809, 0.01038265609741211, 0.01041494369506836, 0.010396063804626465, 0.010374976158142089, 0.010462847709655762, 0.010541376113891601, 0.010652288436889649, 0.010414463996887207, 0.01037007999420166, 0.010439807891845704, 0.010443231582641602]",tokens/s,24393.566791025434,kWh,3.431983375147018e-07,3.783589778108301e-08,2.2630583452407013e-07,6.07340069819855e-07,tokens/kWh,421510143.52788043,MB,1218.269184,609.15712,0.0,201.326592,184.528384,s,25,10.12959649658203,0.40518385986328126,0.004919799178749007,0.40542120361328127,0.40999371948242186,0.4109174865722656,0.4166025048828125,"[0.4047490234375, 0.39681671142578123, 0.4053542175292969, 0.4055596923828125, 0.4014530944824219, 0.4110566711425781, 0.4010058898925781, 0.4103607482910156, 0.4067662048339844, 0.40542120361328127, 0.4089087219238281, 0.4082795104980469, 0.40944317626953125, 0.40081253051757815, 0.407805908203125, 0.39951266479492187, 0.3957163696289063, 0.4001940307617188, 0.41835382080078126, 0.4073577880859375, 0.4045588073730469, 0.3997763061523438, 0.40915234375, 0.40476751708984376, 0.4064135437011719]",tokens/s,155.48496927112967,kWh,1.1434372355484986e-05,1.2610269749405768e-06,4.364320060197409e-06,1.705971939062297e-05,tokens/kWh,3692909.5114324405,,s,1575,10.117378850460058,0.0064237326034667,0.00023948435789487624,0.006399040222167969,0.006530944156646728,0.006600761699676514,0.007333048276901245,"[0.006305280208587646, 0.0065354561805725095, 0.006521312236785889, 0.006501344203948975, 0.006511392116546631, 0.006486015796661377, 0.006481696128845215, 0.006534880161285401, 0.006525055885314941, 0.006526336193084716, 0.0065465598106384274, 0.006597856044769287, 0.006517407894134522, 0.006493696212768555, 0.00643123197555542, 0.006483615875244141, 0.006457759857177735, 0.006574368000030518, 0.006440608024597168, 0.00637065601348877, 0.006427296161651611, 0.006379519939422608, 0.006360608100891113, 0.00641209602355957, 0.006372064113616943, 0.006414432048797608, 0.006438464164733887, 0.006332543849945069, 0.006367392063140869, 0.00636518383026123, 0.006395808219909668, 0.006373472213745117, 0.006362720012664795, 0.006395999908447266, 0.006375775814056396, 0.006434207916259766, 0.006370175838470459, 0.006403840065002441, 0.00637667179107666, 0.006382304191589356, 0.006423679828643799, 0.006372223854064942, 0.006340288162231446, 0.006373407840728759, 0.007237919807434082, 0.006307839870452881, 0.006292768001556396, 0.0062507839202880855, 0.006257279872894287, 0.006257823944091797, 0.006307680130004883, 0.006287551879882812, 0.006292096138000488, 0.006313983917236328, 0.006268896102905273, 0.006352287769317627, 0.0063218560218811035, 0.0063879361152648926, 0.00639628791809082, 0.006392159938812256, 0.00649948787689209, 0.006372191905975342, 0.006391808032989502, 0.005998879909515381, 0.006312960147857666, 0.0062576642036437985, 0.00629750394821167, 0.006269023895263672, 0.006330592155456543, 0.0063357119560241695, 0.006334047794342041, 0.0063497920036315915, 0.00633241605758667, 0.006305856227874756, 0.006287295818328858, 0.006290592193603515, 0.0062841281890869145, 0.006221151828765869, 0.006253536224365234, 0.006227263927459717, 0.006281599998474121, 0.006229055881500244, 0.006246912002563477, 0.006296031951904297, 0.006256608009338379, 0.0066007041931152345, 0.006301375865936279, 0.006256159782409668, 0.006244128227233887, 0.006228127956390381, 0.006237023830413818, 0.006292992115020752, 0.006322976112365723, 0.0062353920936584475, 0.006261216163635254, 0.006264832019805908, 0.0062873601913452145, 0.006300992012023926, 0.006269728183746338, 0.006344607830047607, 0.006375423908233643, 0.006268928050994873, 0.00623638391494751, 0.006295296192169189, 0.006239520072937012, 0.006198016166687012, 0.006273344039916993, 0.006276927947998047, 0.006487872123718262, 0.0062362241744995115, 0.006253888130187988, 0.00623686408996582, 0.006352511882781983, 0.006300000190734864, 0.0062746238708496094, 0.006266848087310791, 0.006320576190948487, 0.006326335906982422, 0.006445184230804444, 0.006537087917327881, 0.006338431835174561, 0.006286655902862549, 0.006255551815032959, 0.006297696113586426, 0.0062708802223205564, 0.006265984058380127, 0.006023168087005615, 0.006303743839263916, 0.006244031906127929, 0.00687340784072876, 0.006317984104156494, 0.006270336151123047, 0.0063170561790466305, 0.006307551860809326, 0.0063482880592346195, 0.00627180814743042, 0.006311615943908691, 0.006292992115020752, 0.006304255962371826, 0.006307424068450928, 0.006318016052246094, 0.0068793601989746095, 0.0064617919921875, 0.006436863899230957, 0.006463615894317627, 0.006422656059265137, 0.006534207820892334, 0.006388031959533692, 0.006361663818359375, 0.006393631935119629, 0.006320159912109375, 0.006387712001800537, 0.006362880229949951, 0.0063480958938598635, 0.006427487850189209, 0.006330463886260986, 0.006366367816925049, 0.006438752174377441, 0.006466559886932373, 0.006415679931640625, 0.0063731842041015625, 0.00646611213684082, 0.006359360218048096, 0.00635920000076294, 0.006408160209655762, 0.006342559814453125, 0.0064297599792480465, 0.006490784168243408, 0.006420735836029053, 0.006446656227111817, 0.00642409610748291, 0.0063743038177490234, 0.006381343841552734, 0.006484479904174805, 0.006397151947021484, 0.006428607940673828, 0.00639958381652832, 0.00643779182434082, 0.006924352169036865, 0.006443295955657959, 0.006440767765045166, 0.006375328063964843, 0.006415359973907471, 0.006376448154449463, 0.007439775943756104, 0.0064704642295837405, 0.006706975936889649, 0.006422495841979981, 0.0063918399810791015, 0.006064991950988769, 0.006389696121215821, 0.006326335906982422, 0.0063610877990722655, 0.006311552047729492, 0.0063942399024963376, 0.006354944229125976, 0.0063119039535522465, 0.007235616207122803, 0.006341663837432862, 0.006306272029876709, 0.006356800079345703, 0.006322336196899414, 0.006351359844207764, 0.006332704067230225, 0.006354688167572021, 0.006342656135559082, 0.006329951763153076, 0.00633516788482666, 0.006583424091339112, 0.006382175922393799, 0.0064245758056640625, 0.006468671798706055, 0.006483007907867431, 0.006469503879547119, 0.0064412798881530765, 0.00643884801864624, 0.006432064056396484, 0.006513088226318359, 0.006451295852661133, 0.006424479961395264, 0.006455296039581298, 0.006445055961608887, 0.006479328155517578, 0.006472224235534668, 0.0064430079460144046, 0.0064471039772033695, 0.006416384220123291, 0.006450623989105225, 0.006442560195922852, 0.006429728031158447, 0.006445024013519287, 0.0064430079460144046, 0.006401055812835693, 0.006417376041412353, 0.0064471039772033695, 0.006521024227142334, 0.006475584030151367, 0.0064143362045288085, 0.006536831855773926, 0.006467936038970948, 0.006513728141784668, 0.006461503982543945, 0.00649513578414917, 0.006438144207000732, 0.006433568000793457, 0.0064587841033935545, 0.0064635519981384275, 0.006426144123077393, 0.006439328193664551, 0.00641648006439209, 0.006402368068695068, 0.006420544147491455, 0.006106048107147217, 0.006385439872741699, 0.006408095836639404, 0.00637724781036377, 0.006385727882385254, 0.006426784038543701, 0.006569183826446533, 0.006396031856536865, 0.006418496131896973, 0.006429279804229736, 0.0063772802352905274, 0.006410560131072998, 0.006352767944335937, 0.00640121603012085, 0.006396192073822022, 0.006349343776702881, 0.006376480102539063, 0.006353888034820556, 0.0065989761352539065, 0.006379199981689453, 0.006359039783477783, 0.0063569917678833006, 0.00638486385345459, 0.006453055858612061, 0.006418560028076172, 0.006411104202270508, 0.0064143362045288085, 0.006361216068267822, 0.006377312183380127, 0.006397984027862549, 0.00633241605758667, 0.006313471794128418, 0.006300511837005615, 0.0063463678359985355, 0.00629804801940918, 0.006374623775482178, 0.006369247913360596, 0.006348415851593018, 0.006350719928741455, 0.006346687793731689, 0.006374495983123779, 0.006340479850769043, 0.006324543952941895, 0.0063422398567199705, 0.0063339838981628415, 0.006414912223815918, 0.006334752082824707, 0.006327616214752197, 0.006293087959289551, 0.006302080154418945, 0.0063240318298339844, 0.0063220481872558595, 0.006345056056976319, 0.0063564801216125484, 0.006383520126342773, 0.006355967998504639, 0.006313983917236328, 0.00635475206375122, 0.0063203201293945315, 0.006292895793914795, 0.006303616046905517, 0.00632528018951416, 0.006364128112792969, 0.006125696182250976, 0.006363135814666748, 0.0063201279640197755, 0.006372960090637207, 0.006312352180480957, 0.006339968204498291, 0.006368959903717041, 0.00669536018371582, 0.006439424037933349, 0.006405151844024658, 0.006427487850189209, 0.006493599891662598, 0.0065474557876586915, 0.006584479808807373, 0.006571616172790527, 0.006564832210540771, 0.0072353601455688475, 0.007833824157714844, 0.0075443840026855466, 0.007920063972473145, 0.006651904106140137, 0.006569983959197998, 0.006600895881652832, 0.006577792167663574, 0.006838655948638916, 0.00661411190032959, 0.006529248237609863, 0.006517248153686523, 0.006485951900482178, 0.006420544147491455, 0.0064633598327636715, 0.006430912017822266, 0.006431807994842529, 0.00639408016204834, 0.007348896026611328, 0.006415840148925781, 0.006425248146057129, 0.006344575881958008, 0.006391744136810303, 0.0063591041564941405, 0.006378592014312744, 0.00639628791809082, 0.006359583854675293, 0.006395904064178467, 0.006371327877044678, 0.006374527931213379, 0.0063515200614929195, 0.006344064235687256, 0.006370143890380859, 0.006370880126953125, 0.0063455038070678715, 0.006339392185211182, 0.006356128215789795, 0.0063721280097961425, 0.006340640068054199, 0.006337408065795898, 0.006360095977783203, 0.006355936050415039, 0.00638976001739502, 0.006323552131652832, 0.006372000217437744, 0.006342656135559082, 0.006352128028869629, 0.006159071922302246, 0.006537312030792236, 0.006510496139526367, 0.006452672004699707, 0.006472352027893066, 0.006419904232025146, 0.006364863872528076, 0.006342495918273926, 0.0063681921958923336, 0.006338560104370118, 0.006307616233825684, 0.00632041597366333, 0.006389472007751465, 0.006322400093078613, 0.00633241605758667, 0.006291647911071777, 0.006373023986816407, 0.006324384212493896, 0.006336832046508789, 0.006313663959503174, 0.006311520099639893, 0.0063162240982055666, 0.006342271804809571, 0.006287168025970459, 0.006298399925231933, 0.006278783798217774, 0.006293407917022705, 0.006283616065979004, 0.006325664043426514, 0.006316991806030273, 0.006323200225830078, 0.006367839813232422, 0.006336639881134034, 0.006514272212982178, 0.006329152107238769, 0.006358687877655029, 0.006366496086120606, 0.006318111896514892, 0.006341311931610107, 0.006303743839263916, 0.006344704151153564, 0.006308896064758301, 0.006288352012634277, 0.006334496021270752, 0.00630508804321289, 0.006611616134643555, 0.006368480205535889, 0.006305791854858398, 0.006330848217010498, 0.006347263813018799, 0.0063240318298339844, 0.00627836799621582, 0.006355743885040283, 0.006352735996246338, 0.006379039764404297, 0.0064139838218688966, 0.006375840187072754, 0.006418015956878662, 0.006454239845275879, 0.0064102401733398436, 0.006488351821899414, 0.0063935680389404295, 0.006430751800537109, 0.00617196798324585, 0.006466239929199219, 0.006444767951965332, 0.006418655872344971, 0.006453311920166015, 0.006490111827850342, 0.006491136074066162, 0.006468192100524902, 0.006441023826599121, 0.006479904174804687, 0.006461472034454346, 0.006422848224639893, 0.006473504066467285, 0.006455488204956054, 0.006448416233062744, 0.006453055858612061, 0.006448287963867188, 0.0064962878227233884, 0.006464928150177002, 0.00654585599899292, 0.00651251220703125, 0.0064778242111206055, 0.0070388798713684085, 0.006533215999603272, 0.006616384029388428, 0.006566624164581299, 0.006537439823150634, 0.0065780158042907716, 0.006536223888397217, 0.006520768165588379, 0.006529983997344971, 0.006526976108551025, 0.006475168228149414, 0.006475967884063721, 0.0064757437705993655, 0.006416831970214844, 0.0064899840354919434, 0.006459392070770263, 0.006406271934509277, 0.006399744033813477, 0.006417984008789062, 0.006436927795410156, 0.00646998405456543, 0.006421792030334473, 0.006381824016571045, 0.006408768177032471, 0.006606592178344727, 0.0064947838783264164, 0.006913919925689697, 0.0065474557876586915, 0.006494207859039307, 0.006528639793395996, 0.0065289921760559085, 0.006539680004119873, 0.006474944114685058, 0.00641926383972168, 0.0063751678466796875, 0.0064956479072570805, 0.006479872226715088, 0.006486879825592041, 0.0064791679382324215, 0.007381760120391846, 0.006494143962860108, 0.0061975998878479, 0.006475967884063721, 0.0064766077995300295, 0.006500895977020264, 0.006492767810821533, 0.006649119853973389, 0.006498784065246582, 0.006487040042877197, 0.006455967903137207, 0.006463039875030518, 0.006449952125549317, 0.0064204797744750975, 0.006454495906829834, 0.006445856094360352, 0.006408415794372559, 0.006444831848144531, 0.006395391941070557, 0.00644927978515625, 0.0064347200393676755, 0.006389567852020264, 0.006436607837677002, 0.006402976036071777, 0.006422048091888428, 0.006447616100311279, 0.006422751903533935, 0.006493951797485352, 0.006491360187530518, 0.006484767913818359, 0.006470655918121338, 0.006462463855743408, 0.006450240135192871, 0.006425536155700683, 0.00643071985244751, 0.006518784046173095, 0.006450592041015625, 0.006466400146484375, 0.006505824089050293, 0.0065335359573364254, 0.006474783897399902, 0.006454112052917481, 0.006408160209655762, 0.006415999889373779, 0.0063903040885925294, 0.006411871910095215, 0.006399648189544678, 0.006400767803192139, 0.0064280638694763186, 0.0064496641159057615, 0.006446591854095459, 0.006435679912567139, 0.006463232040405273, 0.00645743989944458, 0.0064429759979248045, 0.006442944049835205, 0.006453120231628418, 0.006459807872772217, 0.006460800170898438, 0.006441311836242676, 0.006520832061767578, 0.006454592227935791, 0.006388415813446045, 0.006430592060089111, 0.006439104080200196, 0.006159840106964111, 0.006457952022552491, 0.0063961601257324216, 0.006381599903106689, 0.006352735996246338, 0.0063949117660522465, 0.006427487850189209, 0.006367231845855713, 0.006380703926086426, 0.006363999843597412, 0.006383264064788818, 0.006433375835418701, 0.006419392108917237, 0.006650335788726806, 0.006618879795074463, 0.00650710391998291, 0.006475776195526123, 0.006450399875640869, 0.006422304153442383, 0.006445280075073242, 0.0064048957824707035, 0.006389887809753418, 0.006410111904144287, 0.006516863822937011, 0.006469759941101074, 0.006416224002838135, 0.006433951854705811, 0.006415103912353516, 0.006465536117553711, 0.006405344009399414, 0.0064354238510131835, 0.006391295909881592, 0.006383967876434326, 0.006420127868652344, 0.006445759773254394, 0.00643891191482544, 0.006412288188934326, 0.0064245758056640625, 0.006430016040802002, 0.006410943984985352, 0.006400032043457031, 0.0064141440391540525, 0.006350912094116211, 0.006379615783691406, 0.006402048110961914, 0.00638976001739502, 0.006440864086151123, 0.0063714241981506345, 0.006625408172607422, 0.006430751800537109, 0.006403584003448487, 0.00644547176361084, 0.006409823894500732, 0.006363615989685058, 0.006475647926330566, 0.006390079975128173, 0.006401631832122803, 0.006508800029754638, 0.006373504161834717, 0.006438432216644287, 0.006506688117980957, 0.006504447937011719, 0.006475456237792969, 0.006160384178161621, 0.006466815948486328, 0.006494624137878418, 0.006488224029541015, 0.00645907211303711, 0.0065090560913085935, 0.006462783813476563, 0.006459296226501465, 0.006526048183441162, 0.006377344131469727, 0.006416192054748535, 0.006402048110961914, 0.0064215359687805175, 0.006573247909545899, 0.0065000319480896, 0.006475488185882569, 0.006447679996490478, 0.006430528163909912, 0.006475776195526123, 0.006510591983795166, 0.006457344055175781, 0.006500448226928711, 0.007337215900421142, 0.0064784960746765135, 0.006529151916503906, 0.0064503679275512696, 0.006394688129425049, 0.006407296180725098, 0.006417151927947998, 0.006456895828247071, 0.00646608018875122, 0.006448416233062744, 0.00647811222076416, 0.006443359851837158, 0.0064839677810668945, 0.006463488101959228, 0.006455296039581298, 0.006519904136657715, 0.006471583843231201, 0.006489088058471679, 0.006536543846130371, 0.006500895977020264, 0.006551680088043213, 0.006472928047180176, 0.006460192203521728, 0.006487775802612305, 0.006459392070770263, 0.006465824127197266, 0.006456448078155518, 0.006458047866821289, 0.006453440189361572, 0.00645743989944458, 0.006436960220336914, 0.006452223777770996, 0.006430560111999511, 0.0065270719528198245, 0.006626175880432129, 0.006442463874816894, 0.006500192165374756, 0.0064899840354919434, 0.006553728103637696, 0.006513343811035156, 0.0064430079460144046, 0.0061296639442443845, 0.006388031959533692, 0.006413568019866944, 0.006488192081451416, 0.006426943778991699, 0.006395008087158203, 0.006395071983337402, 0.006448512077331543, 0.0064997758865356445, 0.006460288047790527, 0.006528223991394043, 0.00657692813873291, 0.006512576103210449, 0.00653715181350708, 0.006510719776153564, 0.006481887817382813, 0.006469664096832276, 0.0064297599792480465, 0.0064419198036193845, 0.0065922880172729495, 0.0064019842147827145, 0.0064041919708251955, 0.006348991870880127, 0.0064093761444091795, 0.006423232078552246, 0.006441120147705078, 0.006506400108337402, 0.006432928085327148, 0.006509984016418457, 0.006504896163940429, 0.006482336044311523, 0.006481632232666016, 0.006458528041839599, 0.006497087955474853, 0.006465248107910156, 0.00645904016494751, 0.006499040126800537, 0.00648796796798706, 0.006413856029510498, 0.006461919784545899, 0.006422304153442383, 0.006504479885101318, 0.0065170879364013674, 0.006464384078979492, 0.006515456199645996, 0.00647382402420044, 0.006445184230804444, 0.0064711041450500485, 0.006425055980682373, 0.006484288215637207, 0.006457119941711426, 0.006468768119812012, 0.006488959789276123, 0.006467455863952636, 0.006487936019897461, 0.006465536117553711, 0.006465760231018067, 0.006516736030578613, 0.006580575942993164, 0.006573728084564209, 0.0066068158149719235, 0.006618144035339356, 0.006601247787475586, 0.00634281587600708, 0.006684671878814697, 0.006606912136077881, 0.006594079971313477, 0.006547872066497803, 0.006559455871582031, 0.006539552211761474, 0.006536960124969482, 0.006541056156158447, 0.006566207885742187, 0.00654969596862793, 0.0065147199630737306, 0.00647983980178833, 0.00653107213973999, 0.006520832061767578, 0.006526976108551025, 0.0065177597999572755, 0.006508639812469482, 0.0064910402297973634, 0.0064471039772033695, 0.006446815967559814, 0.006479648113250733, 0.006461023807525635, 0.0064910402297973634, 0.0065075201988220215, 0.006441023826599121, 0.006491199970245361, 0.006464928150177002, 0.006457824230194092, 0.006465760231018067, 0.0064878082275390625, 0.0066943359375, 0.006535999774932862, 0.006735551834106446, 0.006506591796875, 0.006518879890441895, 0.006473631858825684, 0.006479872226715088, 0.006460544109344483, 0.00644595193862915, 0.006444672107696533, 0.00640883207321167, 0.006432511806488037, 0.006426239967346191, 0.006395679950714111, 0.006432447910308838, 0.006412288188934326, 0.006909952163696289, 0.006778048038482666, 0.006442527770996093, 0.006437056064605713, 0.006441088199615479, 0.0064039678573608394, 0.0064011201858520505, 0.006403007984161377, 0.006444863796234131, 0.006414624214172363, 0.006379392147064209, 0.006371200084686279, 0.00636736011505127, 0.006405663967132568, 0.00638105583190918, 0.0063787841796875, 0.006033120155334473, 0.0063712639808654785, 0.006404448032379151, 0.006399328231811523, 0.0063515200614929195, 0.0063504638671875, 0.006379903793334961, 0.0063582401275634765, 0.006367680072784424, 0.006381279945373535, 0.006354623794555664, 0.006394815921783447, 0.006367455959320068, 0.0063649601936340335, 0.006354944229125976, 0.006420703887939453, 0.006397727966308594, 0.0064204797744750975, 0.006408095836639404, 0.006522304058074952, 0.006431072235107422, 0.006407519817352295, 0.00641267204284668, 0.0063695359230041505, 0.006625951766967774, 0.006377151966094971, 0.0064430079460144046, 0.006364255905151367, 0.006386591911315918, 0.006416160106658936, 0.0063463358879089354, 0.00639244794845581, 0.006379519939422608, 0.006397535800933838, 0.006359583854675293, 0.006356895923614502, 0.006379583835601807, 0.006369184017181397, 0.0063508481979370115, 0.006352896213531494, 0.006346752166748047, 0.006363135814666748, 0.006365439891815186, 0.006329984188079834, 0.006416512012481689, 0.00632422399520874, 0.006318079948425293, 0.006240255832672119, 0.006270976066589356, 0.006287583827972412, 0.006268703937530517, 0.006281216144561768, 0.006291711807250976, 0.006305535793304443, 0.006275072097778321, 0.006311935901641846, 0.006301663875579834, 0.006268735885620117, 0.006309311866760254, 0.006286176204681396, 0.006320064067840576, 0.006245791912078857, 0.006310495853424072, 0.005974016189575195, 0.006301023960113525, 0.0062984957695007325, 0.006343647956848144, 0.006298431873321533, 0.006313151836395263, 0.00629641580581665, 0.006295135974884033, 0.00635532808303833, 0.006268928050994873, 0.006336512088775635, 0.006318079948425293, 0.0063281598091125485, 0.007385280132293701, 0.006432096004486084, 0.0065030078887939454, 0.00859875202178955, 0.007103424072265625, 0.0066128640174865725, 0.006550559997558594, 0.006575232028961182, 0.0065288000106811525, 0.006492224216461182, 0.0065413122177124024, 0.006522047996520996, 0.006488895893096923, 0.006543360233306885, 0.006502336025238037, 0.006504511833190918, 0.006496255874633789, 0.006449408054351806, 0.006450240135192871, 0.00644323205947876, 0.006459871768951416, 0.006448192119598389, 0.006433728218078613, 0.006494495868682861, 0.006483359813690186, 0.006414463996887207, 0.0064791679382324215, 0.0064551677703857425, 0.006398975849151611, 0.006408415794372559, 0.006373151779174805, 0.006422751903533935, 0.00637500810623169, 0.006377600193023682, 0.006422016143798828, 0.006359551906585693, 0.006404287815093994, 0.006365056037902832, 0.006345823764801025, 0.006382495880126953, 0.006372384071350098, 0.006408991813659668, 0.006332831859588623, 0.006388927936553955, 0.006343264102935791, 0.006336512088775635, 0.006368576049804687, 0.006378176212310791, 0.006321695804595948, 0.006332191944122314, 0.006023200035095215, 0.006301663875579834, 0.0063143038749694826, 0.006301375865936279, 0.006299295902252197, 0.006291808128356934, 0.006313920021057129, 0.006322400093078613, 0.006256480216979981, 0.0063610877990722655, 0.007538496017456054, 0.006288896083831787, 0.006355296134948731, 0.006356639862060547, 0.006307839870452881, 0.006399903774261475, 0.006302720069885254, 0.006323967933654785, 0.006299071788787842, 0.006304448127746582, 0.006307744026184082, 0.0063201279640197755, 0.006313248157501221, 0.006248767852783203, 0.006324831962585449, 0.006326079845428467, 0.006291584014892578, 0.006307551860809326, 0.006330527782440185, 0.006329376220703125, 0.0062969279289245604, 0.006397568225860596, 0.006391808032989502, 0.006369279861450195, 0.006369088172912598, 0.006328512191772461, 0.006387584209442139, 0.006354047775268555, 0.006305791854858398, 0.0063170561790466305, 0.006305280208587646, 0.006351327896118164, 0.006332448005676269, 0.0063285760879516605, 0.006341856002807617, 0.006296160221099853, 0.006299583911895752, 0.0062993597984313966, 0.00632041597366333, 0.006300960063934326, 0.006284095764160156, 0.006547359943389892, 0.006260735988616943, 0.0062791681289672855, 0.006242368221282959, 0.006320064067840576, 0.006275072097778321, 0.006301695823669433, 0.006295551776885986, 0.006258272171020508, 0.006308256149291992, 0.006338560104370118, 0.006289408206939697, 0.00602569580078125, 0.006383264064788818, 0.006302048206329345, 0.006322175979614258, 0.006309120178222656, 0.006292319774627685, 0.006264351844787598, 0.00637721586227417, 0.0063496317863464356, 0.006341824054718018, 0.0062798080444335935, 0.006359039783477783, 0.006270976066589356, 0.006291776180267334, 0.006252223968505859, 0.006322175979614258, 0.006256703853607178, 0.006233183860778808, 0.00626361608505249, 0.006246560096740723, 0.006250463962554932, 0.006209440231323242, 0.006268928050994873, 0.006205440044403076, 0.006309792041778565, 0.00623583984375, 0.006209248065948486, 0.0061914238929748535, 0.006216063976287842, 0.0061922879219055175, 0.006204480171203613, 0.0062154560089111325, 0.006187071800231933, 0.0063446397781372075, 0.006205408096313476, 0.0062280001640319825, 0.0062341117858886715, 0.006227968215942382, 0.006219007968902588, 0.0061979198455810545, 0.006243743896484375, 0.006242720127105713, 0.006249023914337158, 0.006237247943878174, 0.006224544048309326, 0.006422368049621582, 0.006281375885009766, 0.006285120010375977, 0.00619539213180542, 0.006339744091033935, 0.006269536018371582, 0.006205664157867431, 0.006245471954345703, 0.006235072135925293, 0.006264832019805908, 0.006250624179840088, 0.006330239772796631, 0.0062849922180175785, 0.00625871992111206, 0.006947103977203369, 0.006329823970794677, 0.006337215900421142, 0.006279007911682129, 0.0059985918998718265, 0.0062707200050354005, 0.006320735931396484, 0.0062709121704101565, 0.006238080024719238, 0.006258495807647705, 0.006262080192565918, 0.006256671905517578, 0.006230368137359619, 0.006277184009552002, 0.006270271778106689, 0.006275072097778321, 0.006337056159973145, 0.006304192066192627, 0.006294847965240479, 0.006281919956207275, 0.006291456222534179, 0.006312096118927002, 0.006330207824707031, 0.006389791965484619, 0.006354015827178955, 0.0063434882164001465, 0.0065567679405212406, 0.006362080097198486, 0.006709248065948486, 0.006362688064575195, 0.006371327877044678, 0.006754271984100342, 0.006320352077484131, 0.006333759784698486, 0.0063006081581115725, 0.006453536033630371, 0.006436063766479492, 0.00640390396118164, 0.006370240211486816, 0.006321919918060303, 0.006346752166748047, 0.006254591941833496, 0.0062197761535644535, 0.007120384216308594, 0.0062984638214111325, 0.0062986879348754885, 0.006265567779541016, 0.006235136032104492, 0.006281631946563721, 0.006228447914123535, 0.0062845759391784665, 0.006284287929534912, 0.006303264141082764, 0.006301887989044189, 0.006301695823669433, 0.006313695907592774, 0.006294015884399414, 0.006297376155853272, 0.0062863359451293946, 0.0063211522102355954, 0.006362207889556885, 0.006443935871124268, 0.006464799880981445, 0.006457471847534179, 0.006417344093322754, 0.006401919841766358, 0.006396031856536865, 0.006118400096893311, 0.006386847972869873, 0.006331200122833252, 0.006283360004425049, 0.0063056640625, 0.006331840038299561, 0.006343552112579346, 0.006315775871276855, 0.00632422399520874, 0.006326272010803223, 0.006305247783660889, 0.006291679859161377, 0.006299967765808105, 0.006326272010803223, 0.006313983917236328, 0.006498303890228272, 0.006844480037689209, 0.006610623836517334, 0.006901247978210449, 0.00668342399597168, 0.0068198080062866215, 0.006694911956787109, 0.012613632202148438, 0.008443903923034669, 0.006449151992797852, 0.006561791896820069, 0.006459392070770263, 0.0064495038986206055, 0.006485568046569824, 0.006559648036956787, 0.006512832164764404, 0.006506080150604248, 0.006547071933746338, 0.00655785608291626, 0.006623936176300049, 0.006490047931671143, 0.006477407932281494, 0.006474112033843994, 0.006506527900695801, 0.006449151992797852, 0.006428927898406983, 0.006482687950134278, 0.006481120109558106, 0.006489952087402343, 0.006676415920257569, 0.006551328182220459, 0.006580448150634766, 0.006612736225128174, 0.006614304065704345, 0.0067266240119934085, 0.006524384021759033, 0.006562655925750733, 0.006530752182006836, 0.006530208110809326, 0.006502848148345947, 0.0065886077880859375, 0.0066152639389038086, 0.006531424045562744, 0.006591775894165039, 0.006694464206695556, 0.006584159851074219, 0.006548448085784912, 0.006526624202728272, 0.006287263870239258, 0.0065717120170593266, 0.006576735973358155, 0.0065569281578063965, 0.006519360065460205, 0.006576128005981445, 0.00656828784942627, 0.006518432140350342, 0.006523935794830322, 0.006458335876464843, 0.006498367786407471, 0.006470623970031738, 0.006461599826812744, 0.006510464191436768, 0.006472703933715821, 0.0064347200393676755, 0.0064553279876708985, 0.006440671920776367, 0.006504320144653321, 0.006482336044311523, 0.006434815883636475, 0.006471680164337158, 0.006516736030578613, 0.006501984119415283, 0.006635424137115479, 0.006795775890350342, 0.006483232021331787, 0.006500576019287109, 0.006500480175018311, 0.006547008037567139, 0.006550335884094238, 0.006496384143829346, 0.00648528003692627, 0.006460288047790527, 0.006435872077941895, 0.006529727935791016, 0.0064646081924438475, 0.006449600219726562, 0.006425055980682373, 0.006391808032989502, 0.0064204797744750975, 0.00636518383026123, 0.006405312061309815, 0.006407167911529541, 0.006328127861022949, 0.006362368106842041, 0.0063450241088867185, 0.006398591995239258, 0.006352799892425537, 0.006364480018615722, 0.006381696224212647, 0.006383103847503662, 0.006479231834411621, 0.006377056121826172, 0.006335616111755371, 0.006343679904937744, 0.006367104053497314, 0.006408192157745361, 0.006404255867004395, 0.0063814082145690915, 0.0063836159706115725, 0.006409664154052735, 0.006486815929412842, 0.006401760101318359, 0.006476319789886474, 0.00643071985244751, 0.006457344055175781, 0.00639081621170044, 0.006364128112792969, 0.006375423908233643, 0.0063949117660522465, 0.006341856002807617, 0.006352287769317627, 0.006324575901031494, 0.006371520042419434, 0.006350560188293457, 0.006338272094726562, 0.006360511779785156, 0.00635590410232544, 0.006334464073181153, 0.006338751792907714, 0.0063731842041015625, 0.006368896007537842, 0.006362783908843994, 0.006326943874359131, 0.0063324799537658695, 0.006395808219909668, 0.006373472213745117, 0.006313375949859619, 0.006377823829650879, 0.006305535793304443, 0.006324704170227051, 0.0062930240631103515, 0.006303775787353516, 0.006324096202850342, 0.006323935985565185, 0.006369631767272949, 0.007159391880035401, 0.0064460477828979495, 0.006471839904785157, 0.0064061121940612795, 0.006362912178039551, 0.006775904178619385, 0.0068858561515808105, 0.006566559791564942, 0.006413568019866944, 0.006502175807952881, 0.006503551959991455, 0.00648364782333374, 0.006452320098876953, 0.006463808059692383, 0.0063985600471496585, 0.006436704158782959, 0.006441120147705078, 0.00637337589263916, 0.006391808032989502, 0.006306943893432618, 0.0063640961647033695, 0.006449088096618653, 0.006348864078521728, 0.006356927871704101, 0.006368512153625488, 0.006512479782104492, 0.006404255867004395, 0.0063424639701843265, 0.006460319995880127, 0.00601475191116333, 0.006338111877441406, 0.00629807996749878, 0.0063116798400878905, 0.006285823822021484, 0.0064061121940612795, 0.006402048110961914, 0.006383935928344727, 0.0063909759521484375, 0.006344992160797119, 0.006525152206420899, 0.006363135814666748, 0.006338016033172607, 0.006351391792297364, 0.006356575965881348, 0.006408127784729004, 0.006316512107849121, 0.006342751979827881, 0.006288608074188232, 0.0063034558296203615, 0.006326591968536377, 0.006314655780792236, 0.00636518383026123, 0.006340608119964599, 0.006295296192169189, 0.006283391952514648, 0.0063489279747009275, 0.006275072097778321, 0.006319392204284668, 0.006258975982666015, 0.00629094409942627, 0.006255519866943359, 0.006293536186218262, 0.00626204776763916, 0.006314015865325927, 0.006310463905334473, 0.006323775768280029, 0.00630947208404541, 0.006353312015533448, 0.006488639831542969, 0.006272863864898681, 0.007192736148834229, 0.0068297600746154786, 0.006313759803771973, 0.0063433279991149906, 0.006295328140258789, 0.006260096073150635, 0.0062873921394348145, 0.006314432144165039, 0.0063647360801696775, 0.006257343769073487, 0.006305791854858398, 0.0063134078979492185, 0.006263519763946533, 0.006299488067626953, 0.0062873601913452145, 0.006292992115020752, 0.006273536205291748, 0.006297887802124024, 0.006350304126739502, 0.006265408039093018, 0.006290112018585205, 0.006255616188049316, 0.005999839782714843, 0.006224768161773682, 0.006280576229095459, 0.006279647827148438, 0.00630185604095459, 0.006262784004211426, 0.006307839870452881, 0.006299071788787842, 0.0062674560546875, 0.006453279972076416, 0.006313951969146728, 0.006615039825439453, 0.006309887886047363, 0.00632422399520874, 0.006344704151153564, 0.0063491201400756835, 0.0064652800559997555, 0.006313695907592774, 0.006367455959320068, 0.006408095836639404, 0.006348896026611328, 0.006459392070770263, 0.006391808032989502, 0.006375584125518799, 0.006360928058624268, 0.006385248184204101, 0.0063976640701293944, 0.007331583976745605, 0.006437920093536377, 0.0064629120826721195, 0.006449632167816162, 0.007112703800201416, 0.00771011209487915, 0.007934559822082519, 0.007710720062255859, 0.007201056003570557, 0.00653004789352417, 0.006484576225280762, 0.0064943361282348636, 0.006465824127197266, 0.006491104125976563, 0.006438687801361084, 0.006390751838684082, 0.006378975868225098, 0.006361343860626221, 0.006418496131896973, 0.006363359928131104, 0.006371327877044678, 0.006370719909667969, 0.00639353609085083, 0.006430880069732666, 0.0064122557640075685, 0.006392767906188965, 0.006404255867004395, 0.00637062406539917, 0.006404543876647949, 0.006389696121215821, 0.006399040222167969, 0.006388671875, 0.006371327877044678, 0.006412288188934326, 0.00639737606048584, 0.006384128093719482, 0.006056831836700439, 0.006418560028076172, 0.006722335815429688, 0.006388671875, 0.006404096126556396, 0.006418432235717773, 0.00636627197265625, 0.006362048149108887, 0.006344128131866455, 0.006389696121215821, 0.006373824119567871, 0.006353280067443848, 0.006440767765045166, 0.006364384174346924, 0.006384416103363037, 0.006369279861450195, 0.0063836159706115725, 0.0063991999626159665, 0.0064429759979248045, 0.006386335849761963, 0.006366879940032959, 0.006584640026092529, 0.006442944049835205, 0.0064076800346374516, 0.006448959827423096, 0.00647379207611084, 0.006433472156524658, 0.006434400081634522, 0.006417088031768799, 0.006453152179718018, 0.00643891191482544, 0.006492127895355224, 0.006591872215270996, 0.006481791973114014, 0.006474527835845947, 0.006478943824768066, 0.006419360160827637, 0.006555295944213867, 0.006470016002655029, 0.006420447826385498, 0.00645308780670166, 0.00643017578125, 0.006464000225067139, 0.0065248641967773436, 0.0063994240760803225, 0.006416575908660889, 0.0063760638236999515, 0.006395904064178467, 0.006320288181304931, 0.006348639965057373, 0.006376704216003418, 0.006370048046112061, 0.006416384220123291, 0.006392864227294922, 0.0063740482330322265, 0.006378880023956299, 0.006378015995025635, 0.006408607959747315, 0.006408192157745361, 0.0064124159812927245, 0.006387584209442139, 0.006369279861450195, 0.00643071985244751, 0.006132160186767578, 0.006469727993011475, 0.006444960117340088, 0.006387263774871826, 0.0063710718154907225, 0.006351039886474609, 0.006420447826385498, 0.006377151966094971, 0.0063517122268676755, 0.006393119812011718, 0.006384160041809082, 0.00641871976852417, 0.0063712320327758785, 0.0064204797744750975, 0.006399104118347168, 0.006382304191589356, 0.006391424179077149, 0.006347296237945557, 0.0064074559211730955, 0.006367680072784424, 0.006359327793121338, 0.006404096126556396, 0.006382880210876465, 0.006431136131286621, 0.0063919677734375, 0.006375584125518799, 0.006369279861450195, 0.006356800079345703, 0.006387712001800537, 0.006352128028869629, 0.006370528221130371, 0.006418144226074219, 0.006402048110961914, 0.006496352195739746, 0.006501952171325684, 0.006472032070159912, 0.006478047847747802, 0.0064694080352783205, 0.006448351860046387, 0.006511392116546631, 0.006448959827423096, 0.006452928066253662, 0.006455423831939697, 0.006470016002655029, 0.006510591983795166, 0.006535168170928955, 0.006516736030578613, 0.0064856958389282226, 0.006454751968383789, 0.006509407997131348, 0.006545407772064209, 0.00651910400390625, 0.006487743854522705, 0.00742300796508789, 0.006487296104431152, 0.006532608032226562, 0.006502431869506836, 0.006462656021118164, 0.006433792114257812, 0.006406144142150879, 0.00642252779006958, 0.006440256118774414, 0.006441664218902588]",tokens/s,155.67273137433045,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2177.503232,2492.399616,0.0,2097.152,1986.693632,s,1,9.5335927734375,9.5335927734375,0.0,9.5335927734375,9.5335927734375,9.5335927734375,9.5335927734375,[9.5335927734375],,kWh,7.331160806666807e-05,8.078223592068064e-06,2.9448634669998475e-05,0.0001108384663287346,,MB,2245.156864,2504.982528,0.0,2097.152,1862.377472,s,10,11.278011718750001,1.127801171875,0.0012554719244792937,1.1274984130859376,1.1294764282226564,1.129490106201172,1.1295010485839845,"[1.125908203125, 1.1264990234375, 1.12655517578125, 1.1295037841796876, 1.1271705322265626, 1.127464111328125, 1.12753271484375, 1.1289794921875, 1.12892529296875, 1.129473388671875]",tokens/s,226.99036530915555,kWh,3.2999977701250025e-05,3.6394173935105946e-06,2.1911711973799514e-05,5.855110706856014e-05,tokens/kWh,4372248.669871229,MB,2249.78944,2504.982528,0.0,2097.152,1946.952704,s,10,14.862877807617187,1.4862877807617187,0.02083933928562863,1.4797899169921873,1.5144125610351562,1.5164294860839844,1.5180430261230469,"[1.5184464111328124, 1.510836669921875, 1.495533447265625, 1.51396435546875, 1.487755615234375, 1.4692376708984376, 1.4657362060546875, 1.463328369140625, 1.46621484375, 1.47182421875]",tokens/s,42.387484318624125,kWh,4.332284441624911e-05,4.778796280974686e-06,2.577252061800039e-05,7.387416131522419e-05,tokens/kWh,852801.5598197632,,s,630,14.860007503509511,0.02358731349763416,0.0006191026026769244,0.023459424018859865,0.024075221633911133,0.024343612957000733,0.0255697407913208,"[0.024386816024780274, 0.024013568878173828, 0.0239803524017334, 0.024073759078979493, 0.024008159637451173, 0.02391315269470215, 0.02386124801635742, 0.02432614326477051, 0.02386288070678711, 0.023849376678466795, 0.024115007400512697, 0.023937215805053712, 0.024070144653320313, 0.024086528778076172, 0.02414729690551758, 0.024102912902832032, 0.02407491111755371, 0.024292543411254884, 0.024138463973999023, 0.0240166072845459, 0.02407865524291992, 0.024024320602416993, 0.024973760604858397, 0.024062335968017576, 0.024197120666503907, 0.024004608154296874, 0.02392678451538086, 0.023817888259887697, 0.023935327529907225, 0.024020992279052734, 0.023883615493774414, 0.02383683204650879, 0.02411315155029297, 0.024023040771484375, 0.02410028839111328, 0.024136255264282227, 0.02492416000366211, 0.024465408325195313, 0.02424790382385254, 0.024061439514160156, 0.024017824172973632, 0.023860416412353515, 0.02385145568847656, 0.023857215881347656, 0.02380406379699707, 0.023987775802612306, 0.024031839370727538, 0.023982080459594726, 0.024025087356567384, 0.02391859245300293, 0.02387459182739258, 0.023978975296020506, 0.023957664489746094, 0.024026975631713868, 0.024766464233398438, 0.026572799682617186, 0.024110143661499023, 0.02420172882080078, 0.023980480194091797, 0.02391859245300293, 0.02387923240661621, 0.02374496078491211, 0.02371993637084961, 0.024375360488891603, 0.02394153594970703, 0.023875648498535157, 0.02389580726623535, 0.023852672576904297, 0.023990272521972656, 0.024038015365600587, 0.023974912643432617, 0.024792064666748048, 0.02406393623352051, 0.024342592239379884, 0.024006656646728516, 0.02408038330078125, 0.023963647842407225, 0.024030368804931642, 0.023941984176635744, 0.02399635124206543, 0.02386950492858887, 0.024022176742553712, 0.023996448516845702, 0.02434444808959961, 0.023880128860473634, 0.02391267204284668, 0.024037664413452148, 0.02392268753051758, 0.023668159484863283, 0.02371027183532715, 0.023908351898193358, 0.023957408905029298, 0.023990367889404295, 0.023867391586303712, 0.02388899230957031, 0.024031936645507814, 0.02388422393798828, 0.023926559448242186, 0.023848512649536132, 0.024569440841674804, 0.02477552032470703, 0.024094720840454102, 0.023969791412353517, 0.024047903060913086, 0.023999584197998046, 0.02407084846496582, 0.02388547134399414, 0.024225215911865234, 0.02389017677307129, 0.02401273536682129, 0.023963935852050783, 0.023916927337646485, 0.023814144134521483, 0.024029184341430664, 0.023816192626953125, 0.023783071517944336, 0.023708000183105468, 0.023758848190307616, 0.0237076473236084, 0.02367897605895996, 0.023666688919067383, 0.023856416702270507, 0.023720672607421875, 0.02373222351074219, 0.02390425682067871, 0.024061759948730468, 0.02438969612121582, 0.023949216842651368, 0.024035327911376952, 0.024030431747436524, 0.023966495513916015, 0.02384486389160156, 0.023782527923583985, 0.02398092842102051, 0.023744512557983398, 0.023666496276855468, 0.023811872482299806, 0.024060319900512696, 0.02378700828552246, 0.023671295166015623, 0.02370742416381836, 0.023660768508911134, 0.02382156753540039, 0.023814432144165038, 0.023796064376831055, 0.023661855697631837, 0.023642976760864257, 0.023596864700317383, 0.023941183090209962, 0.02367500877380371, 0.023856544494628908, 0.023771743774414062, 0.0236810245513916, 0.023607295989990236, 0.0237293758392334, 0.023599519729614257, 0.02366092872619629, 0.023629823684692384, 0.02368076705932617, 0.023664064407348633, 0.023628671646118163, 0.02368880081176758, 0.02364246368408203, 0.023723007202148438, 0.023722879409790038, 0.023619712829589842, 0.023681280136108398, 0.023626720428466797, 0.02354240036010742, 0.02353705596923828, 0.023688159942626952, 0.023754528045654297, 0.023760000228881837, 0.02369171142578125, 0.023696832656860352, 0.02378031921386719, 0.02378329658508301, 0.023732383728027343, 0.023733631134033203, 0.02361334419250488, 0.023589920043945313, 0.02351465606689453, 0.02356150436401367, 0.02356038475036621, 0.023472831726074218, 0.02344960021972656, 0.02369932746887207, 0.023823936462402343, 0.023941120147705077, 0.025169919967651368, 0.02459379196166992, 0.024125343322753907, 0.023968320846557617, 0.023946720123291014, 0.023962015151977538, 0.023998783111572265, 0.023928800582885743, 0.023956735610961913, 0.024692800521850584, 0.023818944931030272, 0.02385055923461914, 0.023796159744262694, 0.023613439559936524, 0.0246560001373291, 0.024514015197753907, 0.02400534439086914, 0.02407801628112793, 0.02408448028564453, 0.023971168518066407, 0.023962272644042968, 0.02384841537475586, 0.02475267219543457, 0.024758272171020508, 0.02397011184692383, 0.023795391082763673, 0.026979743957519533, 0.03143945693969727, 0.023740415573120118, 0.02368819236755371, 0.023839744567871093, 0.02371583938598633, 0.023690784454345703, 0.02429305648803711, 0.02388047981262207, 0.023750656127929686, 0.02395955276489258, 0.023928768157958986, 0.0237825927734375, 0.024092735290527342, 0.024687423706054687, 0.0238734073638916, 0.023984256744384765, 0.024137727737426756, 0.023855392456054687, 0.02356809616088867, 0.023615488052368162, 0.023373823165893554, 0.023318527221679687, 0.023322303771972655, 0.023298368453979493, 0.02333695983886719, 0.023172224044799804, 0.023417055130004884, 0.023368192672729493, 0.023290016174316405, 0.02334457588195801, 0.02380009651184082, 0.023344415664672852, 0.02325196838378906, 0.023166976928710937, 0.023195648193359376, 0.023359487533569336, 0.0242774715423584, 0.023863391876220705, 0.023791263580322266, 0.023628480911254884, 0.023474496841430666, 0.023314111709594725, 0.023330816268920897, 0.023234560012817384, 0.023185407638549805, 0.02317673683166504, 0.023113664627075196, 0.023101984024047853, 0.023145824432373046, 0.023359264373779297, 0.027855104446411132, 0.027777536392211914, 0.02375289535522461, 0.023691200256347657, 0.023174911499023437, 0.023146751403808594, 0.023879327774047852, 0.025571680068969725, 0.023247968673706054, 0.02322105598449707, 0.023238752365112306, 0.023212032318115236, 0.023123327255249022, 0.02346044731140137, 0.023118175506591798, 0.02315817642211914, 0.023249183654785156, 0.02327510452270508, 0.024131999969482423, 0.024788991928100586, 0.023293760299682616, 0.02319379234313965, 0.02328384017944336, 0.023313312530517577, 0.02333590316772461, 0.023245824813842773, 0.023177696228027345, 0.023513631820678713, 0.0237260799407959, 0.024396799087524415, 0.02384588813781738, 0.023562240600585937, 0.02352332878112793, 0.023646207809448243, 0.023604671478271486, 0.02352707290649414, 0.02355846405029297, 0.02362224006652832, 0.023557279586791994, 0.023345632553100584, 0.02315507125854492, 0.023197599411010742, 0.02343049621582031, 0.023259904861450194, 0.023302143096923827, 0.02328780746459961, 0.02323276710510254, 0.023137216567993165, 0.023147167205810545, 0.023985151290893555, 0.023528192520141603, 0.023388416290283202, 0.023128063201904296, 0.023179264068603517, 0.02308710479736328, 0.023127424240112306, 0.023128704071044923, 0.02313216018676758, 0.023142400741577147, 0.023224319458007812, 0.023223424911499025, 0.023486751556396485, 0.02345840072631836, 0.023524608612060547, 0.023593536376953127, 0.02413382339477539, 0.023625312805175783, 0.023628192901611327, 0.023560192108154295, 0.023521280288696288, 0.023480319976806642, 0.023382015228271484, 0.023516544342041014, 0.023226591110229493, 0.023316896438598633, 0.023334911346435547, 0.023345151901245118, 0.02327347183227539, 0.02314975929260254, 0.023461023330688478, 0.023268640518188475, 0.023200128555297853, 0.023112768173217772, 0.02314950370788574, 0.023222272872924804, 0.023214080810546874, 0.023179264068603517, 0.023142400741577147, 0.023158784866333007, 0.02427827262878418, 0.02398080062866211, 0.023617536544799804, 0.02330419158935547, 0.023096384048461913, 0.02327142333984375, 0.023114688873291017, 0.02305843162536621, 0.02319923210144043, 0.023089664459228516, 0.023220224380493162, 0.023111679077148437, 0.023077983856201172, 0.023204416275024415, 0.023125728607177733, 0.023138816833496095, 0.023179391860961913, 0.02309939193725586, 0.023175008773803712, 0.02309744071960449, 0.023185216903686523, 0.023262752532958984, 0.023157472610473632, 0.023965696334838867, 0.023488512039184572, 0.02319094467163086, 0.023077375411987306, 0.0230994873046875, 0.02324684715270996, 0.023150592803955077, 0.02314854431152344, 0.023353343963623048, 0.02323865509033203, 0.023136255264282226, 0.02324265670776367, 0.02314249610900879, 0.023130111694335938, 0.023128063201904296, 0.02329542350769043, 0.023172895431518556, 0.02310633659362793, 0.023084575653076173, 0.02310537528991699, 0.023061183929443358, 0.023052223205566408, 0.023162879943847657, 0.0230830078125, 0.023068672180175782, 0.023044095993041993, 0.023053472518920898, 0.023025983810424804, 0.02312656021118164, 0.02332057571411133, 0.02309939193725586, 0.02310553550720215, 0.023189184188842773, 0.02309766387939453, 0.02309280014038086, 0.02313465690612793, 0.023138208389282225, 0.023267423629760742, 0.025564992904663086, 0.02481990432739258, 0.02325017547607422, 0.023149023056030272, 0.02320591926574707, 0.023228256225585938, 0.023263647079467775, 0.023302143096923827, 0.023287328720092773, 0.023376352310180665, 0.02337286376953125, 0.02308105659484863, 0.023104352951049803, 0.02324684715270996, 0.023289440155029296, 0.023464351654052733, 0.02326835250854492, 0.02319580841064453, 0.023163232803344726, 0.02316339111328125, 0.023181343078613283, 0.02310550308227539, 0.02348236846923828, 0.023136255264282226, 0.023127967834472657, 0.024162464141845703, 0.02358665657043457, 0.023254655838012697, 0.02311363220214844, 0.0232391357421875, 0.02329190444946289, 0.023212032318115236, 0.023212032318115236, 0.023214080810546874, 0.023343103408813477, 0.02557542419433594, 0.02324265670776367, 0.023199840545654295, 0.023162591934204103, 0.023107872009277344, 0.023146495819091797, 0.02317430305480957, 0.023124000549316407, 0.023171007156372072, 0.023128799438476563, 0.023197856903076172, 0.02318880081176758, 0.023190208435058594, 0.023172447204589844, 0.023130783081054686, 0.02325503921508789, 0.02330624008178711, 0.02318671989440918, 0.02308892822265625, 0.02314124870300293, 0.0231278076171875, 0.02317251205444336, 0.023133087158203124, 0.023072416305541993, 0.023080543518066408, 0.02314726448059082, 0.023187456130981447, 0.023144447326660156, 0.02311302375793457, 0.023196352005004882, 0.02309724807739258, 0.023134559631347657, 0.02309836769104004, 0.02313430404663086, 0.023141023635864257, 0.023173120498657225, 0.02312406349182129, 0.02306412887573242, 0.023045503616333007, 0.02312041664123535, 0.023173568725585937, 0.02334534454345703, 0.023305471420288087, 0.023229312896728516, 0.02309087944030762, 0.023154687881469727, 0.023134208679199218, 0.02310553550720215, 0.023064863204956054, 0.023100255966186523, 0.02308799934387207, 0.02311689567565918, 0.023135135650634766, 0.024225791931152343, 0.023690975189208985, 0.02348089599609375, 0.023215808868408204, 0.02322230339050293, 0.023115488052368165, 0.023237152099609373, 0.02319657516479492, 0.023161727905273436, 0.023123935699462892, 0.02325503921508789, 0.023183135986328124, 0.023722175598144532, 0.02315881538391113, 0.023285184860229492, 0.02383929634094238, 0.025065376281738282, 0.02342508888244629, 0.023316511154174803, 0.02333286476135254, 0.023262432098388672, 0.023378719329833986, 0.023201791763305665, 0.02311974334716797, 0.023117952346801758, 0.023736320495605468, 0.02326323127746582, 0.02310758399963379, 0.023195648193359376, 0.023152639389038086, 0.023123296737670898, 0.02314672088623047, 0.023192319869995117, 0.02304787254333496, 0.023145952224731446, 0.023088703155517577, 0.02312495994567871, 0.023171072006225587, 0.023117824554443358, 0.023119871139526366, 0.023431167602539063, 0.02309529685974121, 0.023646207809448243, 0.02314396858215332, 0.02309984016418457, 0.0231844482421875, 0.02302012825012207, 0.023128032684326172, 0.02307523155212402, 0.023133535385131836, 0.023165599822998047, 0.023349248886108398, 0.023197311401367188, 0.02315507125854492, 0.023177183151245118, 0.023189279556274416, 0.023131935119628907, 0.023038431167602538, 0.023117536544799804, 0.02311196708679199, 0.02305356788635254, 0.023065343856811523, 0.02317686462402344, 0.024754175186157225, 0.023817279815673827, 0.023759807586669922, 0.023633119583129882, 0.02368489646911621, 0.023472768783569336, 0.02336742401123047, 0.023264991760253907, 0.0233276481628418, 0.023369119644165038, 0.023368288040161132, 0.023369728088378908, 0.02326937675476074, 0.02321023941040039, 0.023132959365844728, 0.023110687255859377, 0.02309084892272949, 0.023107872009277344, 0.023203039169311525, 0.023137056350708007, 0.02314431953430176, 0.023237855911254882, 0.02345471954345703, 0.023465536117553712, 0.023658815383911132, 0.023395679473876954, 0.023290559768676757, 0.02304979133605957, 0.023091648101806642, 0.024012928009033204, 0.023089023590087892, 0.02302726364135742, 0.02311008071899414, 0.02308710479736328, 0.02310291290283203, 0.023084672927856445, 0.02304684829711914, 0.023058687210083008, 0.023125120162963867, 0.02310028839111328, 0.023045536041259765, 0.023265888214111328, 0.023146495819091797, 0.023150047302246093, 0.023167327880859376, 0.023138496398925783, 0.023170944213867187, 0.023156768798828126, 0.023141727447509766, 0.023173887252807616, 0.023166976928710937, 0.02333919906616211, 0.023471071243286134, 0.02369993591308594, 0.023678335189819336, 0.02387455940246582, 0.023688224792480467, 0.023624000549316407, 0.023938816070556642, 0.023579103469848633, 0.02362335968017578, 0.023612159729003906, 0.023584768295288085]",tokens/s,42.39567172837641,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4356.030464,5355.012096,0.0,4959.76448,4769.731072,s,1,11.3029736328125,11.3029736328125,0.0,11.3029736328125,11.3029736328125,11.3029736328125,11.3029736328125,[11.3029736328125],,kWh,0.00013010201382500003,1.4343644621801782e-05,5.7639212778000035e-05,0.00020208487122480185,,MB,1615.03232,5373.886464,0.0,4966.055936,4251.027456,s,10,30.630434814453125,3.0630434814453125,0.004086782872654609,3.0632532958984378,3.067527099609375,3.06789404296875,3.06818759765625,"[3.054669921875, 3.05810400390625, 3.06162646484375, 3.0614765625, 3.064130615234375, 3.0623759765625, 3.066512939453125, 3.065831787109375, 3.068260986328125, 3.067445556640625]",tokens/s,83.577004881173,kWh,8.94033612470832e-05,9.861140676874984e-06,5.9460658679599995e-05,0.00015872516060355817,tokens/kWh,1612850.7857642148,MB,1622.60992,5388.566528,0.0,4980.736,4251.030016,s,10,17.748735351562498,1.77487353515625,0.00824765049756315,1.772857360839844,1.7831903564453124,1.7875395141601562,1.7910188403320313,"[1.7669478759765624, 1.7713026123046876, 1.7784786376953126, 1.774412109375, 1.791888671875, 1.78211083984375, 1.765817626953125, 1.7652471923828126, 1.770305908203125, 1.782223876953125]",tokens/s,35.49548672179274,kWh,5.2776072571249924e-05,5.820655008400578e-06,3.4842416762799994e-05,9.343914434245052e-05,tokens/kWh,674235.6262286357,,s,630,17.744511219024666,0.028165890823848665,0.0004181261219157756,0.028107151985168458,0.028600368690490725,0.02875118350982666,0.02936630439758301,"[0.028757663726806642, 0.028042911529541015, 0.027849695205688477, 0.027927391052246092, 0.02764227294921875, 0.027685312271118163, 0.028643360137939455, 0.02774630355834961, 0.027622623443603514, 0.02810655975341797, 0.027959775924682618, 0.027746559143066406, 0.027622911453247072, 0.02776476860046387, 0.027816352844238282, 0.027843072891235353, 0.027650943756103517, 0.027914623260498046, 0.027714111328125, 0.027873279571533204, 0.0277193603515625, 0.027958784103393555, 0.027814048767089844, 0.028125888824462892, 0.028208768844604493, 0.028195104598999023, 0.028170303344726564, 0.028414976119995116, 0.028423168182373046, 0.02835456085205078, 0.028413951873779295, 0.02857574462890625, 0.028298271179199218, 0.028285919189453126, 0.028167423248291017, 0.028465791702270506, 0.02810483169555664, 0.02815385627746582, 0.028141855239868164, 0.028302719116210937, 0.02793712043762207, 0.02799731254577637, 0.02800124740600586, 0.028077503204345704, 0.02807651138305664, 0.02811494445800781, 0.028052831649780275, 0.027823104858398437, 0.02793747138977051, 0.028513023376464844, 0.0282193603515625, 0.027963584899902343, 0.027841663360595702, 0.02784320068359375, 0.02777529525756836, 0.027811840057373048, 0.02789116859436035, 0.027910495758056642, 0.02814112091064453, 0.028721792221069336, 0.028042720794677733, 0.027892576217651368, 0.027846303939819336, 0.028726207733154298, 0.028196863174438477, 0.028391424179077147, 0.028104703903198244, 0.028073984146118162, 0.02852979278564453, 0.028013439178466798, 0.028069887161254883, 0.02802387237548828, 0.028310367584228516, 0.028136640548706054, 0.028136671066284178, 0.02809823989868164, 0.02812326431274414, 0.02778883171081543, 0.02781737518310547, 0.02778822326660156, 0.028088319778442384, 0.028161535263061522, 0.02805401611328125, 0.027753536224365234, 0.0277142391204834, 0.02769536018371582, 0.02783875274658203, 0.027708511352539062, 0.028186304092407227, 0.027849279403686523, 0.028342655181884766, 0.028146944046020507, 0.028299711227416993, 0.0287172794342041, 0.028082271575927735, 0.028743263244628905, 0.028154272079467774, 0.02817228889465332, 0.028303071975708007, 0.02814828872680664, 0.02831279945373535, 0.028422624588012695, 0.02833523178100586, 0.028218271255493164, 0.028192768096923827, 0.02815180778503418, 0.02788150405883789, 0.02781916809082031, 0.028167232513427735, 0.027930368423461915, 0.02808137512207031, 0.02785321617126465, 0.02791472053527832, 0.027873216629028322, 0.027656160354614257, 0.02751251220703125, 0.027787839889526367, 0.027743423461914062, 0.02931155204772949, 0.029345855712890626, 0.028250207901000978, 0.02843164825439453, 0.0280216007232666, 0.027966367721557618, 0.02769987106323242, 0.02765126419067383, 0.029288448333740235, 0.028362464904785157, 0.0313863353729248, 0.02802409553527832, 0.02834636878967285, 0.02799001693725586, 0.028013280868530274, 0.02817843246459961, 0.028278175354003905, 0.028366464614868164, 0.028150335311889648, 0.027939231872558593, 0.02805868721008301, 0.028332096099853515, 0.02781683158874512, 0.02791219139099121, 0.027914239883422853, 0.028039167404174805, 0.02793471908569336, 0.027860191345214842, 0.027816736221313476, 0.02796303939819336, 0.027801952362060546, 0.027751487731933595, 0.02780361557006836, 0.028097375869750977, 0.028039583206176756, 0.028716096878051756, 0.028463647842407225, 0.02832758331298828, 0.028600799560546876, 0.028571552276611328, 0.028223520278930665, 0.0284467830657959, 0.028264448165893553, 0.028387327194213868, 0.02837708854675293, 0.028380447387695313, 0.02841270446777344, 0.02828486442565918, 0.028094207763671875, 0.02793087959289551, 0.027973728179931642, 0.028026784896850586, 0.028339487075805664, 0.02810870361328125, 0.027978559494018555, 0.027910144805908203, 0.028030048370361327, 0.027921375274658204, 0.02791961669921875, 0.027914623260498046, 0.027859264373779297, 0.02787059211730957, 0.028156383514404297, 0.02846531105041504, 0.02838319969177246, 0.028353567123413085, 0.028313600540161132, 0.028427167892456053, 0.02856559944152832, 0.028372543334960938, 0.028305856704711915, 0.028676383972167967, 0.028227584838867188, 0.02850201606750488, 0.027879199981689452, 0.02774038314819336, 0.02782147216796875, 0.02774403190612793, 0.02765702438354492, 0.027828224182128908, 0.027922431945800782, 0.027889503479003906, 0.027860319137573242, 0.02803158378601074, 0.028133567810058595, 0.028208223342895508, 0.02843289566040039, 0.028154592514038086, 0.028501728057861327, 0.0280710391998291, 0.02815679931640625, 0.028028095245361328, 0.028213184356689455, 0.02804185676574707, 0.02809395217895508, 0.02786790466308594, 0.02794495964050293, 0.028236896514892577, 0.028304288864135742, 0.028317855834960937, 0.028255199432373045, 0.02816924858093262, 0.029072351455688476, 0.028105600357055664, 0.02813132858276367, 0.02809040069580078, 0.028274656295776367, 0.028254463195800782, 0.028608383178710936, 0.028386560440063477, 0.028298944473266602, 0.02807865524291992, 0.028367231369018555, 0.028180479049682617, 0.029063167572021483, 0.02828816032409668, 0.029012832641601562, 0.02813260841369629, 0.028101375579833984, 0.028112255096435546, 0.028094463348388672, 0.028246335983276367, 0.027887392044067382, 0.027886112213134764, 0.027950239181518555, 0.028141984939575194, 0.027832767486572266, 0.028014591217041016, 0.027910144805908203, 0.0279583683013916, 0.028037887573242187, 0.028121248245239257, 0.02792857551574707, 0.028073984146118162, 0.029064287185668947, 0.02865577507019043, 0.028299840927124023, 0.028461055755615236, 0.030914527893066406, 0.028531904220581054, 0.028445280075073243, 0.028430591583251952, 0.028526880264282226, 0.02817398452758789, 0.027817695617675782, 0.02807423973083496, 0.028272735595703126, 0.027799264907836914, 0.027758975982666016, 0.028020639419555664, 0.028249759674072266, 0.028346719741821288, 0.028065183639526366, 0.028190624237060546, 0.028265151977539062, 0.028001440048217772, 0.028139936447143556, 0.028329824447631834, 0.028209760665893556, 0.028247648239135743, 0.02813705635070801, 0.028067935943603517, 0.027987968444824218, 0.028182912826538085, 0.028186975479125978, 0.028128959655761718, 0.02886649513244629, 0.02822563171386719, 0.028250015258789063, 0.0284532470703125, 0.028692480087280273, 0.028602367401123048, 0.028464351654052734, 0.028405920028686523, 0.028813087463378906, 0.028509023666381837, 0.02873103904724121, 0.028575519561767578, 0.028722944259643553, 0.028709760665893556, 0.028553152084350587, 0.0285347843170166, 0.028481536865234375, 0.028577375411987304, 0.028885311126708984, 0.028895328521728516, 0.028544063568115233, 0.028484832763671874, 0.028502847671508787, 0.02854560089111328, 0.02876860809326172, 0.02853398323059082, 0.028613407135009764, 0.028440576553344726, 0.028336160659790038, 0.02836182403564453, 0.028473791122436524, 0.02879324722290039, 0.02814156723022461, 0.028204864501953125, 0.028127424240112303, 0.028114208221435545, 0.02853766441345215, 0.028365856170654298, 0.02840665626525879, 0.02816579246520996, 0.02821334457397461, 0.028152128219604493, 0.028130752563476562, 0.028332544326782227, 0.028118623733520507, 0.027972000122070313, 0.027875232696533202, 0.027940639495849608, 0.028174047470092774, 0.028826208114624025, 0.031318016052246093, 0.028112895965576173, 0.027936767578125, 0.028016639709472657, 0.028028928756713867, 0.027990304946899414, 0.028087263107299806, 0.028183296203613283, 0.028114015579223633, 0.028405855178833008, 0.02830204772949219, 0.028319839477539063, 0.028151391983032226, 0.028377504348754884, 0.029198112487792968, 0.028508384704589843, 0.028462368011474608, 0.02841449546813965, 0.028287168502807616, 0.028331935882568358, 0.028305503845214845, 0.02813273620605469, 0.027995840072631836, 0.02807244873046875, 0.028151744842529296, 0.028099071502685546, 0.028069887161254883, 0.02807923126220703, 0.02801286315917969, 0.0286231689453125, 0.02816640090942383, 0.02815184020996094, 0.028000032424926758, 0.02823520088195801, 0.02791865539550781, 0.029374656677246095, 0.0283155517578125, 0.028194976806640626, 0.02832841682434082, 0.028094175338745118, 0.02811075210571289, 0.028024160385131835, 0.028334848403930663, 0.027883520126342775, 0.02926585578918457, 0.02825484848022461, 0.028057151794433594, 0.027826656341552736, 0.028018655776977538, 0.028101951599121093, 0.028070592880249025, 0.027969375610351562, 0.027812000274658202, 0.028016639709472657, 0.027934591293334962, 0.027875680923461914, 0.028093984603881836, 0.0280762882232666, 0.02897305679321289, 0.029832672119140625, 0.0281810245513916, 0.028286144256591796, 0.02885055923461914, 0.028228031158447266, 0.027926528930664062, 0.027834367752075196, 0.02863315200805664, 0.027845951080322267, 0.0276691837310791, 0.027721376419067384, 0.027735519409179687, 0.027771135330200196, 0.027841087341308593, 0.027989696502685547, 0.027826496124267578, 0.028338176727294922, 0.027887136459350585, 0.027668960571289064, 0.027830272674560546, 0.028214879989624023, 0.028036863327026366, 0.027885919570922853, 0.02784492874145508, 0.027867136001586915, 0.02792038345336914, 0.027875328063964845, 0.027807743072509765, 0.027826175689697266, 0.02767616081237793, 0.027668096542358397, 0.027961727142333984, 0.027910879135131836, 0.02770512008666992, 0.027809375762939452, 0.02773151969909668, 0.02769596862792969, 0.027760255813598634, 0.027865631103515625, 0.027905887603759765, 0.027703296661376952, 0.02775052833557129, 0.027867008209228515, 0.028317407608032228, 0.027953088760375975, 0.02790608024597168, 0.02812550354003906, 0.028117088317871092, 0.028852672576904298, 0.0280043830871582, 0.028004352569580077, 0.02768076705932617, 0.02793471908569336, 0.027611135482788086, 0.027815935134887695, 0.0277359676361084, 0.028149280548095703, 0.028064064025878906, 0.028416255950927734, 0.02816409683227539, 0.028180479049682617, 0.028217344284057616, 0.028237823486328126, 0.028442272186279295, 0.02823740768432617, 0.028248128890991212, 0.028203296661376952, 0.02790991973876953, 0.02800499153137207, 0.028050912857055663, 0.028142112731933594, 0.028026847839355468, 0.028057632446289064, 0.028067840576171874, 0.028073728561401366, 0.027947200775146484, 0.028118688583374022, 0.027830495834350585, 0.027949247360229492, 0.027863040924072265, 0.028932096481323243, 0.028241920471191406, 0.02832793617248535, 0.028221439361572266, 0.028218591690063476, 0.02816076850891113, 0.02848771286010742, 0.028004352569580077, 0.02787308883666992, 0.027860671997070312, 0.027815711975097655, 0.02842428779602051, 0.0277653751373291, 0.027877376556396483, 0.027926271438598632, 0.027859199523925782, 0.027611135482788086, 0.027893760681152343, 0.027660287857055665, 0.027838016510009767, 0.02764182472229004, 0.0278306884765625, 0.027693119049072266, 0.027875328063964845, 0.027729248046875, 0.02792310333251953, 0.027766399383544922, 0.028139904022216798, 0.027756032943725587, 0.027621055603027345, 0.02774304008483887, 0.028835872650146484, 0.028057024002075194, 0.02799398422241211, 0.027677343368530272, 0.027719743728637697, 0.027715776443481447, 0.028087295532226563, 0.027650623321533202, 0.027824127197265625, 0.028188383102416992, 0.02852707290649414, 0.029421567916870117, 0.02836479949951172, 0.028030431747436524, 0.027824159622192382, 0.027747072219848633, 0.027676416397094727, 0.02784992027282715, 0.027744800567626952, 0.027725439071655273, 0.027794080734252928, 0.027715583801269532, 0.02765145683288574, 0.027870943069458008, 0.027808383941650392, 0.02773651123046875, 0.02860630416870117, 0.029122560501098634, 0.02796441650390625, 0.027988576889038087, 0.027918752670288087, 0.02851430320739746, 0.02777292823791504, 0.027897855758666993, 0.028305120468139648, 0.027928287506103516, 0.03185523223876953, 0.02821308708190918, 0.02801263999938965, 0.027807136535644532, 0.027716192245483398, 0.02773580741882324, 0.02787347221374512, 0.02773561668395996, 0.02795516777038574, 0.02776323127746582, 0.02769715118408203, 0.02789580726623535, 0.02806505584716797, 0.028271327972412108, 0.028216575622558592, 0.028681184768676756, 0.02822047996520996, 0.02813395118713379, 0.028170528411865233, 0.028321664810180665, 0.028278783798217775, 0.028368896484375, 0.02808822441101074, 0.028031007766723633, 0.027903167724609376, 0.027851648330688476, 0.027852800369262694, 0.028877216339111327, 0.02817433547973633, 0.028166208267211914, 0.02808143997192383, 0.02880793571472168, 0.028185535430908203, 0.028107744216918945, 0.028130336761474608, 0.028396480560302733, 0.028576000213623047, 0.02833180809020996, 0.028438528060913085, 0.02849523162841797, 0.028596895217895508, 0.02866934394836426, 0.02865023994445801, 0.02864896011352539, 0.028604736328125, 0.02860032081604004, 0.02854617691040039, 0.028443199157714844, 0.028633407592773438, 0.028632959365844726, 0.02873766326904297, 0.028661760330200195, 0.028704767227172853, 0.028383232116699218, 0.028434656143188478, 0.028509279251098633, 0.028308160781860353, 0.02806118392944336, 0.02861926460266113, 0.02765180778503418, 0.027953184127807618, 0.027916543960571288, 0.028332063674926758, 0.02828489685058594, 0.02836275291442871, 0.02831155204772949, 0.02872319984436035, 0.028209184646606444, 0.028374752044677733, 0.028272607803344726, 0.02829280090332031, 0.028121023178100585, 0.02834499168395996, 0.028333887100219727, 0.028038560867309572, 0.028037919998168945, 0.028014144897460937, 0.028085952758789064, 0.02801308822631836, 0.028201183319091796, 0.027850751876831056, 0.027787200927734373, 0.02783647918701172, 0.027778528213500978, 0.02791084861755371, 0.027856735229492186, 0.027883232116699217, 0.027822271347045898, 0.027973728179931642, 0.028092704772949218]",tokens/s,35.503936525709975,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1516.470272,1531.904,0.0,1136.656384,1111.384576,s,1,8.6451328125,8.6451328125,0.0,8.6451328125,8.6451328125,8.6451328125,8.6451328125,[8.6451328125],,kWh,4.834146508333864e-05,5.307537180176209e-06,1.72866804960016e-05,7.093568275951646e-05,,MB,1583.77984,1601.110016,0.0,1193.279488,1029.126656,s,10,5.371971313476563,0.5371971313476562,0.0013318217229636975,0.53732275390625,0.5388457336425782,0.5390548858642578,0.5392222076416016,"[0.5392640380859375, 0.5355089721679688, 0.5380020751953125, 0.536339111328125, 0.537271240234375, 0.5346407470703125, 0.5373074951171875, 0.5387992553710937, 0.5373380126953125, 0.5375003662109376]",tokens/s,476.5475931671073,kWh,1.5963014583771957e-05,1.7604251500054838e-06,1.0564204357789467e-05,2.8287644091566905e-05,tokens/kWh,9049887.617764484,MB,1591.15264,1603.207168,0.0,1195.37664,1083.4944,s,10,10.96520068359375,1.096520068359375,0.007217433858219241,1.0955146484374998,1.1040343994140625,1.1077592041015625,1.1107390478515624,"[1.0915089111328125, 1.095031982421875, 1.084250244140625, 1.089641845703125, 1.1032066650390624, 1.0956351318359374, 1.1114840087890625, 1.10139501953125, 1.0976527099609374, 1.0953941650390624]",tokens/s,57.45448881228528,kWh,3.385353185497731e-05,3.7339509309515524e-06,1.7515929217410514e-05,5.510341200333938e-05,tokens/kWh,1143304.882757207,,s,630,10.962743856430054,0.01740118072449215,0.000302951764663302,0.017318960189819335,0.017811832809448244,0.01798636655807495,0.01838773988723755,"[0.018071903228759765, 0.017671455383300783, 0.018241312026977537, 0.01757689666748047, 0.017505407333374023, 0.01761087989807129, 0.01747020721435547, 0.01762486457824707, 0.017348672866821287, 0.017549375534057617, 0.017196640014648438, 0.017168479919433592, 0.017242528915405272, 0.0171144962310791, 0.017166015625, 0.017195648193359375, 0.017215232849121093, 0.017097280502319335, 0.017111040115356444, 0.017111040115356444, 0.017113088607788086, 0.01704140853881836, 0.017137664794921875, 0.017032768249511717, 0.017078720092773437, 0.01704140853881836, 0.017061567306518553, 0.01700486373901367, 0.017098367691040037, 0.01711552047729492, 0.017121280670166016, 0.01708790397644043, 0.017341024398803712, 0.0172728328704834, 0.017345792770385744, 0.017300224304199217, 0.017283071517944337, 0.017254400253295898, 0.017540895462036132, 0.01738947105407715, 0.017535263061523438, 0.01775209617614746, 0.01752239990234375, 0.017352800369262695, 0.01743276786804199, 0.01740540885925293, 0.01747420883178711, 0.01742425537109375, 0.017526784896850587, 0.01743257522583008, 0.017456287384033202, 0.017256736755371094, 0.01719353675842285, 0.01718230438232422, 0.017170848846435546, 0.017463424682617187, 0.01725632095336914, 0.017216991424560547, 0.01733603286743164, 0.01722800064086914, 0.017287776947021483, 0.017223039627075196, 0.017173120498657227, 0.01806537628173828, 0.017682432174682617, 0.017311744689941407, 0.017268735885620116, 0.017188800811767577, 0.017184831619262694, 0.017229824066162108, 0.01715590476989746, 0.017205408096313476, 0.017282751083374022, 0.01734223937988281, 0.017289567947387695, 0.017222944259643554, 0.017140256881713868, 0.017209760665893553, 0.017108991622924806, 0.01716419219970703, 0.017154144287109374, 0.017145856857299805, 0.017084415435791016, 0.017225439071655274, 0.017127712249755858, 0.017137664794921875, 0.01752992057800293, 0.01814409637451172, 0.018397279739379883, 0.017494016647338868, 0.017416128158569334, 0.0175350399017334, 0.017328128814697266, 0.017313440322875975, 0.017422399520874023, 0.017344736099243165, 0.017336383819580078, 0.017397184371948243, 0.017340383529663085, 0.01752943992614746, 0.017333984375, 0.01751273536682129, 0.017473535537719728, 0.017442815780639647, 0.01745715141296387, 0.01742233657836914, 0.017349632263183593, 0.01734668731689453, 0.0174703369140625, 0.01723948860168457, 0.01721196746826172, 0.017338367462158204, 0.01764892768859863, 0.018029279708862304, 0.01740924835205078, 0.017453855514526367, 0.017481727600097655, 0.01726249694824219, 0.01723196792602539, 0.017315839767456053, 0.0172476806640625, 0.017279552459716796, 0.017367040634155274, 0.01719705581665039, 0.01721139144897461, 0.017612800598144532, 0.018023584365844728, 0.017535839080810547, 0.01742624092102051, 0.017234111785888673, 0.017168479919433592, 0.017219488143920898, 0.01724006462097168, 0.017196672439575195, 0.01747609519958496, 0.01711280059814453, 0.01715203285217285, 0.017141632080078125, 0.017112319946289062, 0.017085439682006837, 0.017113088607788086, 0.01720729637145996, 0.01722287940979004, 0.01717737579345703, 0.017200767517089845, 0.017031551361083985, 0.017203168869018556, 0.017102367401123048, 0.01715177536010742, 0.017150688171386718, 0.0172806396484375, 0.017283456802368164, 0.01739548873901367, 0.01726076889038086, 0.01719705581665039, 0.017194080352783202, 0.01718771171569824, 0.017100831985473634, 0.017124671936035157, 0.01710736083984375, 0.017185056686401367, 0.017176544189453125, 0.017127712249755858, 0.017153535842895508, 0.01714201545715332, 0.017059648513793945, 0.017204864501953125, 0.01722764778137207, 0.017811487197875977, 0.017113759994506837, 0.01715974426269531, 0.01704185676574707, 0.017118623733520508, 0.017220191955566407, 0.017146879196166993, 0.017081344604492187, 0.017160192489624023, 0.017152000427246093, 0.017246208190917968, 0.017243295669555663, 0.01728803253173828, 0.01709174346923828, 0.017159008026123048, 0.017141759872436522, 0.01722777557373047, 0.017176576614379883, 0.017125375747680666, 0.01707827186584473, 0.01713968086242676, 0.018052223205566407, 0.01759228706359863, 0.01737615966796875, 0.017348031997680664, 0.01732268714904785, 0.01728499221801758, 0.01715545654296875, 0.01712745666503906, 0.01708435249328613, 0.017124000549316405, 0.017188863754272463, 0.017119232177734374, 0.017115135192871094, 0.017115135192871094, 0.01720729637145996, 0.017127424240112304, 0.017270368576049806, 0.017136032104492188, 0.01761484718322754, 0.017102495193481445, 0.017850719451904296, 0.017221120834350585, 0.017433088302612306, 0.01742438316345215, 0.01725984001159668, 0.017287103652954102, 0.01736163139343262, 0.0174366397857666, 0.01745033645629883, 0.01754390335083008, 0.017434623718261717, 0.017391616821289063, 0.017480928421020506, 0.017554208755493163, 0.01746112060546875, 0.01743827247619629, 0.017272640228271484, 0.01727769660949707, 0.01732419204711914, 0.017424224853515625, 0.01721343994140625, 0.017516544342041016, 0.017121280670166016, 0.017049280166625977, 0.01732793617248535, 0.017252864837646483, 0.01728499221801758, 0.0173384952545166, 0.01733987236022949, 0.017168928146362304, 0.017136959075927733, 0.017213600158691406, 0.017098623275756834, 0.01705027198791504, 0.0170949764251709, 0.017094335556030273, 0.017106943130493164, 0.017293312072753905, 0.017291263580322267, 0.0172227840423584, 0.017136512756347658, 0.017118976593017577, 0.017126815795898438, 0.018485248565673826, 0.01760665512084961, 0.01720319938659668, 0.01723756790161133, 0.01717673683166504, 0.017258783340454102, 0.017141727447509764, 0.01711836814880371, 0.017154016494750977, 0.01709542465209961, 0.01834614372253418, 0.017579296112060546, 0.017231712341308592, 0.01708121681213379, 0.017123327255249024, 0.017321184158325197, 0.017526784896850587, 0.01748227119445801, 0.017410655975341797, 0.01719606399536133, 0.017246847152709962, 0.01769267272949219, 0.01728335952758789, 0.017227487564086916, 0.01722572708129883, 0.017085504531860352, 0.017158432006835936, 0.017150623321533203, 0.01715193557739258, 0.017129535675048827, 0.017152000427246093, 0.01708201599121094, 0.01717692756652832, 0.017321983337402345, 0.017643007278442382, 0.017719999313354492, 0.017778495788574218, 0.017709056854248048, 0.01785379219055176, 0.017732255935668944, 0.01781760025024414, 0.017804479598999022, 0.017805408477783204, 0.01772755241394043, 0.01779574394226074, 0.01780531120300293, 0.017745439529418944, 0.017732063293457032, 0.017733631134033204, 0.017747167587280274, 0.017785152435302733, 0.017957088470458984, 0.017867263793945314, 0.01783782386779785, 0.017901216506958008, 0.017737888336181642, 0.017637567520141603, 0.017559551239013673, 0.01758415985107422, 0.01750217628479004, 0.01754854393005371, 0.01750502395629883, 0.017503583908081054, 0.01791801643371582, 0.017827327728271485, 0.01740451240539551, 0.017173759460449217, 0.017134239196777343, 0.017139711380004884, 0.01707827186584473, 0.017354751586914064, 0.017328479766845702, 0.017307296752929687, 0.017337568283081056, 0.01726748847961426, 0.01735443115234375, 0.0174466552734375, 0.017684255599975586, 0.01736390495300293, 0.01753487968444824, 0.017388959884643555, 0.01741423988342285, 0.017457599639892577, 0.01742438316345215, 0.017421344757080078, 0.0174748477935791, 0.017464096069335938, 0.017484703063964845, 0.017473440170288086, 0.01726473617553711, 0.017164287567138673, 0.01721548843383789, 0.017129472732543945, 0.017096864700317384, 0.018331071853637696, 0.017461664199829103, 0.017239839553833007, 0.01722390365600586, 0.017100799560546876, 0.017147775650024413, 0.017107072830200194, 0.017121280670166016, 0.017055744171142577, 0.01715782356262207, 0.017140031814575196, 0.017177824020385743, 0.017113887786865234, 0.01725619125366211, 0.017162303924560546, 0.017219072341918946, 0.017855167388916016, 0.0188272647857666, 0.01728102493286133, 0.017315839767456053, 0.017310911178588868, 0.017401952743530274, 0.01735875129699707, 0.017427263259887697, 0.017350656509399414, 0.01741823959350586, 0.017520639419555666, 0.017360895156860352, 0.017854463577270507, 0.017620384216308595, 0.017469919204711915, 0.017447040557861327, 0.01835385513305664, 0.017809791564941405, 0.017452032089233398, 0.017320032119750976, 0.01722051239013672, 0.017180192947387696, 0.017320415496826173, 0.017111040115356444, 0.017141759872436522, 0.017266687393188478, 0.017202880859375, 0.01722956848144531, 0.017360639572143555, 0.01715283203125, 0.01718681526184082, 0.017292287826538084, 0.017317888259887695, 0.01764614486694336, 0.017899967193603514, 0.017588224411010742, 0.017573888778686524, 0.017321119308471678, 0.017285984039306642, 0.017485567092895508, 0.017311296463012694, 0.01727964782714844, 0.017205535888671877, 0.01712892723083496, 0.017213312149047852, 0.017210912704467774, 0.017212127685546873, 0.017492128372192384, 0.017703168869018553, 0.017780479431152345, 0.017881088256835938, 0.01794867134094238, 0.018126848220825196, 0.01790140724182129, 0.01797318458557129, 0.017914079666137697, 0.017975296020507812, 0.018013248443603514, 0.0180982723236084, 0.01791472053527832, 0.018052959442138673, 0.01795907211303711, 0.017946624755859376, 0.017995424270629883, 0.017968832015991212, 0.01804876708984375, 0.018364383697509767, 0.018135583877563477, 0.018501535415649414, 0.018157567977905274, 0.017862752914428712, 0.017656543731689452, 0.017690303802490235, 0.01801398468017578, 0.01779475212097168, 0.017834367752075194, 0.017626527786254884, 0.01734137535095215, 0.017297216415405273, 0.018059072494506837, 0.01771254348754883, 0.01744361686706543, 0.017299455642700197, 0.017164287567138673, 0.01714995193481445, 0.017156095504760743, 0.017211519241333007, 0.017474559783935546, 0.01742323112487793, 0.01744691276550293, 0.017354272842407228, 0.017275360107421874, 0.017133567810058595, 0.017396928787231446, 0.01717740821838379, 0.01726233673095703, 0.017133279800415038, 0.017121824264526367, 0.017182048797607423, 0.017261215209960937, 0.01722777557373047, 0.017296512603759764, 0.017326976776123045, 0.017253568649291992, 0.01723846435546875, 0.017242303848266603, 0.017282560348510743, 0.017650367736816407, 0.01762713623046875, 0.017618207931518554, 0.01752751922607422, 0.017505632400512696, 0.017426816940307618, 0.018227487564086913, 0.018348031997680665, 0.018647039413452148, 0.017680383682250975, 0.018019744873046875, 0.01774620819091797, 0.017766719818115236, 0.01755340766906738, 0.017438528060913085, 0.01785843276977539, 0.017565599441528322, 0.01748543930053711, 0.01745590400695801, 0.017495264053344728, 0.017227680206298827, 0.0171539192199707, 0.017162975311279298, 0.017139999389648438, 0.01715171241760254, 0.017313888549804687, 0.01728531265258789, 0.01742367935180664, 0.017254816055297852, 0.01729676818847656, 0.017177440643310546, 0.01710905647277832, 0.017880191802978517, 0.01988083267211914, 0.017346879959106446, 0.01783184051513672, 0.017525184631347657, 0.017344512939453126, 0.01738515281677246, 0.017242431640625, 0.018280288696289063, 0.01773923110961914, 0.01762348747253418, 0.017535232543945314, 0.01759436798095703, 0.017837312698364256, 0.017338720321655274, 0.017437088012695313, 0.017469440460205078, 0.01718272018432617, 0.017327680587768554, 0.017777088165283204, 0.01745715141296387, 0.017827840805053712, 0.017487360000610352, 0.017259008407592775, 0.017072128295898437, 0.017136640548706054, 0.017177183151245116, 0.017160608291625978, 0.017164287567138673, 0.017118816375732423, 0.01711555290222168, 0.017301055908203126, 0.01716473579406738, 0.017256799697875976, 0.017256128311157228, 0.017405920028686524, 0.017127424240112304, 0.017217504501342774, 0.017242143630981446, 0.017357919692993166, 0.017632160186767578, 0.01756979179382324, 0.01738956832885742, 0.017454751968383787, 0.01741606330871582, 0.017473760604858397, 0.01743017578125, 0.017472095489501953, 0.017467391967773437, 0.01741619110107422, 0.01740595245361328, 0.017516544342041016, 0.017373184204101562, 0.017434623718261717, 0.01743257522583008, 0.017462848663330078, 0.017842559814453124, 0.017413440704345702, 0.01725926399230957, 0.017292448043823242, 0.017383871078491212, 0.01725257682800293, 0.017368511199951173, 0.01747430419921875, 0.017438720703125, 0.017546911239624024, 0.018438207626342774, 0.017780672073364256, 0.017487871170043946, 0.01743052864074707, 0.017295360565185547, 0.017293312072753905, 0.01725644874572754, 0.017231647491455077, 0.017193183898925782, 0.017575935363769533, 0.017473535537719728, 0.017373184204101562, 0.01721139144897461, 0.01719705581665039, 0.01722083282470703, 0.017181472778320314, 0.017340063095092773, 0.017256736755371094, 0.017203264236450196, 0.017094655990600584, 0.017310783386230467, 0.01718092727661133, 0.017407743453979493, 0.017252416610717775, 0.017240959167480467, 0.017309696197509765, 0.01753651237487793, 0.017418752670288085, 0.017814943313598633, 0.017590879440307617, 0.017547264099121093, 0.01724844741821289, 0.017250112533569336, 0.01718272018432617, 0.017252351760864256, 0.01720524787902832, 0.017315839767456053, 0.017207359313964842, 0.017289152145385744, 0.017252256393432617, 0.01744086456298828, 0.017215103149414063, 0.017277311325073243, 0.017108991622924806, 0.01718284797668457, 0.017245792388916017, 0.017448480606079102, 0.01757651138305664, 0.017518335342407227, 0.017438976287841797, 0.017479871749877928, 0.017518592834472657, 0.01747990417480469, 0.01756070327758789, 0.01747420883178711, 0.01745894432067871, 0.01742464065551758, 0.01766579246520996, 0.017550592422485353, 0.017503231048583985, 0.01744076728820801, 0.01750956726074219, 0.017289344787597655]",tokens/s,57.467364762926735,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3142.098944,4602.134528,0.0,4206.886912,4087.771648,s,1,11.3848994140625,11.3848994140625,0.0,11.3848994140625,11.3848994140625,11.3848994140625,11.3848994140625,[11.3848994140625],,kWh,0.00011912927681670831,1.3133468904521388e-05,4.522864729397669e-05,0.0001774913930152064,,MB,3186.556928,4774.100992,0.0,4366.270464,4273.048576,s,10,14.616045898437498,1.46160458984375,0.0024391264378625013,1.4610399780273438,1.4643836791992189,1.465648260498047,1.4666599255371096,"[1.460572998046875, 1.458865234375, 1.4580650634765624, 1.4618348388671876, 1.4611273193359375, 1.4605428466796875, 1.46095263671875, 1.4630694580078125, 1.4641026611328125, 1.466912841796875]",tokens/s,175.14996995690004,kWh,4.293747330542221e-05,4.73315819915845e-06,2.8513161699400546e-05,7.618379320398122e-05,tokens/kWh,3360294.7455577985,MB,3186.556928,4778.295296,0.0,4368.367616,4273.051136,s,10,24.59880004882813,2.4598800048828124,0.0055870374012830905,2.457652099609375,2.4651350830078127,2.4693376831054685,2.4726997631835936,"[2.464201171875, 2.457966064453125, 2.4545078125, 2.457338134765625, 2.457046142578125, 2.4617607421875, 2.457312744140625, 2.453147705078125, 2.473540283203125, 2.461979248046875]",tokens/s,25.611005364060954,kWh,7.094289176040852e-05,7.827931004083778e-06,4.0493560172599746e-05,0.000119264382937092,tokens/kWh,528238.1751241727,,s,630,24.59587634277343,0.03904107355995784,0.0005100896243114708,0.03896319961547851,0.03939310531616211,0.03969704113006592,0.04119656272888184,"[0.03987046432495117, 0.03919257736206055, 0.03948339080810547, 0.03919801712036133, 0.03940217590332031, 0.03913523101806641, 0.03915161514282227, 0.03880303955078125, 0.038876575469970705, 0.03863241577148437, 0.0387410888671875, 0.038720447540283205, 0.039024639129638675, 0.03879446411132813, 0.038781726837158206, 0.038767807006835936, 0.03908041763305664, 0.03907769775390625, 0.03886748886108399, 0.03883785629272461, 0.03877519989013672, 0.03880672073364258, 0.03867116928100586, 0.03895238494873047, 0.03868691253662109, 0.03888572692871094, 0.03888131332397461, 0.038986881256103514, 0.03907468795776367, 0.03908505630493164, 0.039088512420654295, 0.04006358337402344, 0.04022480010986328, 0.03917193603515625, 0.03943859100341797, 0.03906460952758789, 0.04046118545532226, 0.038984737396240234, 0.0392303352355957, 0.03887113571166992, 0.038932479858398435, 0.038913761138916016, 0.03897577667236328, 0.0390041618347168, 0.03904716873168945, 0.03895500946044922, 0.03904307174682617, 0.03989299011230469, 0.03912499237060547, 0.03930112075805664, 0.03906118392944336, 0.03907206344604492, 0.03911679840087891, 0.03912499237060547, 0.03938304138183594, 0.0392305908203125, 0.03923174285888672, 0.039139583587646486, 0.039077472686767575, 0.03930534362792969, 0.03909043121337891, 0.038989471435546874, 0.03904358291625976, 0.039609886169433596, 0.03925040054321289, 0.03909008026123047, 0.039034976959228515, 0.039293216705322265, 0.03900592041015625, 0.038934528350830076, 0.0388218879699707, 0.03885465621948242, 0.03890790557861328, 0.038718528747558593, 0.038960254669189454, 0.038774593353271485, 0.038629375457763675, 0.03863654327392578, 0.03944502258300781, 0.03942351913452148, 0.038847232818603514, 0.038931774139404296, 0.038740863800048826, 0.03905542373657227, 0.03874105453491211, 0.03896790313720703, 0.03858598327636719, 0.03893929672241211, 0.03890585708618164, 0.039049121856689455, 0.039008350372314454, 0.03904886245727539, 0.038916446685791015, 0.039174144744873046, 0.03881369781494141, 0.0390530891418457, 0.039526622772216795, 0.038950912475585936, 0.03874630355834961, 0.038597663879394534, 0.039027488708496094, 0.03863929748535156, 0.03877654266357422, 0.038525600433349606, 0.03868384170532226, 0.03884489440917969, 0.03889385604858398, 0.03906086349487305, 0.0390431022644043, 0.039137535095214844, 0.03931785583496094, 0.038778881072998046, 0.03885158538818359, 0.03881062316894531, 0.03898076629638672, 0.03868716812133789, 0.03871376037597656, 0.0424601936340332, 0.03903078460693359, 0.03934102249145508, 0.03912908935546875, 0.039094432830810544, 0.039065471649169924, 0.03894185638427734, 0.03886368179321289, 0.038983295440673825, 0.03948505783081055, 0.03890796661376953, 0.038615360260009765, 0.03853302383422851, 0.038306049346923825, 0.038618335723876955, 0.03881760025024414, 0.03958777618408203, 0.03929792022705078, 0.03897958374023437, 0.03885260772705078, 0.03918195343017578, 0.03890214538574219, 0.038784801483154295, 0.03896112060546875, 0.0387869758605957, 0.039022014617919924, 0.03898886489868164, 0.03823980712890625, 0.038363040924072264, 0.03816198348999023, 0.03840252685546875, 0.03819472122192383, 0.03837535858154297, 0.0384983024597168, 0.038923137664794924, 0.038557441711425784, 0.038921951293945316, 0.03909686279296875, 0.03901411056518555, 0.038881568908691405, 0.039679550170898435, 0.03978899383544922, 0.03885260772705078, 0.03894879913330078, 0.0397938232421875, 0.03896207809448242, 0.03904716873168945, 0.038637569427490234, 0.03878297424316406, 0.03878860855102539, 0.03881625747680664, 0.03896223831176758, 0.038978111267089846, 0.03862531280517578, 0.03897174453735352, 0.03891404724121094, 0.03884646224975586, 0.03851459121704102, 0.0384186897277832, 0.03850425720214844, 0.03824233627319336, 0.03847372817993164, 0.04012851333618164, 0.040656993865966794, 0.03879212951660156, 0.038701473236083986, 0.038924671173095705, 0.03953478240966797, 0.04266556930541992, 0.03905177688598633, 0.03912003326416016, 0.038828479766845704, 0.03952844619750977, 0.039000064849853515, 0.03937279891967774, 0.03913113784790039, 0.03891712188720703, 0.03867750549316406, 0.03851449584960937, 0.03858041763305664, 0.03865190505981445, 0.03910246276855469, 0.03870217514038086, 0.038556385040283206, 0.03874403381347656, 0.038629600524902344, 0.038787071228027346, 0.03851264190673828, 0.038333633422851565, 0.038886207580566406, 0.03878297424316406, 0.03870105743408203, 0.04410572814941406, 0.042813438415527344, 0.038633663177490236, 0.03858179092407227, 0.03883375930786133, 0.03878572845458984, 0.038572032928466796, 0.03845513534545898, 0.038674591064453125, 0.038809600830078124, 0.03886278533935547, 0.039811134338378906, 0.04007321548461914, 0.03876192092895508, 0.038523456573486325, 0.039051136016845706, 0.03880563354492188, 0.040630271911621094, 0.03916595077514649, 0.038862335205078126, 0.03898601531982422, 0.03890332794189453, 0.038949344635009764, 0.03892364883422852, 0.03893091201782226, 0.03892671966552735, 0.03871881484985352, 0.0386952018737793, 0.038879104614257816, 0.03891804885864258, 0.03884502410888672, 0.038819839477539066, 0.0389857292175293, 0.03905305480957031, 0.038912063598632814, 0.03903302383422851, 0.03886489486694336, 0.0388587532043457, 0.03871539306640625, 0.038422527313232424, 0.0384266242980957, 0.03831193542480469, 0.03808256149291992, 0.039667713165283204, 0.039144447326660156, 0.03926496124267578, 0.0388256950378418, 0.03868662261962891, 0.038507198333740236, 0.03860847854614258, 0.039042911529541015, 0.03883065414428711, 0.03842444610595703, 0.03989311981201172, 0.038940673828125, 0.03875423812866211, 0.038809249877929684, 0.03882844924926758, 0.03877478408813476, 0.0390041618347168, 0.0387454719543457, 0.03896793746948242, 0.03910041427612305, 0.03896284866333008, 0.0391703987121582, 0.03969023895263672, 0.03902668762207031, 0.0391978874206543, 0.0389415054321289, 0.03912879943847656, 0.03921948623657227, 0.03896031951904297, 0.039172927856445314, 0.038819103240966796, 0.03887177658081055, 0.0398803825378418, 0.03888569641113281, 0.0388218879699707, 0.03864985656738281, 0.038834175109863284, 0.03915155029296875, 0.03880352020263672, 0.03975955200195312, 0.039231231689453125, 0.03879302215576172, 0.03885670471191406, 0.03873049545288086, 0.03940284729003906, 0.039010974884033205, 0.038954463958740235, 0.039031295776367186, 0.03897721481323242, 0.039046592712402343, 0.03920374298095703, 0.038927520751953125, 0.039088993072509765, 0.03887868881225586, 0.03864412689208984, 0.03913945770263672, 0.03902054214477539, 0.038980926513671875, 0.038752609252929685, 0.038695262908935546, 0.03890300750732422, 0.038789920806884766, 0.03888332748413086, 0.039701663970947265, 0.039228256225585935, 0.03905027389526367, 0.0391011848449707, 0.03913328170776367, 0.03911203384399414, 0.03893648147583008, 0.03882400131225586, 0.040325439453125, 0.039596542358398434, 0.039032318115234374, 0.03907459259033203, 0.03899772644042969, 0.0393175048828125, 0.03922534561157227, 0.038948863983154294, 0.03919404983520508, 0.038814273834228516, 0.038866462707519533, 0.038795745849609375, 0.03889152145385742, 0.038959102630615236, 0.03907174301147461, 0.03889129638671875, 0.039096126556396486, 0.03897574234008789, 0.03930444717407226, 0.039217121124267576, 0.03910329437255859, 0.039128768920898435, 0.038969791412353516, 0.03889766311645508, 0.039000064849853515, 0.03903692626953125, 0.03903241729736328, 0.0385, 0.03842940902709961, 0.03870841598510742, 0.03853398513793945, 0.03906345748901367, 0.03896124649047852, 0.040187904357910156, 0.03906355285644531, 0.03890585708618164, 0.03873382568359375, 0.03885027313232422, 0.039911487579345706, 0.03908134460449219, 0.03872854232788086, 0.038835777282714846, 0.03894454574584961, 0.03908233642578125, 0.0390843505859375, 0.03896319961547851, 0.03896115112304688, 0.039001823425292965, 0.03896758270263672, 0.03892140960693359, 0.039619392395019534, 0.039159263610839844, 0.03927664184570313, 0.03900806427001953, 0.03914543914794922, 0.03969139099121094, 0.03901529693603516, 0.03888092803955078, 0.03920316696166992, 0.03903641510009766, 0.038879425048828124, 0.03885702514648438, 0.03908607864379883, 0.03891417694091797, 0.03875980758666992, 0.03898624038696289, 0.03873174285888672, 0.03901955032348633, 0.038895713806152345, 0.03891292953491211, 0.038768638610839845, 0.03907993698120117, 0.038787071228027346, 0.0390098876953125, 0.038776798248291014, 0.03864963150024414, 0.038722206115722656, 0.03873382568359375, 0.03869283294677734, 0.038725215911865236, 0.03894316864013672, 0.03869279861450195, 0.03871955108642578, 0.038893089294433594, 0.039286945343017576, 0.0387259521484375, 0.03892633438110352, 0.03876249694824219, 0.03876192092895508, 0.03876287841796875, 0.039044704437255856, 0.039126720428466794, 0.03909084701538086, 0.03893065643310547, 0.038940704345703125, 0.03916185760498047, 0.03930316925048828, 0.03890585708618164, 0.04023651123046875, 0.039141918182373045, 0.039067649841308595, 0.03896319961547851, 0.03918000030517578, 0.03889932632446289, 0.0391761589050293, 0.038940574645996096, 0.03902684783935547, 0.03897407913208008, 0.03911030578613281, 0.03920483016967773, 0.03947148895263672, 0.03906268692016602, 0.03920323181152344, 0.03911315155029297, 0.03917350387573242, 0.039016414642333984, 0.03931625747680664, 0.03895199966430664, 0.039574462890625, 0.03878815841674805, 0.038832801818847656, 0.038750015258789065, 0.03868691253662109, 0.03871683120727539, 0.03875683212280273, 0.03871347045898438, 0.03884646224975586, 0.03885055923461914, 0.038739776611328124, 0.03879459381103516, 0.03909513473510742, 0.03890790557861328, 0.03893040084838867, 0.03879939270019531, 0.04004249572753906, 0.03907788848876953, 0.0397209587097168, 0.03895238494873047, 0.03885523223876953, 0.038834175109863284, 0.03874220657348633, 0.03882579040527344, 0.03907939147949219, 0.03887772750854492, 0.03889766311645508, 0.03886399841308594, 0.03880988693237305, 0.03866182327270508, 0.03875727844238281, 0.039241344451904296, 0.03909260940551758, 0.03893407821655273, 0.03852947235107422, 0.038540287017822264, 0.03842764663696289, 0.03885657501220703, 0.03861305618286133, 0.039230751037597655, 0.0389005126953125, 0.038959102630615236, 0.03914137649536133, 0.038967201232910156, 0.03912303924560547, 0.03965449523925781, 0.039721950531005856, 0.03889350509643555, 0.03911811065673828, 0.03899875259399414, 0.03871728134155274, 0.0387457275390625, 0.03892063903808594, 0.03887318420410156, 0.03876454544067383, 0.03869081497192383, 0.03890995025634766, 0.03882915115356445, 0.03874031829833984, 0.03882352066040039, 0.03919708633422852, 0.039110366821289065, 0.0388164176940918, 0.039627361297607425, 0.03956684875488281, 0.039147201538085936, 0.03890476989746094, 0.03886812973022461, 0.03895904159545899, 0.038984416961669925, 0.03904915237426758, 0.03920300674438477, 0.03969030380249024, 0.039058464050292965, 0.03923452758789062, 0.039202400207519535, 0.039167583465576174, 0.03921798324584961, 0.03909791946411133, 0.038984127044677734, 0.03895024108886719, 0.03899663925170899, 0.039180286407470705, 0.03935644912719727, 0.03893363189697266, 0.039057441711425785, 0.03947743988037109, 0.03936336135864258, 0.03901036834716797, 0.03896847915649414, 0.0393201904296875, 0.03906569671630859, 0.039185409545898435, 0.04003728103637695, 0.04297024154663086, 0.039443328857421876, 0.0390978889465332, 0.03906777572631836, 0.0390285758972168, 0.038877536773681644, 0.0391824951171875, 0.038942718505859376, 0.03900422286987305, 0.039091487884521485, 0.03959820938110352, 0.041269790649414065, 0.039182334899902346, 0.03887308883666992, 0.03916508865356445, 0.03939209747314453, 0.039624702453613284, 0.04179302215576172, 0.03908454513549805, 0.039139328002929685, 0.03892019271850586, 0.03915724945068359, 0.038852161407470706, 0.03858963012695313, 0.03893222427368164, 0.03885670471191406, 0.03870924758911133, 0.03889110565185547, 0.03897590255737305, 0.038860801696777345, 0.038800609588623046, 0.03901520156860352, 0.0395181770324707, 0.03904716873168945, 0.039188480377197264, 0.038830078125, 0.038978721618652346, 0.038875999450683596, 0.03873382568359375, 0.03867766571044922, 0.03865686416625977, 0.03920876693725586, 0.03901663970947265, 0.039239681243896485, 0.03895296096801758, 0.039121982574462894, 0.03938729476928711, 0.03922000122070313, 0.039070945739746094, 0.03907657623291016, 0.038979648590087894, 0.038778881072998046, 0.03894892883300781, 0.04101728057861328, 0.03911884689331055, 0.03867577743530273, 0.03891865539550781, 0.038739646911621094, 0.03896985626220703, 0.038948863983154294, 0.03899801635742187, 0.038878623962402346, 0.039000545501708984, 0.03889324951171875, 0.03895248031616211, 0.038876319885253904, 0.0389826545715332, 0.039146240234375, 0.0392171516418457, 0.039272449493408204, 0.03892559814453125, 0.039118560791015625, 0.039174686431884764, 0.039098846435546876, 0.0389911994934082, 0.03917023849487305, 0.038885025024414065, 0.039036769866943356, 0.03900902557373047, 0.03909961700439453, 0.03910131072998047, 0.03968819046020508, 0.03901043319702149, 0.03900620651245117, 0.03895280075073242, 0.038988990783691405, 0.0388966064453125, 0.03922313690185547, 0.03909836959838867, 0.03907600021362305, 0.03913657760620117, 0.03915750503540039, 0.03927257537841797, 0.03899679946899414, 0.03942371368408203]",tokens/s,25.61404973826442,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1031.5776,965.67296,0.0,570.425344,525.840896,s,1,8.07862255859375,8.07862255859375,0.0,8.07862255859375,8.07862255859375,8.07862255859375,8.07862255859375,[8.07862255859375],,kWh,3.357288517912214e-05,3.6961730086164217e-06,1.2070287433996185e-05,4.933934562173475e-05,,MB,1262.501888,1005.518848,0.0,597.68832,584.940544,s,10,1.3383642425537108,0.13383642425537107,0.0006236488037619517,0.1337007522583008,0.13432077026367187,0.1349051055908203,0.13537257385253906,"[0.13548944091796875, 0.1332783966064453, 0.13394464111328125, 0.13326112365722656, 0.13419091796875, 0.1338299560546875, 0.1333603515625, 0.13388128662109375, 0.13357154846191407, 0.13355657958984374]",tokens/s,1912.7827228223807,kWh,4.107531149247839e-06,4.5298668327437093e-07,2.7255191557223735e-06,7.2860369882445834e-06,tokens/kWh,35135698.65388204,MB,1296.502784,1020.198912,0.0,612.368384,597.290496,s,10,13.734800659179687,1.3734800659179687,0.01288839497495525,1.3706681518554689,1.3920146484375,1.3957022094726563,1.3986522583007812,"[1.37127294921875, 1.3993897705078124, 1.3911951904296875, 1.35689794921875, 1.380880126953125, 1.3673121337890626, 1.3628775634765624, 1.3700633544921874, 1.3601531982421875, 1.3747584228515626]",tokens/s,45.8688855872792,kWh,3.944550456283639e-05,4.350508324748603e-06,1.5790801212874922e-05,5.958681410045991e-05,tokens/kWh,1057280.892611336,,s,630,13.728747289657596,0.02179166236453586,0.0005926802038786021,0.021665040016174315,0.022311522483825685,0.02247623510360718,0.024017830162048347,"[0.021702592849731445, 0.021878816604614257, 0.02166275215148926, 0.021687616348266603, 0.021984960556030272, 0.021578912734985353, 0.02175062370300293, 0.02166771125793457, 0.02207084846496582, 0.022034048080444336, 0.022055200576782227, 0.022097984313964845, 0.021896095275878907, 0.021925920486450194, 0.02201795196533203, 0.021746719360351562, 0.021730016708374024, 0.021786880493164063, 0.0215467529296875, 0.02178236770629883, 0.021530176162719728, 0.021594623565673828, 0.021624927520751954, 0.02163523292541504, 0.021725120544433593, 0.021622079849243164, 0.021665760040283203, 0.02189276885986328, 0.021909887313842774, 0.021782495498657226, 0.021547008514404296, 0.021490304946899415, 0.021520479202270508, 0.02208332824707031, 0.021560543060302733, 0.021574783325195312, 0.021687423706054688, 0.02184671974182129, 0.021788671493530275, 0.02183788871765137, 0.02208064079284668, 0.022030336380004883, 0.021713184356689452, 0.021602848052978515, 0.02157369613647461, 0.021520320892333984, 0.021427520751953123, 0.02154719924926758, 0.021602176666259767, 0.021635040283203125, 0.021746496200561523, 0.021525823593139648, 0.021412384033203124, 0.021428192138671875, 0.021442592620849608, 0.021647520065307617, 0.021646400451660157, 0.02192799949645996, 0.021881568908691407, 0.02209791946411133, 0.022228191375732422, 0.022180959701538085, 0.022189504623413087, 0.02251740837097168, 0.02215872001647949, 0.02196086311340332, 0.02212713623046875, 0.02216771125793457, 0.022167680740356445, 0.022293792724609376, 0.022042783737182617, 0.022104639053344727, 0.021972991943359374, 0.022137088775634764, 0.022148992538452147, 0.022161279678344727, 0.022689727783203124, 0.022320287704467773, 0.02234432029724121, 0.02226848030090332, 0.02253385543823242, 0.02214303970336914, 0.02224051284790039, 0.022389440536499022, 0.02219798469543457, 0.02225724792480469, 0.022399904251098633, 0.022361024856567383, 0.022383583068847655, 0.022339456558227538, 0.022437887191772463, 0.022374048233032226, 0.022333791732788086, 0.022386304855346678, 0.022335872650146485, 0.022401023864746093, 0.022107295989990235, 0.02224166488647461, 0.02240086364746094, 0.02238640022277832, 0.022272672653198242, 0.022290111541748047, 0.02229110336303711, 0.02220863914489746, 0.022261056900024414, 0.022250207901000976, 0.022134880065917968, 0.022046367645263673, 0.022053056716918946, 0.022072160720825195, 0.022250560760498046, 0.022071327209472656, 0.022026111602783203, 0.022214656829833986, 0.0219869441986084, 0.021733728408813477, 0.021549087524414062, 0.021560928344726563, 0.021816831588745117, 0.021932064056396486, 0.022317024230957033, 0.022172319412231446, 0.022048831939697266, 0.02206287956237793, 0.021989791870117188, 0.022829055786132812, 0.022126272201538087, 0.022487615585327147, 0.02252288055419922, 0.02257734489440918, 0.022480928421020507, 0.02247964859008789, 0.02232089614868164, 0.022429311752319336, 0.02242188835144043, 0.022504543304443358, 0.022623136520385743, 0.02257257652282715, 0.022532928466796876, 0.022433439254760743, 0.022069440841674805, 0.022092832565307616, 0.021948640823364257, 0.021979103088378905, 0.021967456817626952, 0.021934080123901366, 0.021824800491333007, 0.021609184265136718, 0.0214466552734375, 0.02166579246520996, 0.021884927749633788, 0.023183359146118163, 0.022454143524169922, 0.021768320083618165, 0.02170457649230957, 0.02168623924255371, 0.021787872314453127, 0.021869056701660155, 0.021882976531982422, 0.021916000366210938, 0.02190745544433594, 0.02315673637390137, 0.029048032760620117, 0.02218662452697754, 0.022224992752075196, 0.021939584732055664, 0.021932064056396486, 0.021545631408691406, 0.021518335342407227, 0.02147942352294922, 0.021503135681152342, 0.021646175384521484, 0.021743167877197267, 0.021652191162109376, 0.0216529598236084, 0.021575935363769533, 0.02153267288208008, 0.021415103912353517, 0.02144054412841797, 0.02143084716796875, 0.02143459129333496, 0.021527904510498047, 0.021469856262207033, 0.021635040283203125, 0.02154044723510742, 0.02142457580566406, 0.021702079772949218, 0.02213462448120117, 0.02201468849182129, 0.02181830406188965, 0.02204457664489746, 0.02193116760253906, 0.021977407455444336, 0.021991968154907227, 0.021759008407592772, 0.021577823638916017, 0.021551103591918946, 0.021502559661865234, 0.021528863906860353, 0.021522432327270507, 0.021646976470947266, 0.021635295867919922, 0.021679712295532227, 0.021643199920654298, 0.0224586238861084, 0.021559680938720703, 0.02149171257019043, 0.021429855346679686, 0.021495967864990233, 0.02136294364929199, 0.021389375686645506, 0.021343360900878905, 0.02148547172546387, 0.021523296356201174, 0.021720191955566407, 0.02193212890625, 0.021789535522460938, 0.021488832473754882, 0.021361024856567382, 0.02126892852783203, 0.021317632675170898, 0.021340160369873046, 0.02146086311340332, 0.021521888732910156, 0.021400224685668944, 0.02136819267272949, 0.021422719955444335, 0.02159984016418457, 0.021445215225219725, 0.021520191192626954, 0.02136172866821289, 0.021476255416870118, 0.02134223937988281, 0.02130534362792969, 0.02128873634338379, 0.021713119506835937, 0.021321727752685548, 0.021235712051391603, 0.021247648239135743, 0.021389375686645506, 0.021274528503417968, 0.021299264907836915, 0.021408063888549805, 0.021352287292480468, 0.0213504638671875, 0.021667903900146484, 0.02183359909057617, 0.021438623428344728, 0.021362688064575194, 0.021378175735473633, 0.021601152420043946, 0.021433504104614257, 0.0213175048828125, 0.021526016235351563, 0.021439104080200194, 0.0212620792388916, 0.021495519638061525, 0.021406240463256836, 0.02153990364074707, 0.022257919311523437, 0.026421279907226564, 0.026288799285888672, 0.022142431259155274, 0.02246441650390625, 0.022035104751586914, 0.021995840072631837, 0.021937503814697265, 0.023060800552368164, 0.021854015350341798, 0.02177043151855469, 0.021626880645751953, 0.021522432327270507, 0.02141961669921875, 0.021419519424438475, 0.02145372772216797, 0.02163692855834961, 0.021555391311645508, 0.021440383911132812, 0.021353887557983398, 0.02162761688232422, 0.02409062385559082, 0.022184032440185547, 0.021687904357910157, 0.021862207412719728, 0.02162067222595215, 0.02166431999206543, 0.021594112396240234, 0.021573631286621094, 0.021780479431152345, 0.022183935165405275, 0.02387721633911133, 0.021660064697265623, 0.022101631164550783, 0.02465830421447754, 0.02158729553222656, 0.021487775802612304, 0.02139187240600586, 0.021417600631713867, 0.021430656433105467, 0.021319679260253906, 0.021555200576782226, 0.02135161590576172, 0.02143929672241211, 0.02133331108093262, 0.02139411163330078, 0.021384191513061524, 0.021436704635620116, 0.02138185691833496, 0.021420032501220702, 0.021888448715209962, 0.021533248901367187, 0.02147327995300293, 0.021407743453979493, 0.021549055099487305, 0.021398944854736326, 0.021373056411743165, 0.021515743255615234, 0.021507776260375977, 0.02141881561279297, 0.02145712089538574, 0.021477184295654296, 0.02146713638305664, 0.02146303939819336, 0.02192793655395508, 0.021510143280029297, 0.02147478485107422, 0.021602848052978515, 0.0215285758972168, 0.02148761558532715, 0.021372415542602538, 0.022249824523925783, 0.021416095733642577, 0.02141779136657715, 0.021572864532470704, 0.021654464721679687, 0.021766368865966796, 0.021690143585205077, 0.02177008056640625, 0.022055072784423826, 0.02206924819946289, 0.021929983139038087, 0.021942272186279296, 0.02198297691345215, 0.02184623908996582, 0.021725471496582032, 0.021806848526000976, 0.021556480407714844, 0.021584640502929686, 0.021392831802368163, 0.021752384185791014, 0.02187264060974121, 0.021738847732543944, 0.02169718360900879, 0.021997312545776367, 0.02132371139526367, 0.021420095443725588, 0.021418239593505858, 0.021415327072143556, 0.021463487625122072, 0.02142633628845215, 0.021366239547729492, 0.021563936233520507, 0.021381120681762695, 0.021409791946411134, 0.021653600692749023, 0.02192095947265625, 0.02189753532409668, 0.021774751663208008, 0.02178767967224121, 0.021828575134277342, 0.02196678352355957, 0.021846208572387695, 0.021886240005493163, 0.021848672866821288, 0.021755903244018555, 0.02210028839111328, 0.023135135650634766, 0.02203126335144043, 0.021719039916992186, 0.02205695915222168, 0.021729183197021485, 0.02155945587158203, 0.021429695129394532, 0.02132633590698242, 0.02134864044189453, 0.021337823867797853, 0.02127462387084961, 0.02138444709777832, 0.02133011245727539, 0.021354143142700194, 0.022588319778442383, 0.021823200225830078, 0.02189651107788086, 0.021855199813842773, 0.021624767303466796, 0.021817407608032226, 0.021544639587402343, 0.02153081512451172, 0.0216060791015625, 0.021567264556884767, 0.02148198318481445, 0.021662111282348632, 0.021562751770019532, 0.021561088562011718, 0.02154969596862793, 0.021431488037109377, 0.021582015991210936, 0.02154355239868164, 0.02146665573120117, 0.021345760345458983, 0.021371488571166993, 0.021223072052001954, 0.021174495697021484, 0.021523040771484377, 0.021548160552978514, 0.021545791625976564, 0.021733312606811522, 0.021442432403564454, 0.02148953628540039, 0.02141209602355957, 0.02160214424133301, 0.021381343841552734, 0.021405696868896484, 0.02137196731567383, 0.021410207748413086, 0.02113795280456543, 0.02137276840209961, 0.021481632232666015, 0.021962080001831055, 0.022200159072875977, 0.022096832275390624, 0.021919904708862306, 0.021890111923217773, 0.02185487937927246, 0.021747711181640626, 0.022179840087890625, 0.021781856536865235, 0.02171356773376465, 0.022060895919799806, 0.02214076805114746, 0.02211187171936035, 0.02206230354309082, 0.022309759140014648, 0.022353279113769532, 0.02225619125366211, 0.022468576431274413, 0.02231091117858887, 0.022223007202148436, 0.022251264572143555, 0.022472063064575196, 0.02260479927062988, 0.022452064514160156, 0.022458240509033202, 0.022255104064941408, 0.022212287902832032, 0.022600255966186523, 0.022460800170898437, 0.022235231399536134, 0.022191904067993165, 0.021941984176635742, 0.02187708854675293, 0.0217640323638916, 0.021583520889282226, 0.021584096908569335, 0.021477504730224608, 0.021265632629394533, 0.021238304138183593, 0.021360895156860352, 0.021475360870361327, 0.021449823379516602, 0.0213239688873291, 0.021285568237304688, 0.02128611183166504, 0.02131817626953125, 0.021125375747680666, 0.021245023727416993, 0.021207967758178712, 0.021242944717407227, 0.02142464065551758, 0.021344703674316408, 0.02126848030090332, 0.021237024307250975, 0.021572032928466798, 0.0217458553314209, 0.021592159271240235, 0.02130454444885254, 0.021158239364624024, 0.021221824645996094, 0.021280096054077147, 0.021330591201782226, 0.021235712051391603, 0.021273664474487305, 0.021277631759643555, 0.021456127166748048, 0.02161347198486328, 0.021903263092041016, 0.021936063766479493, 0.02186240005493164, 0.02189004707336426, 0.02178755187988281, 0.02165155220031738, 0.021813247680664064, 0.02212563133239746, 0.021910207748413086, 0.02211043167114258, 0.02199692726135254, 0.021827583312988282, 0.02170911979675293, 0.021618879318237305, 0.021418367385864258, 0.02136819267272949, 0.02137766456604004, 0.021579776763916016, 0.021348352432250976, 0.02123097610473633, 0.02207916831970215, 0.021271488189697266, 0.021226879119873046, 0.021596799850463866, 0.021733184814453126, 0.02163020706176758, 0.021308416366577147, 0.021261407852172853, 0.02114406394958496, 0.02109270477294922, 0.021130399703979494, 0.021230432510375978, 0.021208127975463866, 0.021429183959960938, 0.021269535064697264, 0.02134048080444336, 0.021324447631835938, 0.021481472015380858, 0.02127257537841797, 0.02157360076904297, 0.021927967071533203, 0.022042623519897463, 0.02186467170715332, 0.021880064010620117, 0.021869087219238283, 0.021760000228881835, 0.02184339141845703, 0.022709856033325194, 0.02205820846557617, 0.02190140724182129, 0.021976383209228515, 0.022002016067504883, 0.022052831649780273, 0.02197711944580078, 0.021712896347045898, 0.021544960021972655, 0.021401599884033205, 0.021405183792114257, 0.021868703842163086, 0.02148796844482422, 0.02171801567077637, 0.021696575164794923, 0.02139846420288086, 0.02128233528137207, 0.021367200851440428, 0.021368703842163085, 0.02129142379760742, 0.02128463935852051, 0.02125791931152344, 0.021264543533325197, 0.021666271209716797, 0.021560415267944336, 0.021874912261962892, 0.02232694435119629, 0.023127647399902345, 0.022128639221191407, 0.022046560287475585, 0.021928640365600587, 0.021840896606445313, 0.02188582420349121, 0.0218951358795166, 0.021994911193847656, 0.02191231918334961, 0.021745471954345702, 0.021575872421264648, 0.02407526397705078, 0.025135135650634764, 0.021981632232666016, 0.02176630401611328, 0.021655071258544923, 0.021474111557006837, 0.02142620849609375, 0.021397504806518555, 0.021338111877441408, 0.02122127914428711, 0.021424224853515625, 0.02128281593322754, 0.02134556770324707, 0.021357280731201172, 0.02142207908630371, 0.02139116859436035, 0.02132601547241211, 0.02228953552246094, 0.021865215301513672, 0.021739648818969726, 0.021379423141479493, 0.02167363166809082, 0.021344255447387696, 0.021331647872924804, 0.021600576400756837, 0.021398815155029297, 0.021350303649902345, 0.021289535522460937, 0.02183737564086914, 0.0230447998046875, 0.021299104690551757, 0.02127676773071289, 0.021215200424194336, 0.02127177619934082, 0.021342655181884766, 0.021270591735839842, 0.021283136367797852, 0.021348031997680664, 0.021360960006713867, 0.021540864944458008, 0.02153267288208008, 0.02186854362487793, 0.022068384170532227, 0.022172512054443358, 0.0221693115234375, 0.022091487884521484, 0.02220921516418457, 0.02226982307434082, 0.022742528915405274, 0.022331327438354493]",tokens/s,45.88911039790236,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1551.089664,1823.408128,0.0,1428.160512,1322.516992,s,1,8.658736328125,8.658736328125,0.0,8.658736328125,8.658736328125,8.658736328125,8.658736328125,[8.658736328125],,kWh,4.8651675595783675e-05,5.3587232173526815e-06,1.8511681475999464e-05,7.252208028913582e-05,,MB,1581.109248,1842.282496,0.0,1434.451968,1320.892416,s,10,5.406551818847656,0.5406551818847657,0.0018826069571755957,0.5399053039550781,0.5420017639160156,0.5439743316650391,0.5455523858642578,"[0.5459468994140625, 0.5397939453125, 0.5397861938476562, 0.5399390258789063, 0.5397224731445313, 0.5403477783203126, 0.5389154663085938, 0.53987158203125, 0.5415634155273438, 0.5406650390625]",tokens/s,473.49957713817565,kWh,1.6123777890129133e-05,1.778195864228297e-06,1.0764438436105498e-05,2.8666412190462927e-05,tokens/kWh,8930311.833204193,MB,1588.71552,1842.282496,0.0,1434.451968,1373.031936,s,10,13.761650512695315,1.3761650512695311,0.012817914920764084,1.3816153564453124,1.388759130859375,1.3899785034179688,1.3909540014648436,"[1.387599609375, 1.36535986328125, 1.362013427734375, 1.3884881591796876, 1.3596466064453125, 1.3572025146484374, 1.3869117431640625, 1.3838553466796875, 1.3793753662109376, 1.3911978759765624]",tokens/s,45.7793924804889,kWh,3.9648751469035636e-05,4.373167839454321e-06,1.952119690349422e-05,6.354311621198417e-05,tokens/kWh,991452.7923029098,,s,630,13.758981756210328,0.021839653581286235,0.00047273162063906005,0.021813504219055174,0.022227411842346192,0.022362587451934812,0.023115124950408936,"[0.022732927322387696, 0.02229055976867676, 0.022194175720214843, 0.02215488052368164, 0.022108543395996095, 0.02215116882324219, 0.02215884780883789, 0.022047231674194336, 0.022034431457519533, 0.022158624649047852, 0.022198047637939453, 0.02199852752685547, 0.022642303466796875, 0.02233996772766113, 0.0221200008392334, 0.022086080551147462, 0.021977088928222657, 0.021987327575683592, 0.021929983139038087, 0.021877983093261718, 0.021990175247192382, 0.022074367523193358, 0.02205183982849121, 0.02214297676086426, 0.02205286407470703, 0.022256927490234377, 0.022110944747924806, 0.0218603515625, 0.0218984317779541, 0.021882944107055664, 0.02193280029296875, 0.021741567611694337, 0.02198944091796875, 0.021686208724975585, 0.02184169578552246, 0.02183788871765137, 0.02205302429199219, 0.022338623046875, 0.021892032623291015, 0.02186240005493164, 0.021952512741088868, 0.022035808563232423, 0.022062847137451172, 0.022008544921875, 0.022433183670043946, 0.022162208557128905, 0.02211020851135254, 0.02181475257873535, 0.02178236770629883, 0.021785280227661134, 0.021991424560546875, 0.021917695999145507, 0.021968896865844727, 0.021804767608642577, 0.021894655227661132, 0.02188368034362793, 0.021996959686279297, 0.021502559661865234, 0.022294015884399415, 0.02218239974975586, 0.021835775375366212, 0.02166169548034668, 0.02159414482116699, 0.022390560150146486, 0.021788543701171875, 0.02145948791503906, 0.02142799949645996, 0.021549535751342774, 0.02152556800842285, 0.021461952209472657, 0.021558624267578125, 0.021744287490844727, 0.021548128128051756, 0.02155641555786133, 0.021513952255249023, 0.02176326370239258, 0.021537599563598634, 0.02156883239746094, 0.02148624038696289, 0.02143235206604004, 0.021972383499145508, 0.021589599609375, 0.021505023956298826, 0.021550304412841798, 0.021590688705444335, 0.021573631286621094, 0.02153615951538086, 0.02148137664794922, 0.02144499206542969, 0.02152707290649414, 0.021484479904174805, 0.021455072402954103, 0.021498207092285157, 0.021509599685668946, 0.02144268798828125, 0.021506879806518556, 0.021865535736083984, 0.021571968078613283, 0.021571744918823244, 0.021923871994018556, 0.021454816818237306, 0.02149622344970703, 0.02151219177246094, 0.021716991424560548, 0.021843967437744142, 0.02159008026123047, 0.021374624252319337, 0.021610784530639648, 0.021411840438842773, 0.02143020820617676, 0.021456064224243163, 0.021568384170532227, 0.021540864944458008, 0.021626880645751953, 0.028711999893188477, 0.021480384826660155, 0.02139311981201172, 0.021430559158325195, 0.021357791900634766, 0.02151641654968262, 0.021504255294799806, 0.021391775131225584, 0.0214835205078125, 0.021388639450073244, 0.021406368255615236, 0.02151628875732422, 0.022323392868041993, 0.021788671493530275, 0.021395456314086913, 0.02134601593017578, 0.02140166473388672, 0.021393632888793944, 0.021426111221313476, 0.021405759811401366, 0.021536447525024413, 0.021333696365356446, 0.02143824005126953, 0.02131235122680664, 0.021409631729125977, 0.021681535720825196, 0.021392160415649415, 0.02147737693786621, 0.021553152084350585, 0.021355712890625, 0.021819360733032228, 0.02155196762084961, 0.02147327995300293, 0.021397247314453124, 0.021975072860717773, 0.0214932804107666, 0.021492351531982423, 0.02139491271972656, 0.02139187240600586, 0.02141606330871582, 0.021422048568725587, 0.021429952621459962, 0.02146131134033203, 0.021401472091674804, 0.02133932876586914, 0.02142617607116699, 0.021398271560668945, 0.021440992355346678, 0.021495264053344728, 0.02139084815979004, 0.021424896240234376, 0.02159187126159668, 0.02139308738708496, 0.021527040481567384, 0.02146672058105469, 0.021402015686035156, 0.021566495895385743, 0.021586912155151367, 0.021729280471801758, 0.02133795166015625, 0.021384864807128905, 0.021385440826416014, 0.021597471237182617, 0.021590944290161132, 0.02170844841003418, 0.02198476791381836, 0.02207427215576172, 0.02220444869995117, 0.022140928268432617, 0.022216384887695312, 0.022279903411865233, 0.02229216003417969, 0.02230521583557129, 0.022499296188354494, 0.02236672019958496, 0.02249932861328125, 0.022130624771118164, 0.022227008819580077, 0.022287519454956054, 0.0221210880279541, 0.021995712280273437, 0.0220098876953125, 0.023076864242553712, 0.021917695999145507, 0.021898880004882812, 0.02189072036743164, 0.021869440078735352, 0.021883968353271485, 0.021932832717895506, 0.021812768936157228, 0.021905311584472655, 0.02187321662902832, 0.021942272186279296, 0.021977088928222657, 0.02183286476135254, 0.021875551223754883, 0.021803007125854493, 0.021857280731201172, 0.02189411163330078, 0.022249216079711913, 0.021819679260253907, 0.021970943450927736, 0.022188032150268554, 0.02192793655395508, 0.021876735687255858, 0.02191360092163086, 0.021962751388549806, 0.02187059211730957, 0.021816991806030275, 0.022091264724731444, 0.021862592697143555, 0.022067520141601564, 0.021899839401245118, 0.021927711486816406, 0.022129728317260743, 0.021965408325195314, 0.02183612823486328, 0.021805055618286134, 0.02197491264343262, 0.02208576011657715, 0.022122495651245116, 0.021991424560546875, 0.022486047744750978, 0.023318784713745117, 0.022468639373779298, 0.021936832427978517, 0.021925888061523437, 0.022003456115722655, 0.022155519485473632, 0.02228223991394043, 0.021978431701660157, 0.022098560333251954, 0.022263103485107422, 0.021723264694213866, 0.0216408634185791, 0.022010847091674803, 0.022099967956542968, 0.021968896865844727, 0.02210358428955078, 0.021953216552734377, 0.021868064880371095, 0.021566112518310546, 0.021503231048583985, 0.021520639419555666, 0.021477888107299805, 0.021386592864990235, 0.02145552062988281, 0.021403648376464843, 0.021370336532592772, 0.021404191970825194, 0.021360288619995116, 0.021329504013061523, 0.021325664520263674, 0.021418912887573242, 0.021448703765869142, 0.021381120681762695, 0.021481184005737303, 0.02135478401184082, 0.02147737693786621, 0.021376031875610352, 0.022205215454101562, 0.021833087921142576, 0.022274784088134766, 0.02146108818054199, 0.02161039924621582, 0.021391456604003906, 0.021389312744140625, 0.021559007644653322, 0.021623071670532228, 0.02135759925842285, 0.021814239501953124, 0.02154697608947754, 0.021538719177246094, 0.021579904556274412, 0.021514240264892577, 0.02149580764770508, 0.02135183906555176, 0.02129078483581543, 0.02143836784362793, 0.02127964782714844, 0.02170675277709961, 0.022154943466186523, 0.021844287872314454, 0.021556863784790038, 0.021458976745605467, 0.0215978889465332, 0.02149033546447754, 0.021706560134887695, 0.021409984588623046, 0.021489664077758788, 0.021397504806518555, 0.021728319168090822, 0.02165241622924805, 0.021743616104125976, 0.02282700729370117, 0.021932031631469725, 0.02149580764770508, 0.021211135864257814, 0.021394975662231447, 0.02148963165283203, 0.02158847999572754, 0.022292287826538085, 0.021743711471557618, 0.021350496292114256, 0.02128281593322754, 0.02141798400878906, 0.021569408416748048, 0.021389696121215822, 0.021673055648803712, 0.021422752380371092, 0.021352447509765626, 0.02149068832397461, 0.021351423263549805, 0.02167919921875, 0.021492095947265626, 0.021463584899902344, 0.02156300735473633, 0.021485855102539062, 0.02147689628601074, 0.021402175903320313, 0.021602304458618164, 0.02167788887023926, 0.02170460891723633, 0.021507808685302734, 0.021553695678710936, 0.02134223937988281, 0.021387231826782226, 0.02143951988220215, 0.021351039886474608, 0.021451200485229492, 0.021409727096557616, 0.02146099281311035, 0.021391359329223633, 0.021348127365112303, 0.02150147247314453, 0.021559776306152342, 0.0215185604095459, 0.021406015396118163, 0.021614271163940428, 0.021630975723266603, 0.021581279754638673, 0.02155695915222168, 0.021373247146606444, 0.021422592163085938, 0.021366783142089844, 0.021428224563598632, 0.021436128616333008, 0.021448991775512696, 0.021366783142089844, 0.021421695709228517, 0.021481855392456055, 0.021444351196289062, 0.02144895935058594, 0.021583871841430666, 0.02146713638305664, 0.021853952407836913, 0.021426111221313476, 0.021565759658813476, 0.02155107116699219, 0.022298656463623046, 0.021936128616333008, 0.02188697624206543, 0.021966848373413086, 0.0218603515625, 0.022382848739624022, 0.022038816452026367, 0.02203209686279297, 0.021842208862304688, 0.021766143798828123, 0.02169148826599121, 0.02182032012939453, 0.02187059211730957, 0.021991424560546875, 0.022155263900756835, 0.02230659294128418, 0.022089696884155272, 0.022496736526489258, 0.02226460838317871, 0.022221887588500976, 0.02218079948425293, 0.022231039047241212, 0.02235753631591797, 0.02220460891723633, 0.02210812759399414, 0.022091392517089845, 0.02233145523071289, 0.022061695098876954, 0.021968832015991212, 0.022100351333618165, 0.022066303253173828, 0.022179391860961913, 0.02209689521789551, 0.022101280212402343, 0.02207606315612793, 0.022141088485717775, 0.02204457664489746, 0.02209587287902832, 0.022033727645874024, 0.02186310386657715, 0.021941919326782227, 0.0217030086517334, 0.021700607299804688, 0.021730783462524415, 0.02182601547241211, 0.021897279739379882, 0.021961023330688476, 0.02224892807006836, 0.02221897506713867, 0.021992895126342775, 0.021899839401245118, 0.02185215950012207, 0.021723072052001954, 0.02176211166381836, 0.022271999359130858, 0.021953567504882813, 0.021851104736328127, 0.021839391708374022, 0.021895647048950195, 0.02173891258239746, 0.02186422348022461, 0.02186854362487793, 0.022131519317626955, 0.02197465515136719, 0.021764480590820312, 0.021946367263793946, 0.021825088500976562, 0.021975488662719728, 0.02232150459289551, 0.022534208297729494, 0.02195382308959961, 0.021738208770751954, 0.021739519119262696, 0.021570655822753908, 0.02169036865234375, 0.021733503341674804, 0.021594623565673828, 0.02169990348815918, 0.021857248306274415, 0.02216307258605957, 0.022047103881835936, 0.022265951156616212, 0.02210758399963379, 0.02241539192199707, 0.02245849609375, 0.022432064056396483, 0.022206464767456056, 0.022109600067138673, 0.022280799865722657, 0.022203903198242187, 0.02214246368408203, 0.022150144577026368, 0.021938175201416017, 0.021893119812011717, 0.021798912048339843, 0.02194793510437012, 0.021952543258666992, 0.021936288833618166, 0.02209542465209961, 0.021946720123291016, 0.022083967208862305, 0.021876735687255858, 0.02190745544433594, 0.021716991424560548, 0.021776384353637695, 0.021677343368530274, 0.02213475227355957, 0.021623008728027342, 0.0216724796295166, 0.021675264358520508, 0.02178643226623535, 0.02196575927734375, 0.02199942398071289, 0.022276063919067383, 0.02212681579589844, 0.02251366424560547, 0.022112159729003905, 0.022044767379760744, 0.021982303619384767, 0.022094335556030274, 0.0220546875, 0.02185241508483887, 0.02199795150756836, 0.02193222427368164, 0.021753664016723632, 0.02184601593017578, 0.021763296127319337, 0.021762624740600586, 0.021559167861938476, 0.021690336227416993, 0.02143395233154297, 0.02251363182067871, 0.022097759246826172, 0.02167344093322754, 0.021642112731933595, 0.021679807662963867, 0.022042367935180666, 0.022069280624389648, 0.021778976440429688, 0.021812671661376952, 0.022970592498779297, 0.022690271377563475, 0.021696224212646484, 0.021788576126098632, 0.021741823196411134, 0.021729280471801758, 0.02179852867126465, 0.0217706241607666, 0.021802400588989256, 0.021889503479003907, 0.02537295913696289, 0.022071136474609374, 0.022003231048583986, 0.021711328506469726, 0.021932031631469725, 0.022697984695434572, 0.021491424560546875, 0.023031967163085938, 0.02153875160217285, 0.021563232421875, 0.02159222412109375, 0.021632352828979493, 0.021604192733764647, 0.021445280075073243, 0.021592416763305665, 0.021596160888671875, 0.021736928939819336, 0.021794815063476563, 0.02173187255859375, 0.022340639114379883, 0.021736160278320312, 0.021530879974365234, 0.02149580764770508, 0.022005151748657227, 0.021895551681518556, 0.021725343704223632, 0.02161235237121582, 0.021616416931152342, 0.021498016357421875, 0.02161257553100586, 0.021661983489990235, 0.021651456832885742, 0.022063039779663087, 0.022046592712402342, 0.021790304183959962, 0.021865055084228514, 0.02168832015991211, 0.021784576416015625, 0.021704704284667968, 0.021594112396240234, 0.021761215209960938, 0.021719871520996095, 0.02165328025817871, 0.021745887756347657, 0.022271999359130858, 0.02210406494140625, 0.02221571159362793, 0.022119392395019533, 0.022124544143676757, 0.021991519927978515, 0.021970848083496093, 0.02190745544433594, 0.02203225517272949, 0.02192313575744629, 0.02266604804992676, 0.021987327575683592, 0.022017183303833007, 0.02194927978515625, 0.021976543426513673, 0.022034975051879884, 0.021966848373413086, 0.02205695915222168, 0.022153215408325197, 0.022058176040649413, 0.022296960830688477, 0.022066719055175783, 0.02186537551879883, 0.021855520248413085, 0.021960575103759764, 0.0218438720703125, 0.021783775329589843, 0.021744960784912108, 0.02162719917297363, 0.021547071456909178, 0.02160438346862793, 0.021643264770507813, 0.02161769676208496, 0.021674240112304687, 0.021815519332885742, 0.021922304153442384, 0.021978559494018553, 0.02203001594543457, 0.02213158416748047, 0.022132415771484375, 0.022081855773925782, 0.023324031829833985, 0.02313075256347656, 0.023174816131591797, 0.022206815719604492, 0.022177791595458983, 0.0221343994140625, 0.02210380744934082, 0.021926528930664064, 0.021745664596557617, 0.021782527923583983, 0.02185558319091797, 0.021990047454833985, 0.021946367263793946, 0.021972991943359374, 0.021766143798828123, 0.022494527816772462, 0.0219202880859375, 0.021719200134277344, 0.02168627166748047, 0.02186444854736328, 0.022248767852783204, 0.024951488494873046]",tokens/s,45.788272065673745,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,6652.2112,9171.828736,0.0,8776.58112,8188.314112,s,1,13.7148662109375,13.7148662109375,0.0,13.7148662109375,13.7148662109375,13.7148662109375,13.7148662109375,[13.7148662109375],,kWh,0.00019776101324584942,2.180726305868196e-05,8.863895979999148e-05,0.0003082072361045229,,MB,1611.063296,9184.411648,0.0,8776.58112,7663.089664,s,10,52.020645507812496,5.20206455078125,0.004295593668816398,5.203471435546875,5.2058587890625,5.20640107421875,5.20683490234375,"[5.1930009765625, 5.1960810546875, 5.19998583984375, 5.20151123046875, 5.20285498046875, 5.204087890625, 5.2052236328125, 5.20521826171875, 5.206943359375, 5.20573828125]",tokens/s,49.21123094513576,kWh,0.000151715859933751,1.673465758453569e-05,0.00010085163623680116,0.00026930215375508784,tokens/kWh,950605.0970272401,MB,1615.388672,9184.411648,0.0,8776.58112,7906.289152,s,10,27.204001953125,2.7204001953125,0.005295004331255583,2.719349853515625,2.7275160644531247,2.7290044921875,2.730195234375,"[2.730492919921875, 2.720167236328125, 2.72054541015625, 2.712332275390625, 2.718532470703125, 2.7180791015625, 2.718348876953125, 2.7242158203125, 2.727185302734375, 2.7141025390625]",tokens/s,23.15835740217737,kWh,7.942551686416889e-05,8.759270436297493e-06,5.2842431162797285e-05,0.00014102721846326376,tokens/kWh,446722.2759300957,,s,630,27.200278923034652,0.04317504590957884,0.001029309602478756,0.04311952018737793,0.04354717712402344,0.04367663974761963,0.044888739929199215,"[0.044818687438964847, 0.04318556976318359, 0.042661857604980466, 0.04239555358886719, 0.04245932769775391, 0.04239238357543945, 0.04240758514404297, 0.04236528015136719, 0.04238934326171875, 0.04231129455566406, 0.04267472076416016, 0.042624542236328125, 0.04263103866577148, 0.042783329010009766, 0.042718910217285154, 0.04242243194580078, 0.042463390350341794, 0.042602497100830077, 0.0427334098815918, 0.042920097351074216, 0.042807296752929686, 0.042788864135742184, 0.043043903350830075, 0.04311750411987305, 0.043179264068603514, 0.04288131332397461, 0.04304291152954102, 0.04286880111694336, 0.042764190673828126, 0.04274198532104492, 0.04274335861206055, 0.04271142578125, 0.042905857086181644, 0.04291936111450195, 0.04287539291381836, 0.043031841278076174, 0.042961536407470705, 0.042874366760253906, 0.042887775421142575, 0.04275820922851563, 0.06539222717285156, 0.04241206359863281, 0.04287731170654297, 0.04307558441162109, 0.043268096923828124, 0.04322889709472656, 0.043165313720703126, 0.04331996917724609, 0.043304576873779296, 0.04323113632202148, 0.04316937637329102, 0.04300479888916016, 0.043153408050537106, 0.0430439682006836, 0.04330995178222656, 0.04322304153442383, 0.04343711853027344, 0.0434079360961914, 0.0433790397644043, 0.04329856109619141, 0.04425142288208008, 0.044365825653076174, 0.04395212936401367, 0.04469139099121094, 0.043185344696044924, 0.04275289535522461, 0.042610527038574215, 0.04248710250854492, 0.04231439971923828, 0.04239583969116211, 0.04247504043579101, 0.04245888137817383, 0.042500831604003905, 0.04270284652709961, 0.04273152160644531, 0.043069438934326174, 0.0466874885559082, 0.04258483123779297, 0.04294652938842773, 0.0431426887512207, 0.0437314567565918, 0.04306534576416016, 0.04292800140380859, 0.04302451324462891, 0.043251712799072264, 0.04347289657592773, 0.044889408111572264, 0.043261920928955075, 0.043052833557128904, 0.042955711364746095, 0.04319641494750977, 0.04309196853637695, 0.04300115203857422, 0.04294521713256836, 0.04299980926513672, 0.043044864654541014, 0.04322099304199219, 0.04318207931518555, 0.04308777618408203, 0.043259998321533204, 0.04319641494750977, 0.04310419082641601, 0.043164798736572266, 0.043117279052734374, 0.04301846313476562, 0.0429854736328125, 0.043194366455078126, 0.04326604843139648, 0.04340262222290039, 0.04330950546264648, 0.04329081726074219, 0.043337535858154294, 0.04318220901489258, 0.043179393768310544, 0.043162303924560545, 0.043019454956054685, 0.04294960021972656, 0.04308518218994141, 0.04320099258422851, 0.04322089767456055, 0.043120193481445315, 0.04331164932250976, 0.043294689178466794, 0.043431968688964845, 0.04343555068969727, 0.04349359893798828, 0.045954593658447264, 0.04321571350097656, 0.042686561584472656, 0.04260240173339844, 0.04250783920288086, 0.042396095275878905, 0.04259430313110352, 0.04260454559326172, 0.042471424102783206, 0.04243791961669922, 0.04267475128173828, 0.043077056884765624, 0.04263302230834961, 0.04262742233276367, 0.042774879455566406, 0.04279318237304688, 0.043371742248535156, 0.042711841583251954, 0.04295459365844727, 0.042731422424316406, 0.04287513732910156, 0.043081729888916016, 0.043169792175292966, 0.04296089553833008, 0.04301824188232422, 0.04301126480102539, 0.042875713348388675, 0.04285638427734375, 0.04290351867675781, 0.04292617416381836, 0.04300172805786133, 0.04277670288085938, 0.04258822250366211, 0.04273740768432617, 0.042894561767578124, 0.042863582611083986, 0.042913791656494144, 0.04302627182006836, 0.04298688125610352, 0.043837440490722655, 0.04973852920532226, 0.042571136474609375, 0.043243263244628904, 0.043398017883300784, 0.04322873687744141, 0.04338035202026367, 0.043338272094726564, 0.043256095886230465, 0.04330873489379883, 0.04333804702758789, 0.04362041473388672, 0.04332128143310547, 0.043174049377441404, 0.04308268737792969, 0.04338188934326172, 0.043450145721435546, 0.04333324813842773, 0.04328275299072266, 0.04324723052978516, 0.04338937759399414, 0.04359360122680664, 0.043910560607910154, 0.04346748733520508, 0.044711071014404295, 0.043092830657958985, 0.042708606719970704, 0.04256204986572266, 0.04248767852783203, 0.04247449493408203, 0.0427407341003418, 0.04258611297607422, 0.042627071380615236, 0.04251846313476562, 0.04239558410644531, 0.04252060699462891, 0.04258415985107422, 0.0425082893371582, 0.04279430389404297, 0.04303737640380859, 0.042866687774658206, 0.04270073699951172, 0.04267628860473633, 0.04267827224731445, 0.04280319976806641, 0.042858592987060545, 0.043224990844726564, 0.043245567321777346, 0.04317184066772461, 0.04290150451660156, 0.042872833251953124, 0.04278988647460937, 0.04284928131103516, 0.04300815963745117, 0.043040607452392576, 0.04281760025024414, 0.04310214233398438, 0.04286054229736328, 0.0429752311706543, 0.04298076629638672, 0.04290204620361328, 0.043060928344726565, 0.04306367874145508, 0.04298137664794922, 0.04295065689086914, 0.04322099304199219, 0.04331315231323242, 0.043263233184814454, 0.04352486419677734, 0.043491329193115234, 0.0433616943359375, 0.043298847198486326, 0.04332191848754883, 0.04305920028686523, 0.04320012664794922, 0.04355340957641601, 0.043599006652832034, 0.043517822265625, 0.043375328063964845, 0.04328611373901367, 0.043173343658447265, 0.04349004745483399, 0.04346303939819336, 0.04349747085571289, 0.043593536376953124, 0.043319297790527345, 0.04336588668823242, 0.04476918411254883, 0.043241119384765624, 0.04278112030029297, 0.042517696380615234, 0.042549758911132815, 0.042489696502685546, 0.04260665512084961, 0.04238787078857422, 0.04252057647705078, 0.04266995239257813, 0.042616958618164065, 0.043055103302001956, 0.04300595092773438, 0.04292540740966797, 0.04297798538208008, 0.042784126281738284, 0.04290595245361328, 0.04292633438110351, 0.042891231536865235, 0.04323331069946289, 0.04332134246826172, 0.04316073608398437, 0.0433974723815918, 0.043299327850341796, 0.04333929443359375, 0.04308835220336914, 0.043118015289306644, 0.04307321548461914, 0.04299638366699219, 0.04293273544311523, 0.04304067230224609, 0.04319750213623047, 0.04307839965820313, 0.04309811019897461, 0.04332476806640625, 0.043118846893310546, 0.04307190322875976, 0.04326544189453125, 0.04319907379150391, 0.04310630416870117, 0.04326435089111328, 0.04323907089233398, 0.04342784118652344, 0.043401214599609376, 0.043276287078857424, 0.04331315231323242, 0.043621952056884766, 0.04346681594848633, 0.04331686401367187, 0.04335388946533203, 0.04330582427978515, 0.043186302185058596, 0.043053054809570314, 0.04321203231811523, 0.043375553131103514, 0.043329151153564456, 0.043346111297607424, 0.043276287078857424, 0.04325392150878906, 0.04348419189453125, 0.043533119201660156, 0.043484928131103516, 0.04356940841674805, 0.04481635284423828, 0.04324969482421875, 0.04271104049682617, 0.04259392166137695, 0.04245951843261719, 0.042590206146240234, 0.04264550399780274, 0.042602497100830077, 0.04275404739379883, 0.042692607879638675, 0.04270182418823242, 0.04266912078857422, 0.04276627349853516, 0.04289913558959961, 0.042893630981445316, 0.042856224060058595, 0.04278908920288086, 0.04279075241088867, 0.04280131149291992, 0.04284963226318359, 0.04297385787963867, 0.043224288940429685, 0.04346480178833008, 0.043374622344970706, 0.04333955383300781, 0.04315631866455078, 0.04293756866455078, 0.04297401428222656, 0.0429936637878418, 0.043063297271728515, 0.04308377456665039, 0.04312063980102539, 0.042894783020019533, 0.04293280029296875, 0.04296236801147461, 0.04306796646118164, 0.04342156982421875, 0.04318217468261719, 0.0432243537902832, 0.04318899154663086, 0.0430571517944336, 0.04324777603149414, 0.043300640106201174, 0.04349708938598633, 0.04344380950927734, 0.04354339218139648, 0.04346384048461914, 0.043342689514160156, 0.043243518829345705, 0.04353638458251953, 0.0434277458190918, 0.04330915069580078, 0.04349116897583008, 0.04358486557006836, 0.04325049591064453, 0.04328857421875, 0.04333318328857422, 0.0433422737121582, 0.04335747146606445, 0.04342227172851563, 0.04360153579711914, 0.043510208129882814, 0.043380767822265624, 0.04488710403442383, 0.04327110290527344, 0.042850303649902347, 0.042548225402832034, 0.04247859191894531, 0.04238336181640625, 0.04239353561401367, 0.04236275100708008, 0.0426416015625, 0.04258201599121094, 0.04265369415283203, 0.042684417724609375, 0.042686656951904295, 0.04278252792358399, 0.042987518310546875, 0.042864639282226565, 0.04295859146118164, 0.043030784606933596, 0.04296307373046875, 0.04297488021850586, 0.04280137634277344, 0.042889217376708984, 0.04426137542724609, 0.04341958236694336, 0.043081470489501957, 0.04308367919921875, 0.043136577606201175, 0.04305929565429688, 0.04302684783935547, 0.04291004943847656, 0.042866622924804684, 0.04299987030029297, 0.04358873748779297, 0.043159839630126956, 0.04296969604492187, 0.042807296752929686, 0.04290150451660156, 0.04319174575805664, 0.04342022323608399, 0.04346268844604492, 0.04341347122192383, 0.04322099304199219, 0.04323328018188476, 0.04333539199829101, 0.04333596801757812, 0.04331488037109375, 0.04328851318359375, 0.043481472015380856, 0.043597599029541016, 0.04347926330566406, 0.04372480010986328, 0.04339900970458985, 0.043315361022949216, 0.0431484489440918, 0.04346147155761719, 0.04319382476806641, 0.043231777191162106, 0.043222625732421874, 0.0432275505065918, 0.04334121704101562, 0.04356771087646484, 0.043627967834472654, 0.04376361465454102, 0.04480051040649414, 0.04333766555786133, 0.0427786865234375, 0.042608383178710935, 0.042668289184570315, 0.04263740921020508, 0.042665889739990234, 0.042724769592285154, 0.04272598266601563, 0.043190078735351564, 0.04269456100463867, 0.04280972671508789, 0.042796958923339845, 0.042737598419189456, 0.04276611328125, 0.0430533447265625, 0.043063297271728515, 0.042979328155517575, 0.04313497543334961, 0.04318956756591797, 0.043420352935791016, 0.04355072021484375, 0.04340531158447265, 0.04334796905517578, 0.04318172836303711, 0.04331145477294922, 0.04328857421875, 0.043730945587158204, 0.04348659133911133, 0.04294924926757813, 0.04304281616210937, 0.0429936637878418, 0.042921184539794925, 0.043205406188964846, 0.04329414367675781, 0.043151039123535156, 0.042920799255371095, 0.04310540771484375, 0.04328131103515625, 0.04356464004516602, 0.043430305480957034, 0.04342704010009766, 0.0433221435546875, 0.043464702606201173, 0.04322918319702149, 0.04360332870483399, 0.04364905548095703, 0.04365167999267578, 0.043578369140625, 0.04335923385620117, 0.04338675308227539, 0.04326969528198242, 0.043405376434326175, 0.043442272186279295, 0.04346102523803711, 0.04340252685546875, 0.04338966369628906, 0.04322304153442383, 0.04345446395874023, 0.043546783447265626, 0.04338467025756836, 0.04350537490844727, 0.043701728820800784, 0.044947265625, 0.04333014297485351, 0.04287078475952148, 0.04266556930541992, 0.04267180633544922, 0.04258070373535156, 0.04272537612915039, 0.042657791137695314, 0.042583934783935545, 0.04254528045654297, 0.043055103302001956, 0.04290057754516602, 0.042947040557861325, 0.0427729263305664, 0.042755966186523435, 0.04272281646728516, 0.04265347290039063, 0.04262793731689453, 0.04307558441162109, 0.04309196853637695, 0.043052318572998044, 0.043340511322021484, 0.043655105590820316, 0.043509822845458984, 0.04333977508544922, 0.04327423858642578, 0.04306710433959961, 0.043090206146240234, 0.04318207931518555, 0.043172958374023435, 0.04300483322143555, 0.04288425445556641, 0.04306396865844726, 0.04305657577514648, 0.04334662246704102, 0.04322918319702149, 0.0432386245727539, 0.04326895904541016, 0.04335411071777344, 0.04348441696166992, 0.045204448699951175, 0.04349724960327148, 0.04374649429321289, 0.04343020629882813, 0.043536895751953124, 0.04372275161743164, 0.04366889572143555, 0.043596382141113284, 0.043655166625976564, 0.04327590560913086, 0.04330124664306641, 0.04333488082885742, 0.04358367919921875, 0.043587966918945314, 0.04341385650634766, 0.04350553512573242, 0.04362022399902344, 0.04429414367675781, 0.043606369018554685, 0.04352182388305664, 0.04348640060424805, 0.04357593536376953, 0.04379020690917969, 0.04454604721069336, 0.04319232177734375, 0.04280476760864258, 0.042585727691650394, 0.042420127868652346, 0.04225529479980469, 0.042262527465820314, 0.04246527862548828, 0.04242227172851563, 0.04231897735595703, 0.04242015838623047, 0.04282371139526367, 0.0427344970703125, 0.04262297439575195, 0.042641407012939454, 0.04299760055541992, 0.04286649703979492, 0.04266428756713867, 0.04273971176147461, 0.04291340637207031, 0.04291417694091797, 0.043253761291503906, 0.04335935974121094, 0.043322238922119144, 0.04321843338012695, 0.04301465606689453, 0.04293632125854492, 0.043028480529785154, 0.04302643203735351, 0.04298342514038086, 0.04292716979980469, 0.04282668685913086, 0.04274176025390625, 0.04284415817260742, 0.04295427322387695, 0.042938720703125, 0.04294591903686523, 0.043246337890625, 0.04317184066772461, 0.04307878494262695, 0.04289785766601562, 0.04314502334594727, 0.043391616821289065, 0.04365107345581055, 0.043461982727050784, 0.04334787368774414, 0.04328729629516601, 0.04323097610473633, 0.04325958251953125, 0.04329891204833984, 0.04360444641113281, 0.04341145706176758, 0.04347843170166016, 0.04324208068847656, 0.04334592056274414, 0.04345446395874023, 0.0435214729309082, 0.04355673599243164, 0.04333795166015625, 0.04341939163208008, 0.043541217803955076, 0.04368588638305664, 0.04368297576904297]",tokens/s,23.16152719546129,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,11118.22336,14586.6752,0.0,14191.427584,13325.783552,s,1,18.100052734375,18.100052734375,0.0,18.100052734375,18.100052734375,18.100052734375,18.100052734375,[18.100052734375],,kWh,0.00032731011127499793,3.609732983239266e-05,0.00014919789713599418,0.0005126053382433848,,MB,2054.074368,14599.258112,0.0,14191.427584,12591.007744,s,10,93.63611035156251,9.36361103515625,0.0020158699324482636,9.36380419921875,9.36573466796875,9.366105126953125,9.366401494140625,"[9.3626689453125, 9.362556640625, 9.3612734375, 9.360896484375, 9.3611318359375, 9.36565234375, 9.364939453125, 9.365037109375, 9.365478515625, 9.3664755859375]",tokens/s,27.339879779161304,kWh,0.0002731969215091669,3.0134654500767466e-05,0.00018165378421180003,0.0004849853602217344,tokens/kWh,527850.9847863393,MB,2060.218368,14599.258112,0.0,14191.427584,12934.19776,s,10,43.45631689453125,4.345631689453125,0.003924766702764901,4.344840576171874,4.35171328125,4.351717236328125,4.351720400390626,"[4.3444501953125, 4.34523095703125, 4.34166162109375, 4.340126953125, 4.34248779296875, 4.3425625, 4.34802001953125, 4.35171240234375, 4.34834326171875, 4.35172119140625]",tokens/s,14.497316961513649,kWh,0.00012696021212291729,1.4005015315773754e-05,8.450056760039989e-05,0.00022546579503909082,tokens/kWh,279421.5414762899,,s,630,43.452540962219246,0.06897228724161784,0.0005116757379802847,0.06898217391967773,0.06963376541137696,0.0697523811340332,0.07028416740417481,"[0.07021110534667968, 0.06841190338134766, 0.06775328063964844, 0.06799836730957032, 0.0680796127319336, 0.06837203216552734, 0.06838931274414063, 0.06815478515625, 0.06797523498535156, 0.06823990631103516, 0.0686204833984375, 0.06868716430664062, 0.06849177551269531, 0.06866738891601562, 0.06910313415527344, 0.06851181030273437, 0.06861020660400391, 0.06846208190917968, 0.0685739517211914, 0.06850764465332031, 0.0688493423461914, 0.06874317169189453, 0.06875481414794922, 0.06890386962890625, 0.06897241973876952, 0.06911347198486328, 0.0686556167602539, 0.06891315460205077, 0.06926950073242187, 0.06906470489501954, 0.06896633911132813, 0.0689767074584961, 0.06913024139404297, 0.06895938873291016, 0.0689303970336914, 0.06890086364746094, 0.06876096343994141, 0.06893350219726563, 0.06906259155273438, 0.0689755859375, 0.06918128204345703, 0.06931686401367188, 0.06910950469970703, 0.06900691223144531, 0.06898323059082032, 0.06907698822021484, 0.06951321411132813, 0.06929996490478516, 0.06898700714111328, 0.06901773071289062, 0.06913843536376953, 0.06928998565673829, 0.0690871353149414, 0.06985282897949219, 0.06942355346679688, 0.0694128646850586, 0.06966886138916016, 0.0696463394165039, 0.06963350677490235, 0.06954415893554687, 0.06991686248779297, 0.06968742370605469, 0.06965414428710938, 0.07080188751220703, 0.06823321533203125, 0.0677027816772461, 0.06811750030517578, 0.06774681854248046, 0.06848477172851562, 0.06858377838134766, 0.06816563415527344, 0.06829603576660156, 0.06853849792480468, 0.0689874267578125, 0.06848716735839844, 0.06823321533203125, 0.06833971405029297, 0.06899830627441406, 0.06875392150878906, 0.06843631744384765, 0.0684656982421875, 0.06835913848876952, 0.06884761810302735, 0.06873673248291015, 0.06884905242919923, 0.06856924438476562, 0.06877046203613281, 0.0688128662109375, 0.06898057556152344, 0.06881705474853515, 0.0689459228515625, 0.06911328125, 0.06860591888427735, 0.06891375732421876, 0.06860594940185546, 0.06852819061279297, 0.06892537689208984, 0.06888784027099609, 0.06910851287841797, 0.06886531066894531, 0.06934185791015625, 0.06953164672851563, 0.06935884857177735, 0.06928665924072265, 0.06925721740722657, 0.06966595458984375, 0.0692457275390625, 0.06918985748291015, 0.06931177520751954, 0.0692455062866211, 0.06962515258789062, 0.0691792984008789, 0.06921705627441406, 0.0692838363647461, 0.06964979553222657, 0.06942076873779297, 0.06925212860107421, 0.06970918273925782, 0.06998067474365234, 0.06954598236083985, 0.06941801452636719, 0.0690739517211914, 0.06934630584716797, 0.06940153503417969, 0.06919577789306641, 0.06955411529541015, 0.0703597412109375, 0.06841753387451172, 0.06867356872558594, 0.06779824066162109, 0.06808425903320313, 0.06829395294189453, 0.06835238647460938, 0.06868870544433593, 0.06826512145996094, 0.06838944244384766, 0.06848636627197266, 0.0689142074584961, 0.06865382385253906, 0.06874332427978516, 0.06898111724853516, 0.06910205078125, 0.06861619567871094, 0.06815948486328124, 0.06830054473876954, 0.06850774383544922, 0.06821699523925781, 0.06869529724121094, 0.06850153350830078, 0.0682298583984375, 0.0685998077392578, 0.06890419006347656, 0.06892435455322266, 0.06910342407226562, 0.06894544219970702, 0.06910559844970703, 0.06880719757080078, 0.06861004638671875, 0.06890850830078125, 0.06877347564697266, 0.06857823944091797, 0.06872182464599609, 0.06897135925292969, 0.06902793884277343, 0.06892124938964844, 0.06915424346923828, 0.06894445037841797, 0.06917475128173828, 0.0690898208618164, 0.06942515563964843, 0.06920915222167968, 0.06907177734375, 0.06913632202148437, 0.06896649932861328, 0.06908502197265624, 0.06924918365478516, 0.06907289886474609, 0.06903990173339844, 0.06932819366455079, 0.06927657318115234, 0.06951321411132813, 0.06953778839111328, 0.0694775390625, 0.06959804534912109, 0.06952754974365234, 0.06980812835693359, 0.06912627410888672, 0.06965638732910157, 0.06952543640136719, 0.07057859039306641, 0.0686286392211914, 0.06774153900146485, 0.06746173095703124, 0.06849740600585938, 0.06850355529785156, 0.06833974456787109, 0.06820787048339844, 0.06819843292236329, 0.06828707122802734, 0.06850160217285156, 0.06827552032470703, 0.06829119873046875, 0.0687043228149414, 0.06885078430175781, 0.06884444427490234, 0.06808499145507813, 0.06842985534667968, 0.06825244903564454, 0.06872073364257812, 0.06868889617919922, 0.06858223724365234, 0.06858751678466797, 0.06873897552490234, 0.0687096939086914, 0.06885865783691406, 0.06870832061767577, 0.06884150695800781, 0.06877536010742187, 0.06878470611572265, 0.06880802917480469, 0.0690264663696289, 0.06857341003417969, 0.06863235473632813, 0.06881622314453124, 0.06857724761962891, 0.06846125030517577, 0.06895206451416015, 0.0685869140625, 0.06914054107666015, 0.06884143829345703, 0.06930899047851563, 0.06946771240234376, 0.0691655044555664, 0.06973235321044922, 0.06957615661621094, 0.0692741470336914, 0.06934073638916016, 0.06920646667480469, 0.06895616149902344, 0.06909677124023438, 0.06906745910644531, 0.06947840118408204, 0.06985533142089843, 0.06956022644042968, 0.06929203033447266, 0.06915058898925781, 0.06965055847167968, 0.06948454284667968, 0.06959327697753906, 0.0694345932006836, 0.06952953338623047, 0.069429443359375, 0.07030985260009766, 0.06835369873046875, 0.06812054443359375, 0.0679141082763672, 0.06810332489013672, 0.06809276580810547, 0.06788902282714844, 0.06830502319335938, 0.06850559997558593, 0.06857926177978516, 0.06832032012939453, 0.06810111999511718, 0.06850252532958985, 0.06846361541748047, 0.06895158386230468, 0.06873955535888672, 0.06817791748046875, 0.06833766174316407, 0.06856108856201172, 0.0685299835205078, 0.06856294250488282, 0.06861209869384766, 0.06871858978271485, 0.06874931335449219, 0.06887375640869141, 0.06889724731445312, 0.0688864974975586, 0.06905228424072266, 0.06922255706787109, 0.06915891265869141, 0.06903132629394532, 0.06884822082519532, 0.06872176361083984, 0.06884854125976562, 0.06886003112792968, 0.06887615966796876, 0.06925721740722657, 0.06862384033203126, 0.06892598724365234, 0.06913142395019531, 0.06896428680419922, 0.06984591674804687, 0.0690723876953125, 0.06904640197753906, 0.06972783660888672, 0.06924313354492187, 0.06898537445068359, 0.06887379455566406, 0.06911001586914063, 0.06910988616943359, 0.06974674987792968, 0.06918962860107422, 0.06917324829101562, 0.06945760345458984, 0.06946409606933594, 0.06958735656738281, 0.06944140625, 0.06990009307861328, 0.06947449493408203, 0.06939443206787109, 0.06966230773925781, 0.06949417877197266, 0.06946841430664062, 0.07042867279052735, 0.06888768005371093, 0.06800678253173828, 0.06775398254394531, 0.06796288299560548, 0.06825574493408203, 0.0681553955078125, 0.06809731292724609, 0.06822064208984376, 0.06819737243652343, 0.0685465316772461, 0.06860598754882813, 0.06852995300292969, 0.06886844635009766, 0.06909085083007813, 0.06886640167236328, 0.06850355529785156, 0.06849110412597656, 0.06804291534423829, 0.06820044708251953, 0.06872418975830077, 0.06877964782714843, 0.06847135925292969, 0.06830850982666016, 0.06855709075927735, 0.06860829162597656, 0.06883968353271484, 0.06952896118164062, 0.06911564636230469, 0.06909426879882813, 0.06904630279541016, 0.06911318206787109, 0.06903250885009765, 0.06910572814941406, 0.06905651092529297, 0.06884307098388671, 0.06913581085205078, 0.06882406616210937, 0.06871046447753906, 0.06879225921630859, 0.06918931579589843, 0.06918994903564453, 0.06942931365966797, 0.06964640045166015, 0.06911325073242187, 0.06958879852294922, 0.06911020660400391, 0.06893961334228516, 0.06912783813476563, 0.0694873275756836, 0.06909542083740235, 0.06898438262939453, 0.06900166320800781, 0.06928089904785156, 0.0692580795288086, 0.06976924896240234, 0.06927359771728515, 0.06975698852539063, 0.06952076721191407, 0.0694400634765625, 0.06964617919921876, 0.06942121887207031, 0.06955792236328125, 0.07022128295898437, 0.0684867172241211, 0.06797583770751953, 0.06842195129394531, 0.06780313873291016, 0.06829017639160156, 0.06851519775390626, 0.06856396484375, 0.06842777252197266, 0.068421630859375, 0.06841958618164062, 0.06885190582275391, 0.0686316146850586, 0.06890982055664062, 0.06894319915771484, 0.06886688232421875, 0.06867279815673828, 0.06875398254394531, 0.06842918395996093, 0.06866188812255859, 0.06880665588378906, 0.06863257598876953, 0.06909283447265625, 0.06856246185302735, 0.06887731170654297, 0.0692061767578125, 0.06915833282470703, 0.06968748474121093, 0.07014012908935546, 0.0693043212890625, 0.06922361755371094, 0.06905017852783203, 0.06895942687988281, 0.06898668670654297, 0.0686981430053711, 0.06908515167236329, 0.06880067443847657, 0.06921337890625, 0.06908195495605468, 0.06903788757324218, 0.06902579498291016, 0.06924012756347656, 0.0691739501953125, 0.06919923400878907, 0.06918931579589843, 0.06912095642089844, 0.06903545379638672, 0.06889734649658204, 0.06907891082763672, 0.06904025268554688, 0.06906896209716797, 0.06967836761474609, 0.06938066864013671, 0.06935142517089844, 0.06953164672851563, 0.06949807739257813, 0.06932947540283203, 0.0695031967163086, 0.06963609313964844, 0.06943743896484375, 0.06937395477294922, 0.06929203033447266, 0.06972313690185547, 0.07016454315185547, 0.0684416961669922, 0.06782768249511718, 0.06789715576171874, 0.06831948852539063, 0.06851417541503907, 0.0684051513671875, 0.06845244598388672, 0.06828844451904297, 0.06853564453125, 0.06841027069091797, 0.06909113311767578, 0.06915814208984375, 0.06861305236816406, 0.06879750061035156, 0.06869888305664062, 0.06852607727050782, 0.06850895690917969, 0.06876448059082031, 0.06841744232177735, 0.06891891479492188, 0.06907862091064453, 0.06954035186767578, 0.06872502136230468, 0.06886188507080078, 0.06905657958984375, 0.06931401824951172, 0.0695588150024414, 0.06910361480712891, 0.06912204742431641, 0.06896038055419922, 0.06886182403564453, 0.06875564575195313, 0.06867273712158203, 0.06865776062011719, 0.0690708465576172, 0.06932044982910156, 0.0691673583984375, 0.06957465362548829, 0.0693595199584961, 0.06956764984130859, 0.06934374237060546, 0.0692350082397461, 0.06972633361816406, 0.06918553924560547, 0.06902169799804687, 0.0693259506225586, 0.06924527740478516, 0.0689299545288086, 0.06956249237060547, 0.06966041564941407, 0.06981561279296875, 0.06970368194580077, 0.06941289520263672, 0.06951744079589844, 0.06974646759033203, 0.06946713256835937, 0.0696269760131836, 0.06936863708496094, 0.06951741027832031, 0.0697069091796875, 0.06936252593994141, 0.06968524932861328, 0.07047212982177735, 0.06828313446044922, 0.06796463775634766, 0.06796192169189454, 0.06801708984375, 0.06823101043701171, 0.06837398529052735, 0.06819296264648438, 0.06838377380371094, 0.06891209411621094, 0.06865833282470703, 0.06862659454345703, 0.068797119140625, 0.06901747131347656, 0.06891327667236329, 0.06890032196044922, 0.06865360260009766, 0.06844342041015625, 0.06851248168945312, 0.06849238586425781, 0.06897452545166016, 0.06883017730712891, 0.06874652862548829, 0.06871113586425781, 0.06875341033935548, 0.06899712371826172, 0.06908108520507812, 0.06899107360839844, 0.0692325439453125, 0.06890444946289062, 0.06878189086914062, 0.06865376281738281, 0.06895410919189453, 0.06863660430908203, 0.06863468933105468, 0.06901760101318359, 0.06916460418701172, 0.06903759765625, 0.0691223373413086, 0.06908992004394532, 0.0694824981689453, 0.06992009735107421, 0.06916547393798828, 0.06984320068359375, 0.0693411865234375, 0.06903193664550782, 0.06907698822021484, 0.06899097442626953, 0.06911727905273438, 0.06936377716064453, 0.0695591049194336, 0.06951283264160156, 0.06950313568115235, 0.0695050277709961, 0.06921968078613282, 0.0692804183959961, 0.06972211456298828, 0.06980406188964844, 0.06964425659179688, 0.06967862701416015, 0.06977584075927734, 0.06936716461181641, 0.06987449645996094, 0.07034265899658203, 0.06847257232666015, 0.06819859313964843, 0.06797727966308593, 0.06829875183105469, 0.06802227020263672, 0.0681426544189453, 0.06834585571289062, 0.06838886260986328, 0.06872077178955079, 0.06871183776855469, 0.06922128295898437, 0.06903561401367188, 0.06907433319091796, 0.06927632141113281, 0.06876399993896484, 0.0686173095703125, 0.06825667572021485, 0.06835199737548828, 0.06868396759033203, 0.06851673889160156, 0.06852294158935547, 0.06880662536621093, 0.06869612884521484, 0.06882275390625, 0.06869427490234375, 0.068951904296875, 0.06926966094970703, 0.06922000122070313, 0.06919190216064453, 0.0689644775390625, 0.068780029296875, 0.06896575927734375, 0.06906237030029297, 0.06891567993164062, 0.06895014190673829, 0.06907526397705079, 0.0689560317993164, 0.06913193511962891, 0.0691204833984375, 0.0694901123046875, 0.06949945831298829, 0.06946320343017579, 0.06964514923095703, 0.06928998565673829, 0.06939190673828124, 0.06922879791259766, 0.06968329620361328, 0.06908489227294921, 0.06962422180175781, 0.06953984069824219, 0.06937948608398438, 0.06911622619628906, 0.06945410919189453, 0.06965042877197265, 0.0697786865234375, 0.07013616180419922, 0.06978924560546874, 0.06944649505615234, 0.07017472076416016, 0.06986252593994141, 0.06921884918212891, 0.06978173065185547]",tokens/s,14.498576747163469,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.747328,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3549.069312,5170.46272,0.0,4775.215104,4427.072512,s,1,11.134876953125,11.134876953125,0.0,11.134876953125,11.134876953125,11.134876953125,11.134876953125,[11.134876953125],,kWh,0.0001169266884624998,1.2880390238000146e-05,5.117865205399915e-05,0.0001809857307544991,,MB,1478.483968,5201.92,0.0,4794.089472,4101.02016,s,10,24.864014404296874,2.4864014404296872,0.006186620970669977,2.4875468749999996,2.493498486328125,2.494433203125,2.4951809765625,"[2.4734931640625, 2.48464794921875, 2.479156494140625, 2.483735107421875, 2.486490966796875, 2.488602783203125, 2.4898173828125, 2.493290771484375, 2.489411865234375, 2.495367919921875]",tokens/s,102.96004331294121,kWh,7.252988282750001e-05,7.999841356979298e-06,4.8107510708200134e-05,0.00012863723489267944,tokens/kWh,1990092.528136024,MB,1487.233024,5212.40576,0.0,4802.47808,4101.02272,s,10,19.07873876953125,1.907873876953125,0.008483213307707916,1.9114246826171875,1.9149196533203126,1.9190421630859376,1.9223401708984376,"[1.900708984375, 1.898120361328125, 1.9130579833984376, 1.898367431640625, 1.9231646728515626, 1.910844970703125, 1.912362060546875, 1.91200439453125, 1.9140035400390625, 1.8961043701171876]",tokens/s,33.021050689477974,kWh,5.582348417041655e-05,6.157778698774924e-06,3.662208485319986e-05,9.860334772239133e-05,tokens/kWh,638923.5401760467,,s,630,19.0762770271301,0.030279804804968453,0.0004916393253892293,0.030159024238586424,0.03075938320159912,0.031126678657531735,0.0319873175239563,"[0.03118275260925293, 0.030290048599243165, 0.029965568542480468, 0.029962879180908203, 0.029896480560302734, 0.0297322883605957, 0.02972092819213867, 0.029856224060058594, 0.030783103942871093, 0.030079456329345704, 0.02995609664916992, 0.030874975204467775, 0.03004483222961426, 0.029908447265625, 0.030049055099487305, 0.030252063751220703, 0.03052345657348633, 0.029972511291503905, 0.029966976165771483, 0.030047967910766603, 0.03025129508972168, 0.030443168640136718, 0.0299400634765625, 0.029998912811279296, 0.030236032485961913, 0.030284608840942383, 0.03020185661315918, 0.030269311904907226, 0.029902336120605468, 0.03001612854003906, 0.030105600357055663, 0.030026815414428712, 0.03006355285644531, 0.03007600021362305, 0.030230592727661133, 0.0302191047668457, 0.03032815933227539, 0.030206111907958983, 0.029958656311035156, 0.029841407775878907, 0.030095359802246095, 0.030085119247436523, 0.030164255142211913, 0.030024415969848634, 0.03083180809020996, 0.03008515167236328, 0.0299517765045166, 0.03018649673461914, 0.030217279434204103, 0.030759872436523436, 0.030004671096801758, 0.030077280044555663, 0.030180864334106446, 0.03034819221496582, 0.03068854331970215, 0.03002566337585449, 0.030005855560302733, 0.0299683837890625, 0.030287551879882812, 0.030384096145629882, 0.03037628746032715, 0.029998336791992188, 0.030053119659423828, 0.031326623916625974, 0.03032678413391113, 0.03039580726623535, 0.0313450870513916, 0.03052115249633789, 0.030611616134643554, 0.03101750373840332, 0.03024460792541504, 0.030607263565063478, 0.029818815231323244, 0.02982419204711914, 0.0299299201965332, 0.029815231323242188, 0.029851648330688478, 0.0298242244720459, 0.029659936904907227, 0.029635583877563477, 0.029766847610473633, 0.029794111251831054, 0.029714208602905273, 0.029786336898803712, 0.029812736511230467, 0.029865983963012696, 0.029915136337280275, 0.030074495315551758, 0.029878656387329103, 0.02975129508972168, 0.03189555168151856, 0.03079782485961914, 0.0302379207611084, 0.030118688583374024, 0.02993961524963379, 0.030041887283325196, 0.030128448486328126, 0.030121984481811522, 0.030187135696411134, 0.030797216415405275, 0.029932512283325195, 0.030064704895019532, 0.0299355525970459, 0.03012403106689453, 0.02992076873779297, 0.030296575546264647, 0.030094879150390625, 0.030255584716796874, 0.030132095336914064, 0.030097440719604494, 0.03021219253540039, 0.029968351364135743, 0.029863967895507812, 0.029898752212524415, 0.029865983963012696, 0.030060543060302734, 0.029952287673950195, 0.029918336868286134, 0.02996463966369629, 0.030035295486450196, 0.030157728195190428, 0.030103744506835936, 0.029941503524780273, 0.029757728576660158, 0.029828895568847658, 0.03011327934265137, 0.031227647781372072, 0.030310848236083984, 0.030153247833251955, 0.030007295608520508, 0.02998681640625, 0.030015359878540038, 0.03012620735168457, 0.030289215087890627, 0.03022879981994629, 0.03007036781311035, 0.030109920501708985, 0.03017375946044922, 0.03028096008300781, 0.030051071166992186, 0.030096799850463866, 0.03013078308105469, 0.03060736083984375, 0.030357503890991212, 0.032513534545898434, 0.03163539123535156, 0.030529632568359374, 0.030237152099609376, 0.030220287322998047, 0.030826496124267577, 0.030502912521362304, 0.030513151168823242, 0.030404575347900392, 0.030572608947753908, 0.030316511154174806, 0.03065763282775879, 0.030036895751953126, 0.030215648651123046, 0.030579231262207032, 0.03047350311279297, 0.03047212791442871, 0.030332927703857423, 0.030227231979370116, 0.029987903594970704, 0.030143423080444334, 0.030064640045166017, 0.030371423721313476, 0.030476703643798828, 0.030294015884399415, 0.030316543579101563, 0.030125247955322267, 0.030237344741821288, 0.03013372802734375, 0.030759328842163085, 0.030124223709106446, 0.030658655166625977, 0.030074367523193358, 0.030329343795776367, 0.03096531105041504, 0.030101024627685546, 0.030111743927001954, 0.02994374465942383, 0.030487295150756835, 0.030129600524902343, 0.030479135513305663, 0.030472192764282226, 0.030089311599731446, 0.02993142318725586, 0.030496768951416016, 0.031055200576782228, 0.03154323196411133, 0.031135520935058593, 0.03036390495300293, 0.03098076820373535, 0.029937664031982423, 0.030672895431518556, 0.029603647232055663, 0.02979612731933594, 0.030015167236328126, 0.030141151428222657, 0.03018454360961914, 0.02992736053466797, 0.030153696060180663, 0.03015475273132324, 0.029845279693603517, 0.03030431938171387, 0.02987414360046387, 0.029787424087524415, 0.029746080398559572, 0.02985683250427246, 0.029958272933959963, 0.029963071823120118, 0.03017728042602539, 0.03005788803100586, 0.029782623291015626, 0.029783615112304686, 0.029804895401000977, 0.02967475128173828, 0.029895200729370117, 0.029647167205810548, 0.02982636833190918, 0.0298110408782959, 0.02986969566345215, 0.029938400268554686, 0.029872127532958984, 0.02996553611755371, 0.029878528594970703, 0.029858591079711914, 0.030678783416748047, 0.02996633529663086, 0.030422496795654296, 0.03012784004211426, 0.03014534378051758, 0.030713855743408205, 0.030059776306152343, 0.03019340705871582, 0.030465024948120117, 0.030461151123046874, 0.031052576065063477, 0.03061942481994629, 0.03006835174560547, 0.030024288177490234, 0.030166175842285155, 0.02999795150756836, 0.03002774429321289, 0.030085119247436523, 0.029859840393066408, 0.029954048156738283, 0.030097503662109375, 0.029990079879760743, 0.029979360580444335, 0.03012380790710449, 0.03131366348266602, 0.030595359802246095, 0.030568544387817382, 0.031039264678955077, 0.03018569564819336, 0.030400800704956054, 0.030321727752685546, 0.030360223770141602, 0.030529727935791017, 0.0304881591796875, 0.030312671661376953, 0.030357503890991212, 0.0301527042388916, 0.030107519149780274, 0.030236799240112303, 0.03031449508666992, 0.030517248153686522, 0.030418304443359374, 0.030871328353881837, 0.03060393524169922, 0.03046406364440918, 0.030499967575073242, 0.030286848068237306, 0.030672895431518556, 0.030361600875854492, 0.030304256439208983, 0.031115264892578126, 0.030277151107788086, 0.03024358367919922, 0.03037686347961426, 0.031679103851318356, 0.030432512283325195, 0.030524351119995116, 0.030564031600952148, 0.03062812805175781, 0.030576864242553712, 0.030873407363891603, 0.030199968338012695, 0.030474079132080077, 0.030498336791992188, 0.03060960006713867, 0.030579168319702147, 0.030621503829956053, 0.03037183952331543, 0.030225824356079102, 0.030336736679077148, 0.030358400344848633, 0.030433055877685546, 0.030347488403320313, 0.03020185661315918, 0.030377695083618164, 0.03027587127685547, 0.030398464202880858, 0.03042508888244629, 0.03055177688598633, 0.030361888885498046, 0.030629888534545898, 0.03083247947692871, 0.031848608016967775, 0.031364896774291995, 0.030402496337890626, 0.030230079650878906, 0.030383968353271486, 0.031189311981201173, 0.030156991958618165, 0.03032678413391113, 0.03061350440979004, 0.030257152557373046, 0.03014656066894531, 0.029969919204711915, 0.02996403121948242, 0.0298666877746582, 0.030340768814086913, 0.030529951095581053, 0.030345216751098632, 0.030345216751098632, 0.035870113372802735, 0.030715999603271486, 0.03031091117858887, 0.030302207946777345, 0.03057459259033203, 0.030212095260620117, 0.02992742347717285, 0.029900800704956054, 0.030654111862182618, 0.03055414390563965, 0.030027423858642578, 0.029964960098266602, 0.030010719299316407, 0.02998054313659668, 0.029995807647705076, 0.02985955238342285, 0.03004140853881836, 0.029759552001953123, 0.029856672286987306, 0.029945856094360353, 0.030023679733276368, 0.03000307273864746, 0.030062719345092772, 0.030175296783447266, 0.029880256652832032, 0.029888511657714844, 0.029961919784545897, 0.029960512161254883, 0.0315163516998291, 0.0322861442565918, 0.030823455810546876, 0.03165782356262207, 0.030914560317993164, 0.030268863677978517, 0.030254976272583008, 0.02994041633605957, 0.030136320114135744, 0.03035545539855957, 0.030084352493286132, 0.03004275131225586, 0.029862016677856446, 0.030057695388793944, 0.03007551956176758, 0.030159008026123046, 0.029920671463012697, 0.02984934425354004, 0.02995465660095215, 0.02991948890686035, 0.029915136337280275, 0.030136320114135744, 0.032659488677978514, 0.031340927124023435, 0.03023174476623535, 0.030105535507202148, 0.030288799285888672, 0.03019491195678711, 0.03167654418945313, 0.03202479934692383, 0.030420576095581055, 0.030685375213623047, 0.030236671447753907, 0.03004422378540039, 0.030101728439331055, 0.030271936416625976, 0.030159040451049803, 0.029978368759155275, 0.030115840911865234, 0.03009903907775879, 0.030325151443481444, 0.03019980812072754, 0.030019584655761718, 0.030064640045166017, 0.02991676712036133, 0.02998121643066406, 0.029913248062133788, 0.02990617561340332, 0.03006208038330078, 0.030230752944946288, 0.030148927688598632, 0.030337472915649415, 0.030138368606567382, 0.030091264724731445, 0.030946495056152344, 0.03021232032775879, 0.029988895416259764, 0.030113632202148438, 0.03003670310974121, 0.030512895584106445, 0.03040246391296387, 0.030015487670898438, 0.02999888038635254, 0.030187295913696288, 0.030891904830932616, 0.03023072052001953, 0.030157535552978516, 0.03017932891845703, 0.030140384674072266, 0.030105632781982423, 0.030417055130004884, 0.03234537506103516, 0.030132863998413088, 0.030177120208740235, 0.030047359466552733, 0.0301496639251709, 0.030097343444824218, 0.03046428871154785, 0.030248031616210938, 0.030246591567993163, 0.030389184951782226, 0.03057254409790039, 0.03016908836364746, 0.03020595169067383, 0.030356767654418946, 0.03137945556640625, 0.030498144149780273, 0.03064694404602051, 0.030251007080078125, 0.029973695755004883, 0.030180160522460937, 0.030212095260620117, 0.03013804817199707, 0.03025279998779297, 0.030349599838256837, 0.0302840633392334, 0.030231903076171875, 0.030517919540405274, 0.03044870376586914, 0.030538688659667967, 0.030451072692871093, 0.030654207229614257, 0.0305611515045166, 0.030542879104614257, 0.031648223876953124, 0.030432031631469725, 0.030320159912109373, 0.030144704818725585, 0.030442527770996094, 0.030151647567749025, 0.030926847457885744, 0.030371807098388673, 0.030326816558837892, 0.029904895782470704, 0.030668256759643554, 0.030046880722045897, 0.030317983627319335, 0.030402975082397463, 0.030586944580078126, 0.030509056091308592, 0.030791519165039062, 0.030150815963745116, 0.030040063858032227, 0.029982751846313476, 0.030105728149414063, 0.029960031509399413, 0.029997055053710937, 0.03003171157836914, 0.029881759643554686, 0.029889280319213868, 0.029951263427734375, 0.029952608108520507, 0.030344863891601563, 0.03030428886413574, 0.03057708740234375, 0.030414848327636718, 0.030217632293701172, 0.03111587142944336, 0.030265344619750976, 0.02997657585144043, 0.030019584655761718, 0.030382272720336913, 0.029994943618774413, 0.030302080154418945, 0.02996633529663086, 0.030095104217529298, 0.030673152923583986, 0.031064064025878906, 0.03117846488952637, 0.030828639984130858, 0.03023388862609863, 0.030519327163696288, 0.03161343955993652, 0.030970239639282228, 0.030910463333129884, 0.02997452735900879, 0.02993187141418457, 0.030083999633789063, 0.0304586238861084, 0.030121984481811522, 0.03045097541809082, 0.030014175415039063, 0.030064287185668944, 0.029826719284057616, 0.030579391479492186, 0.029926591873168946, 0.029973152160644532, 0.030042272567749023, 0.030115840911865234, 0.03021731185913086, 0.03017206382751465, 0.030500864028930662, 0.030240768432617186, 0.030311552047729493, 0.030468992233276367, 0.03021824073791504, 0.03013222312927246, 0.030129215240478516, 0.030071584701538086, 0.029997215270996094, 0.030283456802368165, 0.030291839599609374, 0.030523839950561522, 0.03055366325378418, 0.031019456863403322, 0.030348352432250977, 0.030104095458984376, 0.03130780792236328, 0.03015513610839844, 0.03000044822692871, 0.030132640838623048, 0.030602527618408204, 0.03004102325439453, 0.03158227157592773, 0.033490398406982425, 0.03049907112121582, 0.030274879455566405, 0.030796384811401366, 0.03021049690246582, 0.030097471237182618, 0.030192991256713868, 0.030234880447387695, 0.030222623825073243, 0.03018547248840332, 0.03014041519165039, 0.030033567428588866, 0.02994620704650879, 0.029941568374633788, 0.029872608184814454, 0.03007663917541504, 0.030304256439208983, 0.031194047927856447, 0.03054591941833496, 0.030649375915527344, 0.030190559387207033, 0.029976480484008788, 0.029912256240844728, 0.029897920608520506, 0.029848575592041016, 0.029843456268310548, 0.030183935165405275, 0.030157024383544923, 0.030410751342773438, 0.0304005126953125, 0.03021414375305176, 0.030066879272460937, 0.029839168548583983, 0.029880319595336914, 0.02977177619934082, 0.029876224517822264, 0.029915136337280275, 0.029879423141479493, 0.0300184326171875, 0.029863935470581054, 0.030482175827026368, 0.031486207962036133, 0.030981727600097656, 0.02997699165344238, 0.02998681640625, 0.029890560150146486, 0.03010508728027344, 0.03021830368041992, 0.02991289520263672, 0.030144512176513674, 0.03009382438659668, 0.030133760452270508, 0.03045030403137207, 0.02996428871154785, 0.029847488403320313, 0.029773504257202148, 0.02981443214416504, 0.029795040130615236, 0.029779104232788085, 0.02990480041503906, 0.030088127136230467, 0.030201759338378906, 0.030062143325805663, 0.02987676811218262, 0.030035423278808593, 0.029864479064941406, 0.029992704391479493, 0.029956287384033203, 0.030137727737426758, 0.03035615921020508, 0.030052127838134764, 0.029911264419555664, 0.029904895782470704, 0.029960224151611328, 0.029933183670043946, 0.029892927169799806, 0.030652448654174803, 0.02979020881652832, 0.029921119689941406, 0.030001312255859374]",tokens/s,33.025311967530094,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4595.003392,7907.24608,0.0,7511.998464,6895.682048,s,1,12.65923046875,12.65923046875,0.0,12.65923046875,12.65923046875,12.65923046875,12.65923046875,[12.65923046875],,kWh,0.00016308059054586011,1.798142303511062e-05,6.664338664802427e-05,0.000247705400228995,,MB,2460.782592,7921.926144,0.0,7514.095616,6822.139904,s,10,29.046709716796876,2.9046709716796877,0.0033588276174065443,2.9054093017578126,2.9084703857421874,2.9084865356445313,2.9084994555664063,"[2.89770849609375, 2.900326416015625, 2.9035615234375, 2.90337646484375, 2.904469970703125, 2.9063486328125, 2.908466796875, 2.90701220703125, 2.908502685546875, 2.9069365234375]",tokens/s,88.13390655808516,kWh,8.473676948625477e-05,9.34597383415778e-06,5.638348955120387e-05,0.0001504662328716164,tokens/kWh,1701378.4097221938,MB,2465.046528,7926.120448,0.0,7516.192768,6822.142464,s,10,21.404966064453124,2.1404966064453124,0.02118520790531577,2.1430899658203124,2.1637234375000003,2.1649804931640624,2.1659861376953122,"[2.11697216796875, 2.106642578125, 2.12462939453125, 2.160172119140625, 2.15317333984375, 2.166237548828125, 2.159627685546875, 2.121060546875, 2.163444091796875, 2.133006591796875]",tokens/s,29.4324222754191,kWh,6.286705930166284e-05,6.935121888016265e-06,4.151392209999494e-05,0.00011131610328967404,tokens/kWh,565955.8512936558,,s,630,21.402253734588616,0.033971831324743844,0.0005589987614556571,0.034013088226318364,0.03447806396484375,0.034646830940246585,0.035267693634033204,"[0.03500239944458008, 0.03394144058227539, 0.03366620635986328, 0.03341747283935547, 0.033616416931152346, 0.03364054489135742, 0.03351551818847656, 0.033586849212646486, 0.03360726547241211, 0.03352576065063476, 0.034168865203857424, 0.03337862396240234, 0.03338668823242188, 0.033340705871582034, 0.033645503997802736, 0.03369574356079102, 0.0336363525390625, 0.033441150665283206, 0.03334563064575195, 0.03340547180175781, 0.03340185546875, 0.03329945755004883, 0.03320832061767578, 0.033173503875732424, 0.03340288162231445, 0.033323009490966796, 0.03328204727172852, 0.03341519927978515, 0.033456096649169924, 0.033484798431396484, 0.034079936981201174, 0.033421535491943356, 0.033280609130859375, 0.03350223922729492, 0.0333076171875, 0.03377897644042969, 0.03347324752807617, 0.0335810546875, 0.03337625503540039, 0.03459619140625, 0.03340118408203125, 0.03331439971923828, 0.03341107177734375, 0.03323065567016602, 0.033530208587646486, 0.033605728149414066, 0.03351542282104492, 0.03338246536254883, 0.033388641357421874, 0.03365478515625, 0.03358761596679687, 0.03325360107421875, 0.033255199432373046, 0.033560192108154294, 0.03408268737792969, 0.03397011184692383, 0.03395593643188476, 0.03388991928100586, 0.03411235046386719, 0.03408224105834961, 0.034027584075927736, 0.03391670227050781, 0.033790687561035156, 0.034800510406494144, 0.03368255996704102, 0.03335667037963867, 0.033298431396484376, 0.03364195251464844, 0.033245471954345705, 0.033198558807373046, 0.03317465591430664, 0.033813152313232425, 0.03335145568847656, 0.033234622955322264, 0.0331822395324707, 0.03349094390869141, 0.03321446228027344, 0.03327366256713867, 0.033530048370361325, 0.033298431396484376, 0.03320611190795898, 0.03327606582641601, 0.0332674560546875, 0.03336832046508789, 0.033308734893798826, 0.03336150360107422, 0.03311840057373047, 0.033495201110839846, 0.03314412689208984, 0.033077953338623046, 0.0331960334777832, 0.033142784118652346, 0.03322265625, 0.033083393096923826, 0.033157119750976564, 0.03315017700195313, 0.03301011276245117, 0.033028030395507814, 0.03307347106933594, 0.03324937438964844, 0.033658878326416015, 0.03389158248901367, 0.03388083267211914, 0.03379715347290039, 0.033745697021484375, 0.033720512390136716, 0.033866817474365235, 0.03376022338867188, 0.033734622955322265, 0.03421388626098633, 0.03343065643310547, 0.03345878219604492, 0.03365827178955078, 0.03370470428466797, 0.03340505599975586, 0.033853214263916014, 0.033265888214111326, 0.033337249755859374, 0.033314910888671875, 0.033355777740478515, 0.033232288360595705, 0.033265758514404296, 0.03398015975952148, 0.03367808151245117, 0.03344384002685547, 0.03303807830810547, 0.034748382568359375, 0.033763263702392576, 0.033626590728759766, 0.033559776306152346, 0.03338729476928711, 0.03365228652954102, 0.03406480026245117, 0.033740833282470704, 0.03383500671386719, 0.03405171203613281, 0.03426342391967773, 0.034108638763427734, 0.03425769424438477, 0.03384921646118164, 0.03382476806640625, 0.03382284927368164, 0.0339865608215332, 0.0338034553527832, 0.03413423919677734, 0.03358371353149414, 0.03375513458251953, 0.0335994873046875, 0.03350515365600586, 0.03337638473510742, 0.03360153579711914, 0.03346825790405274, 0.033829025268554684, 0.03375667190551758, 0.0338600959777832, 0.033849342346191406, 0.033392929077148435, 0.03341299057006836, 0.0332770881652832, 0.03326393508911133, 0.03324720001220703, 0.03319030380249023, 0.033339393615722655, 0.0333496322631836, 0.03418931198120117, 0.03362166213989258, 0.03379808044433594, 0.03333161544799805, 0.033468414306640625, 0.0334881591796875, 0.033321567535400394, 0.033546302795410154, 0.033475936889648436, 0.03345072174072266, 0.03414425659179687, 0.03382681655883789, 0.033753089904785157, 0.03386368179321289, 0.03379404830932617, 0.03388809585571289, 0.03403174209594727, 0.03401299285888672, 0.03385718536376953, 0.03389702224731445, 0.03395993423461914, 0.033678558349609374, 0.03373136138916016, 0.03354982376098633, 0.03354201507568359, 0.035786014556884765, 0.03423100662231445, 0.03401657485961914, 0.03455244827270508, 0.03406649780273437, 0.03456800079345703, 0.0344384651184082, 0.03427587127685547, 0.03409340667724609, 0.03433420944213867, 0.034294143676757816, 0.034574462890625, 0.03500236892700195, 0.0343548469543457, 0.03414255905151367, 0.03418316650390625, 0.03416883087158203, 0.033957889556884766, 0.033973758697509765, 0.03408537673950195, 0.03388838577270508, 0.03391068649291992, 0.03388412857055664, 0.0340206413269043, 0.033946334838867186, 0.03417292785644531, 0.033947647094726564, 0.034136062622070314, 0.03417497634887695, 0.03402956771850586, 0.03401318359375, 0.03425075149536133, 0.0341044807434082, 0.03409187316894531, 0.03421388626098633, 0.034228446960449216, 0.03438979339599609, 0.03433062362670898, 0.0344268798828125, 0.03434451293945313, 0.03440796661376953, 0.03431673431396484, 0.034689502716064455, 0.034236415863037106, 0.034103294372558594, 0.034275169372558596, 0.034412254333496095, 0.03442678451538086, 0.03449091339111328, 0.03430611038208008, 0.03431612777709961, 0.034340190887451175, 0.03488790512084961, 0.03427824020385742, 0.034246047973632815, 0.03431248092651367, 0.03440822219848633, 0.03461539077758789, 0.034121856689453126, 0.03424611282348633, 0.03414435195922851, 0.03420182418823242, 0.034466014862060544, 0.03509657669067383, 0.03429580688476563, 0.0342852783203125, 0.03396636962890625, 0.0339947509765625, 0.0340555534362793, 0.03386841583251953, 0.0337367057800293, 0.033898239135742185, 0.03400518417358398, 0.03394566345214844, 0.0344637451171875, 0.034277374267578126, 0.034097152709960936, 0.03411372756958008, 0.034033470153808594, 0.03400646209716797, 0.033965694427490235, 0.03376224136352539, 0.03387360000610352, 0.03388857650756836, 0.033828414916992185, 0.033882080078125, 0.033699745178222655, 0.0337632942199707, 0.03363849639892578, 0.03452320098876953, 0.034268703460693356, 0.03417161560058594, 0.033990081787109376, 0.034003711700439455, 0.034203647613525394, 0.03412335968017578, 0.0339890251159668, 0.0340805778503418, 0.03506563186645508, 0.034006561279296875, 0.03407478332519531, 0.03404873657226563, 0.03421990585327148, 0.034578144073486326, 0.03410163116455078, 0.03410943984985351, 0.034240543365478514, 0.03424870300292969, 0.03414425659179687, 0.034514942169189454, 0.03416223907470703, 0.03424489593505859, 0.034313377380371095, 0.03430047988891602, 0.03448428726196289, 0.03447356796264649, 0.03429024124145508, 0.034212062835693356, 0.034551807403564457, 0.034344959259033206, 0.03435696029663086, 0.03435507202148438, 0.034292129516601565, 0.034408447265625, 0.03464614486694336, 0.03427622222900391, 0.035020225524902346, 0.034320960998535155, 0.034176097869873044, 0.034210399627685545, 0.03420191955566406, 0.0341514892578125, 0.034212799072265626, 0.03452844619750976, 0.03414422225952148, 0.034810718536376954, 0.03463359832763672, 0.034238174438476564, 0.03423478317260742, 0.034204830169677736, 0.03441712188720703, 0.03424499130249024, 0.03424870300292969, 0.034307903289794925, 0.03428575897216797, 0.03427689743041992, 0.03447382354736328, 0.03422793579101562, 0.03438454437255859, 0.034253150939941406, 0.03440630340576172, 0.034309825897216796, 0.034277694702148434, 0.03420310211181641, 0.03409564971923828, 0.03422412872314453, 0.03437977600097656, 0.034549758911132815, 0.03411084747314453, 0.034284160614013674, 0.03435715103149414, 0.03432457733154297, 0.03454313659667969, 0.03427779388427735, 0.03461536026000977, 0.03445753479003906, 0.034412609100341794, 0.03445935821533203, 0.0352770881652832, 0.03451084899902344, 0.034616992950439456, 0.034385696411132816, 0.034259201049804684, 0.034355518341064456, 0.03447760009765625, 0.03484924697875977, 0.03433881759643555, 0.03432447814941406, 0.03434249496459961, 0.03473859024047852, 0.03430160140991211, 0.03440470504760742, 0.03440367889404297, 0.034390304565429686, 0.03426134490966797, 0.034231903076171875, 0.03420204925537109, 0.03428761672973633, 0.034460800170898434, 0.035250175476074216, 0.03441766357421875, 0.03496243286132812, 0.034408481597900394, 0.03512112045288086, 0.03460505676269531, 0.03437158584594727, 0.034320384979248046, 0.03424665451049805, 0.03441209411621094, 0.03436793518066406, 0.03423231887817383, 0.034231681823730466, 0.03446028900146485, 0.03427840042114258, 0.034229248046875, 0.03417292785644531, 0.034328575134277346, 0.03453952026367187, 0.0343361930847168, 0.03434092712402344, 0.034867710113525394, 0.034515071868896484, 0.03443699264526367, 0.03407417678833008, 0.03434118270874023, 0.034395870208740235, 0.03437609481811523, 0.03422003173828125, 0.034364513397216793, 0.03427612686157227, 0.03601011276245117, 0.03429897689819336, 0.034182048797607424, 0.0342999038696289, 0.03428054428100586, 0.03455683135986328, 0.03428665542602539, 0.034149375915527344, 0.03411289596557617, 0.03457622528076172, 0.03435977554321289, 0.03449472045898438, 0.03427503967285156, 0.034224414825439455, 0.03440550231933594, 0.034380672454833984, 0.03460710525512695, 0.03420774459838867, 0.03396588897705078, 0.03401110458374024, 0.03387321472167969, 0.03392399978637695, 0.03374899291992187, 0.03348876953125, 0.03364672088623047, 0.0332042236328125, 0.03327910232543945, 0.033361824035644534, 0.03328464126586914, 0.034138240814208985, 0.03478764724731445, 0.03347568130493164, 0.03527484893798828, 0.034372222900390624, 0.03425804901123047, 0.03433561706542969, 0.03430604934692383, 0.03477443313598633, 0.03472035217285156, 0.03597942352294922, 0.034928478240966794, 0.033920448303222654, 0.0339359359741211, 0.03390780639648437, 0.03376118469238281, 0.03358924865722656, 0.033459201812744144, 0.03343324661254883, 0.033487201690673825, 0.03333740615844726, 0.03357894515991211, 0.03329036712646485, 0.03357430267333984, 0.03333091354370117, 0.03328623962402344, 0.03325404739379883, 0.033247230529785156, 0.033271583557128906, 0.03329782485961914, 0.03331676864624023, 0.033395614624023434, 0.033500385284423825, 0.03346518325805664, 0.03352681732177734, 0.03366089630126953, 0.03336521530151367, 0.03365859222412109, 0.03371417617797851, 0.033355777740478515, 0.03434617614746094, 0.033365985870361325, 0.033410049438476565, 0.03376726531982422, 0.03330780792236328, 0.033129310607910155, 0.03342131042480469, 0.03334348678588867, 0.03330188751220703, 0.03348912048339844, 0.03327632141113281, 0.03312998580932617, 0.03342489624023438, 0.033842174530029294, 0.033183135986328126, 0.033269824981689455, 0.03313308715820312, 0.03321673583984375, 0.034359073638916014, 0.033562625885009766, 0.033490177154541015, 0.03346022415161133, 0.0333924789428711, 0.0333034553527832, 0.03336732864379883, 0.033638816833496094, 0.03481068801879883, 0.03409036636352539, 0.034017566680908204, 0.03417478561401367, 0.03442073440551758, 0.03409952163696289, 0.034328800201416015, 0.034295646667480466, 0.03442617416381836, 0.03410409545898437, 0.03448223876953125, 0.03413804626464844, 0.03417007827758789, 0.0343704948425293, 0.034231937408447266, 0.03426537704467773, 0.03427939224243164, 0.03417910385131836, 0.03410662460327148, 0.03482495880126953, 0.034123966217041016, 0.03419472122192383, 0.03399939346313476, 0.03383295822143555, 0.03394124984741211, 0.03405235290527344, 0.03406380844116211, 0.034056640625, 0.034081920623779294, 0.034339839935302735, 0.03422563171386719, 0.03438617706298828, 0.03414863967895508, 0.03458854293823242, 0.03447001647949219, 0.034359294891357424, 0.034285568237304685, 0.03430099105834961, 0.03416979217529297, 0.034362400054931644, 0.03420428848266602, 0.03404947280883789, 0.03400182342529297, 0.03416438293457031, 0.03437807846069336, 0.03464739227294922, 0.03425347137451172, 0.03416585540771484, 0.034460575103759765, 0.03450470352172851, 0.03418931198120117, 0.034317375183105465, 0.03456857681274414, 0.03444601440429688, 0.03466227340698242, 0.034187137603759764, 0.033963329315185545, 0.03397046279907227, 0.04120630264282227, 0.03450998306274414, 0.03369424057006836, 0.033357440948486326, 0.03351827239990234, 0.03487311935424805, 0.03375740814208984, 0.03383859252929688, 0.03356313705444336, 0.03348070526123047, 0.03325279998779297, 0.03333283233642578, 0.03326819229125977, 0.0334730224609375, 0.033500831604003904, 0.0336530876159668, 0.03349708938598633, 0.0335093765258789, 0.03338854217529297, 0.034167999267578124, 0.0334587516784668, 0.03346604919433594, 0.034011390686035155, 0.03369929504394531, 0.03356665420532227, 0.033344223022460935, 0.03343584060668945, 0.0337367057800293, 0.033533119201660154, 0.033741184234619144, 0.033837505340576175, 0.03388393783569336, 0.0338741455078125, 0.03387801742553711, 0.03371942520141601, 0.03374310302734375, 0.033954433441162106, 0.03439804840087891, 0.03408707046508789, 0.034238208770751954, 0.03411299133300781, 0.03413190460205078, 0.03420560073852539, 0.034344928741455075, 0.034227169036865235, 0.034256481170654295, 0.03434947204589844, 0.03528511810302734, 0.034396095275878905, 0.03398630523681641, 0.03406988906860352, 0.034286529541015624, 0.03405401611328125, 0.03389004898071289, 0.03397228622436523, 0.033695934295654296, 0.03362319946289063, 0.03382563018798828, 0.03387302398681641, 0.033825248718261716, 0.03382236862182617, 0.03356121444702148, 0.03360371017456055, 0.03361587142944336, 0.03378995132446289, 0.03385331344604492, 0.034087039947509765, 0.03377318572998047]",tokens/s,29.436152276890542,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,861.9008,690.946048,0.0,295.698432,277.263872,s,1,8.0635947265625,8.0635947265625,0.0,8.0635947265625,8.0635947265625,8.0635947265625,8.0635947265625,[8.0635947265625],,kWh,2.460304224586404e-05,2.7061199168236418e-06,7.568061610008225e-06,3.487722377269591e-05,,MB,1191.817216,743.374848,0.0,335.54432,313.833472,s,14,0.4949744300842286,0.03535531643458775,0.0012292449171760196,0.03517835235595704,0.035903210830688474,0.03715352516174316,0.03892500129699707,"[0.039367870330810545, 0.03465631866455078, 0.03466857528686523, 0.035565631866455075, 0.035257183074951175, 0.03448144149780273, 0.035961185455322266, 0.035099521636962894, 0.03540876770019531, 0.03440915298461914, 0.035350238800048825, 0.034116863250732425, 0.03486374282836914, 0.03576793670654297]",tokens/s,7240.778072899887,kWh,1.2312160439345533e-06,1.3578023718483657e-07,8.150780078986655e-07,2.1820742890180554e-06,tokens/kWh,117319562.07375567,MB,1225.547776,768.540672,0.0,360.710144,313.836032,s,14,10.121002380371094,0.7229287414550781,0.00635519996754101,0.7221958007812499,0.7312540771484375,0.7341511932373047,0.7368073382568359,"[0.7176253662109375, 0.7140076904296875, 0.7209071655273438, 0.72089501953125, 0.7236705932617188, 0.7178732299804688, 0.72321728515625, 0.7286656494140625, 0.7237313232421875, 0.72117431640625, 0.7139846801757812, 0.7323634033203125, 0.7374713745117187, 0.725415283203125]",tokens/s,87.1455184824945,kWh,2.0780348703982563e-05,2.2917244849225763e-06,8.175504581388059e-06,3.12475777702932e-05,tokens/kWh,2016156.2749959312,,s,882,10.113854816436769,0.011466955574191346,0.00022937992220901338,0.011431695938110352,0.01170497932434082,0.01181095232963562,0.012260950794219969,"[0.011403488159179688, 0.011644927978515626, 0.0115316162109375, 0.011442432403564453, 0.011473440170288085, 0.0114335355758667, 0.011442463874816894, 0.011468159675598144, 0.011324064254760742, 0.011296992301940917, 0.011361056327819825, 0.01130780792236328, 0.011359904289245605, 0.011339808464050293, 0.011444448471069336, 0.011673439979553222, 0.011359968185424806, 0.011336223602294922, 0.011352319717407227, 0.011333727836608886, 0.01260758399963379, 0.013354144096374511, 0.011467391967773438, 0.011323391914367676, 0.01133743953704834, 0.011254048347473145, 0.011313376426696777, 0.011245344161987305, 0.011288576126098633, 0.011302495956420898, 0.011268095970153809, 0.011286016464233398, 0.01129155158996582, 0.011314528465270996, 0.011327775955200195, 0.01144870376586914, 0.011360095977783204, 0.011386112213134765, 0.011315168380737305, 0.011229887962341308, 0.011311615943908691, 0.011306752204895019, 0.011230239868164063, 0.011289504051208496, 0.011214912414550781, 0.01126195240020752, 0.011260095596313477, 0.011236224174499512, 0.011250144004821777, 0.011237183570861816, 0.011293343544006348, 0.011284640312194824, 0.011264863967895508, 0.01123356819152832, 0.011227871894836426, 0.011220992088317871, 0.01125376033782959, 0.011231295585632324, 0.011275775909423828, 0.011335743904113769, 0.011250047683715821, 0.011274304389953613, 0.0113372802734375, 0.011330112457275391, 0.01147209644317627, 0.011428192138671875, 0.011411199569702149, 0.01130355167388916, 0.011405247688293457, 0.01128172779083252, 0.011393600463867188, 0.011323519706726074, 0.011408800125122071, 0.011369407653808594, 0.01132464027404785, 0.01136684799194336, 0.01131926441192627, 0.011272576332092285, 0.011351519584655762, 0.011419903755187989, 0.01136633586883545, 0.011435775756835938, 0.011355551719665527, 0.011346079826354981, 0.011326144218444825, 0.011337727546691894, 0.011345919609069824, 0.011345919609069824, 0.01140236759185791, 0.011309727668762207, 0.011190303802490234, 0.011291839599609376, 0.01121337604522705, 0.01121507167816162, 0.01124988842010498, 0.01120787239074707, 0.011440959930419922, 0.011239744186401368, 0.012365504264831543, 0.011552576065063477, 0.011360063552856445, 0.011300928115844726, 0.011266112327575684, 0.01124777603149414, 0.011228608131408692, 0.011326111793518067, 0.01123532772064209, 0.011172096252441406, 0.011218688011169434, 0.011181440353393555, 0.011229696273803711, 0.01124675178527832, 0.011240256309509278, 0.011159711837768555, 0.011259903907775879, 0.01129417610168457, 0.011256352424621582, 0.01129593563079834, 0.011239744186401368, 0.011225888252258301, 0.011259008407592773, 0.01123151969909668, 0.01130732822418213, 0.011308927536010743, 0.011290399551391602, 0.011388928413391113, 0.011482560157775878, 0.011681695938110352, 0.011526816368103028, 0.011593728065490723, 0.011624447822570801, 0.011556672096252442, 0.0116779842376709, 0.01159158420562744, 0.011636735916137696, 0.011517248153686523, 0.011556608200073242, 0.011480192184448242, 0.01146275234222412, 0.011404159545898438, 0.011451231956481933, 0.011536383628845214, 0.011481087684631347, 0.011493696212768554, 0.011534015655517578, 0.011474143981933594, 0.011467840194702148, 0.011396832466125488, 0.011404864311218262, 0.011401791572570801, 0.01139724826812744, 0.011450176239013671, 0.011431455612182617, 0.011364768028259278, 0.011393024444580077, 0.011313152313232423, 0.011290847778320312, 0.012256031990051269, 0.011883808135986327, 0.012088800430297851, 0.011522303581237793, 0.011390303611755371, 0.011479167938232422, 0.011391615867614746, 0.011430015563964844, 0.011533408164978028, 0.011322367668151855, 0.011304351806640625, 0.011335968017578125, 0.011358048439025879, 0.011309087753295899, 0.011308768272399902, 0.011266112327575684, 0.011313023567199708, 0.011321824073791504, 0.011324864387512207, 0.011278911590576172, 0.011325440406799316, 0.011284671783447265, 0.011266880035400391, 0.01124998378753662, 0.011247296333312989, 0.011187423706054687, 0.011220704078674317, 0.011242655754089356, 0.01119651222229004, 0.011250335693359375, 0.011190784454345704, 0.011285759925842286, 0.011575231552124024, 0.011571040153503418, 0.011489119529724121, 0.011396575927734375, 0.011297696113586426, 0.011331680297851563, 0.011224127769470215, 0.011281215667724609, 0.01123680019378662, 0.011340319633483887, 0.011176159858703613, 0.011200384140014648, 0.011235296249389648, 0.011216896057128906, 0.011278335571289062, 0.011333919525146485, 0.011385600090026855, 0.011286687850952149, 0.011305312156677245, 0.011327263832092285, 0.011418304443359374, 0.01136796760559082, 0.011430368423461915, 0.011419551849365234, 0.012865407943725586, 0.011578751564025878, 0.01147100830078125, 0.01140118408203125, 0.011432607650756836, 0.011395199775695801, 0.011441856384277345, 0.01144972801208496, 0.011476160049438477, 0.011587264060974121, 0.011507840156555175, 0.011484383583068847, 0.011430815696716308, 0.011513343811035156, 0.01143126392364502, 0.011440799713134766, 0.01140940761566162, 0.011530495643615723, 0.011558239936828613, 0.01160371208190918, 0.011573151588439941, 0.011569952011108399, 0.011495200157165528, 0.011482912063598633, 0.011414175987243652, 0.011422847747802735, 0.011434880256652832, 0.011519359588623048, 0.011436672210693359, 0.0114171199798584, 0.011534079551696778, 0.011498208045959473, 0.011470848083496094, 0.011420736312866211, 0.011387871742248536, 0.011331680297851563, 0.01136787223815918, 0.011262175559997558, 0.01124687957763672, 0.011088607788085937, 0.011468768119812012, 0.01138252830505371, 0.011340031623840332, 0.01134598445892334, 0.011241151809692382, 0.011415552139282227, 0.011268095970153809, 0.011351231575012208, 0.01121065616607666, 0.011311455726623535, 0.011274815559387207, 0.011770112037658691, 0.011855615615844726, 0.011703519821166993, 0.01183414363861084, 0.011359935760498046, 0.011292991638183593, 0.011280384063720703, 0.011249695777893066, 0.011220959663391114, 0.011225088119506836, 0.011243519783020019, 0.011216896057128906, 0.011234592437744141, 0.011248160362243652, 0.01118841552734375, 0.011231295585632324, 0.01126147174835205, 0.011248031616210937, 0.011356160163879395, 0.011566271781921387, 0.011623231887817383, 0.011593728065490723, 0.011503392219543458, 0.011526368141174316, 0.011777536392211914, 0.01170307159423828, 0.011705056190490722, 0.011633664131164552, 0.011556703567504883, 0.011509535789489746, 0.011535840034484863, 0.01151692771911621, 0.011590592384338379, 0.011651424407958985, 0.01161894416809082, 0.011617664337158204, 0.011775775909423828, 0.011679712295532227, 0.011608223915100097, 0.01166425609588623, 0.011669440269470215, 0.011741087913513184, 0.011821056365966797, 0.011821056365966797, 0.011597408294677734, 0.011573375701904297, 0.011591967582702636, 0.011632991790771484, 0.011294367790222169, 0.011347359657287597, 0.011409055709838868, 0.011405311584472656, 0.011651040077209473, 0.011698207855224609, 0.01165721607208252, 0.011586560249328613, 0.01154355239868164, 0.011553824424743652, 0.011398112297058106, 0.01145241641998291, 0.011576607704162597, 0.011461024284362792, 0.011354432106018067, 0.011395071983337402, 0.011447775840759277, 0.01131935977935791, 0.011526111602783204, 0.011369183540344238, 0.01131497573852539, 0.011371999740600587, 0.011373087882995606, 0.011346112251281739, 0.011360063552856445, 0.011386879920959473, 0.011419615745544433, 0.011374624252319335, 0.011337887763977051, 0.011381759643554687, 0.011508671760559083, 0.011468704223632813, 0.011585536003112793, 0.011557951927185058, 0.011334591865539551, 0.011310463905334472, 0.011267904281616212, 0.01127507209777832, 0.011332703590393066, 0.01130793571472168, 0.011272192001342773, 0.011259903907775879, 0.011263999938964844, 0.011261311531066895, 0.011276000022888184, 0.011248543739318847, 0.01125369644165039, 0.011231103897094726, 0.011288224220275879, 0.011293472290039063, 0.011259231567382812, 0.011260319709777832, 0.011262271881103516, 0.011361760139465332, 0.011584768295288086, 0.011307040214538574, 0.011324383735656739, 0.011296704292297363, 0.01134012794494629, 0.011361984252929688, 0.011298815727233886, 0.011330975532531738, 0.011333919525146485, 0.011348511695861816, 0.0114169282913208, 0.011497152328491212, 0.012027487754821778, 0.011574111938476562, 0.0116428804397583, 0.011390975952148438, 0.011563008308410644, 0.0116428804397583, 0.011363743782043457, 0.011469247817993165, 0.011325599670410157, 0.01152950382232666, 0.011551199913024902, 0.011322655677795411, 0.011316320419311524, 0.011231103897094726, 0.011314623832702637, 0.011254688262939454, 0.01140662384033203, 0.011229439735412598, 0.011254112243652343, 0.011211615562438965, 0.011282431602478027, 0.011326399803161621, 0.011257184028625489, 0.011262016296386719, 0.011207263946533203, 0.011232383728027344, 0.011305407524108886, 0.011262399673461914, 0.011304960250854493, 0.01124556827545166, 0.011244864463806152, 0.011252415657043458, 0.011378080368041991, 0.01124617576599121, 0.011280287742614746, 0.011660927772521973, 0.012281920433044434, 0.012151519775390624, 0.011476672172546387, 0.011294015884399414, 0.011532992362976073, 0.011392448425292969, 0.011419679641723633, 0.01149392032623291, 0.011507328033447266, 0.011538816452026366, 0.011731103897094726, 0.01148031997680664, 0.011518560409545898, 0.011707648277282715, 0.011471263885498047, 0.011491616249084473, 0.011447360038757324, 0.01148412799835205, 0.01168182373046875, 0.012041312217712402, 0.011553343772888183, 0.011510111808776856, 0.011593215942382813, 0.011566975593566894, 0.011558655738830567, 0.011801088333129883, 0.011587264060974121, 0.011406880378723144, 0.011686367988586426, 0.011440128326416015, 0.011627840042114259, 0.011637439727783203, 0.011429887771606445, 0.011537440299987793, 0.011457344055175782, 0.011647263526916503, 0.011495295524597169, 0.011575200080871583, 0.011534239768981934, 0.011588895797729492, 0.011454591751098633, 0.011549471855163574, 0.011757568359375, 0.011622400283813476, 0.011550623893737793, 0.011527968406677245, 0.011519424438476563, 0.011487839698791504, 0.011532575607299804, 0.0116779842376709, 0.011550399780273437, 0.011546655654907227, 0.01155081558227539, 0.01157145595550537, 0.011563936233520507, 0.011567584037780762, 0.011536800384521484, 0.011599488258361816, 0.011583744049072265, 0.011562911987304688, 0.011529888153076172, 0.011477439880371094, 0.01144422435760498, 0.011487232208251954, 0.011521535873413086, 0.011497535705566407, 0.011549087524414062, 0.011487263679504395, 0.011661312103271485, 0.011429887771606445, 0.01145036792755127, 0.011460800170898438, 0.01144927978515625, 0.01141756820678711, 0.011411840438842774, 0.011421440124511718, 0.01155907154083252, 0.011528736114501953, 0.012009568214416504, 0.011638976097106934, 0.011622207641601563, 0.011583488464355468, 0.01160912036895752, 0.011588255882263183, 0.011632960319519043, 0.011699423789978028, 0.011747103691101074, 0.011604991912841797, 0.011628543853759766, 0.01157744026184082, 0.011522144317626952, 0.011870559692382812, 0.01173459243774414, 0.011749216079711914, 0.011684736251831055, 0.011573311805725097, 0.011674495697021484, 0.011569952011108399, 0.011797696113586427, 0.011747232437133789, 0.011868831634521484, 0.011695648193359375, 0.011712320327758789, 0.011699104309082031, 0.011755135536193848, 0.01174892807006836, 0.011548959732055664, 0.011411999702453613, 0.011452735900878907, 0.011359071731567383, 0.011346783638000488, 0.01132953643798828, 0.011319104194641114, 0.011277983665466309, 0.011364895820617676, 0.011281439781188965, 0.011302176475524903, 0.011283167839050293, 0.011279328346252442, 0.011243519783020019, 0.011241472244262696, 0.011304448127746582, 0.011274784088134766, 0.01127939224243164, 0.01125062370300293, 0.01132153606414795, 0.01126585578918457, 0.011243776321411133, 0.011277888298034668, 0.011392704010009766, 0.011475520133972168, 0.011506848335266114, 0.011553567886352539, 0.011544575691223144, 0.011504672050476073, 0.011539423942565918, 0.011489279747009277, 0.01148249626159668, 0.011401856422424316, 0.011388192176818848, 0.011440863609313965, 0.011406975746154785, 0.011524479866027832, 0.011514016151428223, 0.011478879928588866, 0.011489279747009277, 0.011495776176452636, 0.011515616416931152, 0.011445216178894043, 0.011524224281311035, 0.011443103790283203, 0.011438143730163575, 0.011558624267578125, 0.01145088005065918, 0.011946432113647461, 0.012744159698486328, 0.013918304443359375, 0.011760064125061035, 0.011454239845275879, 0.011450592041015626, 0.01135148811340332, 0.011385408401489258, 0.011295999526977539, 0.0112991361618042, 0.01128934383392334, 0.011409088134765625, 0.011450207710266114, 0.011407872200012208, 0.011488032341003419, 0.011385727882385253, 0.011243040084838868, 0.011333727836608886, 0.011247039794921875, 0.011389887809753418, 0.01143398380279541, 0.011334848403930665, 0.011411904335021972, 0.01136473560333252, 0.011292672157287598, 0.011271583557128905, 0.01128825569152832, 0.011278719902038574, 0.011726911544799806, 0.011370623588562011, 0.011608415603637696, 0.011376832008361816, 0.011354080200195312, 0.011351967811584473, 0.011384767532348633, 0.011304127693176269, 0.011352928161621093, 0.011325504302978516, 0.011401120185852051, 0.01155072021484375, 0.011531776428222656, 0.011538944244384765, 0.011388319969177246, 0.011485631942749024, 0.011309215545654297, 0.011259615898132324, 0.011268256187438964, 0.011288127899169921, 0.01122976016998291, 0.01128825569152832, 0.011302847862243653, 0.011351840019226074, 0.011283040046691895, 0.01130239963531494, 0.011294560432434082, 0.011342720031738281, 0.011259679794311523, 0.011272192001342773, 0.011261792182922363, 0.011323712348937988, 0.011277376174926758, 0.011351872444152832, 0.011060768127441407, 0.01130083179473877, 0.011373087882995606, 0.011223039627075194, 0.011313152313232423, 0.011272192001342773, 0.011229056358337403, 0.01152012825012207, 0.011280384063720703, 0.01134335994720459, 0.011264639854431152, 0.01119206428527832, 0.011238783836364745, 0.011266143798828124, 0.011231776237487792, 0.011243359565734863, 0.011309344291687012, 0.011237343788146973, 0.011253791809082031, 0.011226176261901855, 0.01130406379699707, 0.011236895561218261, 0.011174240112304688, 0.011255743980407715, 0.011229184150695801, 0.011157504081726074, 0.011214112281799317, 0.011211135864257812, 0.011233632087707519, 0.011228192329406738, 0.011156255722045899, 0.01122537612915039, 0.011193471908569336, 0.011192223548889161, 0.011273088455200196, 0.011206656455993653, 0.011258208274841308, 0.01126959991455078, 0.011335871696472167, 0.011396575927734375, 0.011434528350830078, 0.011341823577880859, 0.011403264045715332, 0.011432255744934082, 0.011663040161132813, 0.011469887733459474, 0.011458975791931152, 0.011501279830932618, 0.011456928253173827, 0.01139958381652832, 0.011423744201660157, 0.011406368255615234, 0.01171174430847168, 0.01147056007385254, 0.011601920127868653, 0.011392640113830567, 0.011360671997070313, 0.011433856010437012, 0.011335935592651367, 0.011564736366271972, 0.011388223648071289, 0.01135206413269043, 0.011373408317565918, 0.011412896156311036, 0.011710656166076661, 0.011675583839416503, 0.01176419162750244, 0.01172480010986328, 0.011798848152160644, 0.011900704383850097, 0.011863903999328613, 0.012111712455749511, 0.01194985580444336, 0.011942336082458496, 0.011911168098449706, 0.011839232444763183, 0.011695679664611816, 0.011718848228454589, 0.011803135871887208, 0.011666912078857422, 0.011622336387634278, 0.011560544013977051, 0.011570015907287597, 0.011670720100402833, 0.011633631706237792, 0.012658880233764649, 0.011720288276672363, 0.011671775817871094, 0.01166323184967041, 0.01162662410736084, 0.011562656402587891, 0.011684415817260742, 0.011495200157165528, 0.011542688369750977, 0.011478879928588866, 0.011486720085144043, 0.011510272026062012, 0.011491392135620117, 0.011511743545532227, 0.011468799591064453, 0.011578559875488281, 0.011553600311279296, 0.011573247909545899, 0.011522047996520996, 0.011537631988525391, 0.011473695755004882, 0.011417920112609864, 0.01140937614440918, 0.011400544166564941, 0.011421855926513671, 0.011374272346496583, 0.01138742446899414, 0.011400351524353028, 0.011523072242736816, 0.011544416427612305, 0.011511551856994628, 0.011464768409729004, 0.011424256324768066, 0.011446047782897949, 0.011374496459960937, 0.011367456436157226, 0.01142796802520752, 0.01181987190246582, 0.011470656394958496, 0.011558879852294921, 0.011698271751403809, 0.011947936058044433, 0.012320768356323243, 0.011929920196533203, 0.01181446361541748, 0.011773344039916991, 0.011782719612121581, 0.011998720169067383, 0.011981439590454102, 0.011710176467895508, 0.011808671951293946, 0.011704287528991698, 0.011635135650634766, 0.011674847602844238, 0.011926303863525391, 0.01163468837738037, 0.011686016082763671, 0.011825056076049804, 0.01174665641784668, 0.011768159866333007, 0.011624064445495606, 0.011600543975830078, 0.011790335655212402, 0.011699872016906738, 0.011752960205078124, 0.011708383560180663, 0.01176159954071045, 0.011840448379516602, 0.011638784408569336, 0.011569087982177734, 0.011511808395385742, 0.011429951667785645, 0.011514080047607423, 0.011498944282531738, 0.011546719551086425, 0.01162880039215088, 0.011642463684082031, 0.011549087524414062, 0.011648063659667968, 0.011690336227416992, 0.011649279594421387, 0.01178048038482666, 0.011672608375549317, 0.01165510368347168, 0.01176579189300537, 0.011651264190673829, 0.011694879531860351, 0.01164083194732666, 0.011595775604248047, 0.011849408149719238, 0.011887040138244629, 0.011870368003845215, 0.011785216331481933, 0.011793120384216309, 0.011653120040893555, 0.01176531219482422, 0.01154032039642334, 0.011472672462463378, 0.011524928092956544, 0.011470848083496094, 0.011583488464355468, 0.011517951965332032, 0.011391039848327636, 0.011452447891235352, 0.011499263763427735, 0.01184547233581543, 0.011791872024536134, 0.011622912406921386, 0.01146662425994873, 0.011416095733642578, 0.011463680267333985, 0.011471232414245606, 0.011428480148315429, 0.011526144027709961, 0.011431936264038087, 0.011519840240478516, 0.011452608108520508, 0.0116080322265625, 0.01155081558227539, 0.011462143898010254, 0.011532704353332519, 0.01162649631500244, 0.01164902400970459, 0.01181107234954834, 0.011560704231262207, 0.01189254379272461, 0.011679936408996582, 0.011687328338623047, 0.011589664459228516, 0.011514431953430175, 0.011530400276184082, 0.011517024040222168, 0.011432543754577636, 0.01145257568359375, 0.011462656021118164, 0.011502880096435547, 0.011419391632080078, 0.011514528274536133, 0.011438400268554687, 0.011436032295227052, 0.011448543548583984, 0.011387680053710937, 0.011536416053771972, 0.01142416000366211, 0.011557215690612794, 0.011524319648742676, 0.011521504402160645, 0.011532832145690917, 0.011560959815979004, 0.011745280265808105, 0.011413503646850585, 0.011466464042663575, 0.011520159721374513, 0.011449503898620605, 0.01150051212310791, 0.011435263633728028, 0.011440095901489257, 0.01132857608795166, 0.011366111755371094, 0.011436063766479493, 0.011390463829040527, 0.011350496292114258, 0.011357312202453613, 0.011274463653564453, 0.011427680015563964, 0.011375295639038085, 0.011364480018615723]",tokens/s,87.20710510562175,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1480.830976,1823.408128,0.0,1428.160512,1322.516992,s,1,9.0790751953125,9.0790751953125,0.0,9.0790751953125,9.0790751953125,9.0790751953125,9.0790751953125,[9.0790751953125],,kWh,4.854528277079074e-05,5.330303647538533e-06,1.8832515065986266e-05,7.270810148431555e-05,,MB,1523.392512,1842.282496,0.0,1434.451968,1320.892416,s,10,5.394534423828125,0.5394534423828125,0.0015562314511154997,0.5389523620605469,0.5410210266113281,0.542272232055664,0.5432731964111328,"[0.5435234375, 0.5383721313476563, 0.5386351318359375, 0.5384713745117188, 0.5396533203125, 0.5378704833984375, 0.5407429809570312, 0.5386964111328125, 0.53936083984375, 0.5392083129882812]",tokens/s,474.55439132842656,kWh,1.6081650892326186e-05,1.773542507303471e-06,1.0747742516316255e-05,2.8602935915945914e-05,tokens/kWh,8950130.180772178,MB,1527.717888,1842.282496,0.0,1434.451968,1373.031936,s,10,13.756556762695311,1.3756556762695311,0.008469377521908704,1.3762078857421876,1.3836558349609374,1.3867905639648437,1.3892983471679687,"[1.382959228515625, 1.3754798583984376, 1.3769359130859375, 1.38992529296875, 1.35745751953125, 1.374190673828125, 1.3677625732421874, 1.38070458984375, 1.3712379150390626, 1.3799031982421874]",tokens/s,45.79634358129633,kWh,3.9714110341008635e-05,4.380432631009805e-06,1.953750393408206e-05,6.36320469061005e-05,tokens/kWh,990067.1605451702,,s,630,13.754047189712523,0.021831820936051623,0.00044146758512300676,0.02171879959106445,0.022216588401794433,0.022392360305786133,0.02359489480972292,"[0.02289436721801758, 0.02641001510620117, 0.02230054473876953, 0.022132095336914064, 0.022047359466552733, 0.02209190368652344, 0.021983104705810545, 0.021936128616333008, 0.021938175201416017, 0.021868320465087892, 0.02159756851196289, 0.021566303253173828, 0.021570880889892577, 0.023116159439086913, 0.021795135498046875, 0.021960704803466798, 0.02183359909057617, 0.021584415435791017, 0.021737056732177733, 0.02147532844543457, 0.021600255966186522, 0.021560319900512694, 0.02166681671142578, 0.021626880645751953, 0.02152448081970215, 0.02148761558532715, 0.02170585632324219, 0.021988224029541016, 0.021835775375366212, 0.02187059211730957, 0.02184806442260742, 0.02186454391479492, 0.022320575714111328, 0.02202003288269043, 0.022766111373901367, 0.02197212791442871, 0.022512447357177733, 0.023052288055419923, 0.021954591751098634, 0.021936128616333008, 0.02184806442260742, 0.02198121643066406, 0.02196847915649414, 0.021666175842285158, 0.02157766342163086, 0.0217457275390625, 0.021733375549316408, 0.02166988754272461, 0.021655168533325195, 0.021579456329345704, 0.02187129592895508, 0.021710687637329102, 0.021581632614135742, 0.02147324752807617, 0.02159654426574707, 0.021733375549316408, 0.021544960021972655, 0.021643264770507813, 0.02164735984802246, 0.02189689636230469, 0.02212486457824707, 0.02166969680786133, 0.021823680877685547, 0.02243132781982422, 0.021907743453979493, 0.021675359725952147, 0.021564064025878907, 0.02166374397277832, 0.0215644474029541, 0.02158060836791992, 0.021532768249511718, 0.021700672149658203, 0.021579296112060546, 0.02166831970214844, 0.02147942352294922, 0.02162073516845703, 0.02215936088562012, 0.021534719467163087, 0.021555200576782226, 0.0215285758972168, 0.021703680038452147, 0.021787647247314454, 0.021590015411376954, 0.02150156784057617, 0.02161497688293457, 0.021569536209106444, 0.021540864944458008, 0.02145884895324707, 0.021561439514160157, 0.021554431915283202, 0.021557024002075195, 0.021530879974365234, 0.021610847473144533, 0.021598495483398438, 0.021642911911010743, 0.021469535827636718, 0.02207753562927246, 0.022587392807006838, 0.022265247344970703, 0.02207548713684082, 0.021600191116333007, 0.02183558464050293, 0.021625600814819335, 0.022115488052368164, 0.0218570556640625, 0.021648639678955077, 0.02152454376220703, 0.021604736328125, 0.021706239700317383, 0.021705120086669923, 0.021682655334472656, 0.021727231979370116, 0.021741535186767576, 0.02167184066772461, 0.02182566452026367, 0.021755903244018555, 0.022105951309204102, 0.024666271209716796, 0.022394880294799805, 0.022292255401611328, 0.02208176040649414, 0.022323200225830078, 0.02266111946105957, 0.022039840698242188, 0.022088159561157228, 0.021897472381591798, 0.022680160522460937, 0.02209382438659668, 0.022060831069946288, 0.022089759826660157, 0.022026432037353515, 0.0219238395690918, 0.022007776260375977, 0.022054943084716797, 0.022024383544921877, 0.02183558464050293, 0.021809152603149414, 0.021651424407958985, 0.02171014404296875, 0.02156959915161133, 0.021765792846679687, 0.021804128646850586, 0.021823392868041993, 0.021740800857543947, 0.021697280883789062, 0.021594112396240234, 0.02150115203857422, 0.02163587188720703, 0.021659648895263672, 0.021550752639770507, 0.02146544075012207, 0.02150809669494629, 0.021600255966186522, 0.021691551208496095, 0.021652320861816406, 0.021708799362182618, 0.021960704803466798, 0.021968704223632812, 0.021813440322875976, 0.021927743911743163, 0.02194438362121582, 0.022171775817871095, 0.02202137565612793, 0.02188569641113281, 0.021804031372070314, 0.0220765438079834, 0.021831552505493165, 0.0218666877746582, 0.021956544876098633, 0.02182649612426758, 0.02208393669128418, 0.02175644874572754, 0.021608160018920897, 0.021500255584716795, 0.021572704315185546, 0.021681407928466796, 0.021769056320190428, 0.021764511108398436, 0.021688512802124024, 0.0237194881439209, 0.02326697540283203, 0.021738784790039063, 0.021693151473999025, 0.021609472274780273, 0.02164735984802246, 0.021555200576782226, 0.021665056228637694, 0.02164556884765625, 0.021723615646362306, 0.022250591278076173, 0.021885215759277345, 0.022164096832275392, 0.02175551986694336, 0.021551488876342773, 0.021476800918579102, 0.021465311050415038, 0.02212019157409668, 0.0218100471496582, 0.02149081611633301, 0.02150163269042969, 0.02140457534790039, 0.02147123146057129, 0.021626720428466795, 0.021558687210083007, 0.02174847984313965, 0.02189107131958008, 0.022155263900756835, 0.022437887191772463, 0.022708223342895507, 0.022375839233398438, 0.022221343994140625, 0.02234988784790039, 0.022226943969726562, 0.02231737518310547, 0.022171327590942383, 0.022428991317749024, 0.02225632095336914, 0.02227596855163574, 0.022130815505981446, 0.022158624649047852, 0.02218671989440918, 0.02212620735168457, 0.02201366424560547, 0.022061695098876954, 0.022310848236083983, 0.022134880065917968, 0.02208358383178711, 0.022157344818115234, 0.022206655502319338, 0.022230239868164064, 0.022127168655395508, 0.021981184005737304, 0.022071647644042968, 0.02201753616333008, 0.02217385673522949, 0.021997568130493163, 0.022368255615234374, 0.022207775115966798, 0.02229487991333008, 0.022161792755126954, 0.021964384078979493, 0.022784128189086914, 0.022044960021972655, 0.022025983810424806, 0.022415456771850587, 0.022016096115112304, 0.022038591384887694, 0.022221824645996095, 0.022021120071411132, 0.021940223693847655, 0.021927871704101563, 0.021961984634399415, 0.022281152725219727, 0.02176540756225586, 0.021535455703735353, 0.02152448081970215, 0.02150297546386719, 0.021878015518188475, 0.021577119827270508, 0.02158585548400879, 0.021481536865234376, 0.021987680435180665, 0.021458080291748047, 0.021500768661499022, 0.021602304458618164, 0.021501951217651367, 0.021497024536132812, 0.02156422424316406, 0.02188083267211914, 0.021531999588012694, 0.021456703186035157, 0.021433408737182618, 0.021518112182617188, 0.02160153579711914, 0.02150783920288086, 0.0214866886138916, 0.021772159576416015, 0.02170992088317871, 0.021707712173461916, 0.0214835205078125, 0.021395456314086913, 0.021424287796020507, 0.02151203155517578, 0.02148524856567383, 0.021433696746826172, 0.021470176696777345, 0.021538816452026367, 0.02146505546569824, 0.021517568588256836, 0.02150480079650879, 0.021551103591918946, 0.021481792449951173, 0.02147225570678711, 0.02145964813232422, 0.021458879470825195, 0.021477439880371093, 0.021414976119995117, 0.021502687454223634, 0.021465311050415038, 0.021355712890625, 0.021416032791137695, 0.021500640869140625, 0.021450080871582032, 0.021507072448730468, 0.021616128921508788, 0.021558528900146486, 0.021502880096435546, 0.021524576187133788, 0.021425247192382812, 0.021625663757324217, 0.021559295654296876, 0.02144256019592285, 0.021452064514160155, 0.021465824127197265, 0.02146303939819336, 0.022627040863037108, 0.021825536727905274, 0.021417695999145506, 0.02146544075012207, 0.021523872375488282, 0.021460832595825194, 0.02149622344970703, 0.021353759765625, 0.02243881607055664, 0.021430303573608398, 0.02158585548400879, 0.021579904556274412, 0.021577407836914062, 0.02154323196411133, 0.021445823669433595, 0.021450847625732423, 0.021573856353759767, 0.021508415222167968, 0.021461183547973633, 0.02141798400878906, 0.021893280029296875, 0.023903488159179687, 0.02158064079284668, 0.021572864532470704, 0.021591552734375, 0.02170572853088379, 0.021655647277832032, 0.021774240493774414, 0.021790719985961913, 0.02207744026184082, 0.02590105628967285, 0.022167295455932618, 0.022152576446533203, 0.022475648880004882, 0.022040639877319336, 0.022054079055786133, 0.021960800170898437, 0.021674655914306642, 0.02171494483947754, 0.02146303939819336, 0.021536928176879883, 0.021671775817871095, 0.021670976638793946, 0.021533632278442384, 0.02149344062805176, 0.02152275276184082, 0.02147635269165039, 0.021445632934570313, 0.021476831436157227, 0.022389280319213868, 0.02149488067626953, 0.021566368103027343, 0.021517440795898436, 0.021838272094726562, 0.021801599502563475, 0.02158777618408203, 0.022114112854003908, 0.02214726448059082, 0.02188083267211914, 0.021683616638183592, 0.02159676742553711, 0.021571584701538086, 0.021609823226928712, 0.022196287155151366, 0.021909408569335938, 0.021610591888427736, 0.021492767333984374, 0.021428512573242187, 0.021520063400268553, 0.021572608947753907, 0.021522432327270507, 0.02156716728210449, 0.02149407958984375, 0.02147327995300293, 0.021514240264892577, 0.02150339126586914, 0.021535263061523438, 0.021891136169433594, 0.02195235252380371, 0.021743776321411133, 0.021798976898193358, 0.021669824600219725, 0.021499807357788087, 0.02155276870727539, 0.021467424392700194, 0.021532447814941406, 0.021431711196899413, 0.02147020721435547, 0.021379072189331053, 0.02147737693786621, 0.02147020721435547, 0.021511167526245118, 0.021618623733520508, 0.021727296829223634, 0.021630975723266603, 0.0216407356262207, 0.021602783203125, 0.02164121627807617, 0.021649728775024413, 0.02163475227355957, 0.021771520614624024, 0.021678207397460937, 0.02166579246520996, 0.021658239364624025, 0.021544960021972655, 0.021659648895263672, 0.02168627166748047, 0.02162483215332031, 0.0216760311126709, 0.021640512466430666, 0.02198294448852539, 0.021563711166381835, 0.02144937515258789, 0.02149113655090332, 0.021434112548828124, 0.021467967987060545, 0.02190540885925293, 0.02328985595703125, 0.022039871215820312, 0.021566144943237303, 0.02211020851135254, 0.021532928466796875, 0.021443552017211914, 0.02188163185119629, 0.023802879333496094, 0.022623231887817383, 0.022318431854248047, 0.022086591720581056, 0.022245376586914063, 0.022183359146118162, 0.022246015548706054, 0.02215116882324219, 0.02221670341491699, 0.02207472038269043, 0.02219468879699707, 0.022210464477539063, 0.022181760787963866, 0.02220070457458496, 0.021989376068115234, 0.021714336395263673, 0.021626815795898438, 0.021671808242797852, 0.021811071395874022, 0.022137760162353515, 0.022337087631225584, 0.022216575622558594, 0.021805631637573243, 0.02196512031555176, 0.02164908790588379, 0.02180905532836914, 0.021700511932373046, 0.02159404754638672, 0.021647615432739256, 0.02167398452758789, 0.0219237117767334, 0.021808832168579102, 0.021917535781860353, 0.021912128448486327, 0.02225155258178711, 0.021972991943359374, 0.02191974449157715, 0.02163711929321289, 0.021576927185058593, 0.021553279876708985, 0.02153334426879883, 0.021594112396240234, 0.02157360076904297, 0.02152787208557129, 0.021597183227539063, 0.02153388786315918, 0.021475008010864258, 0.02168284797668457, 0.02156563186645508, 0.021452192306518555, 0.021510080337524416, 0.021813631057739258, 0.021885183334350584, 0.022249343872070313, 0.022140832901000978, 0.022128896713256838, 0.022245376586914063, 0.02210767936706543, 0.02203081512451172, 0.022171648025512695, 0.022263744354248046, 0.022082944869995118, 0.022080448150634764, 0.022123903274536134, 0.021938560485839843, 0.02262828826904297, 0.022160160064697267, 0.021970655441284178, 0.021778976440429688, 0.02172083282470703, 0.021651456832885742, 0.02163711929321289, 0.021966848373413086, 0.02239897537231445, 0.02191276741027832, 0.021852447509765626, 0.021764223098754882, 0.021811456680297853, 0.02170275115966797, 0.021665855407714842, 0.021777631759643555, 0.021779232025146485, 0.021741567611694337, 0.02162646484375, 0.022229087829589843, 0.021588287353515624, 0.021755456924438477, 0.02200783920288086, 0.0217174072265625, 0.021508256912231447, 0.021607999801635743, 0.021682464599609375, 0.021612543106079102, 0.021696287155151366, 0.02153606414794922, 0.021611423492431642, 0.02149567985534668, 0.02144883155822754, 0.02161065673828125, 0.02153251266479492, 0.021492031097412108, 0.021582752227783202, 0.021490463256835936, 0.0215285758972168, 0.021555200576782226, 0.02156745529174805, 0.021622079849243164, 0.021668479919433593, 0.021583520889282226, 0.02168876838684082, 0.021757951736450197, 0.021760000228881835, 0.02176185607910156, 0.02176838493347168, 0.02169036865234375, 0.02165760040283203, 0.021772064208984376, 0.021680351257324218, 0.02179017639160156, 0.02185641670227051, 0.021922176361083984, 0.021908544540405275, 0.021994272232055665, 0.021931903839111328, 0.02201772880554199, 0.02192854309082031, 0.02194972801208496, 0.021872928619384766, 0.022202783584594727, 0.021967103958129883, 0.021839872360229492, 0.022194175720214843, 0.02267750358581543, 0.02230067253112793, 0.022144704818725585, 0.02216569519042969, 0.02201612854003906, 0.021900768280029296, 0.021923999786376953, 0.02213052749633789, 0.022867712020874023, 0.022124671936035158, 0.022082559585571288, 0.022183584213256835, 0.022214656829833986, 0.022155263900756835, 0.02188287925720215, 0.021855968475341797, 0.021924127578735353, 0.021847679138183595, 0.021730783462524415, 0.021699039459228516, 0.021714815139770506, 0.021692991256713867, 0.021811199188232423, 0.021720191955566407, 0.021552000045776367, 0.02145894432067871, 0.021573888778686524, 0.02181497573852539, 0.021726560592651368, 0.02166815948486328, 0.021657695770263673, 0.02175820732116699, 0.021725248336791993, 0.021764095306396485, 0.021793888092041015, 0.021810079574584963, 0.021673919677734375, 0.021553216934204103, 0.021563392639160156, 0.02184163284301758, 0.02173753547668457, 0.021758176803588866, 0.02167807960510254, 0.021647071838378905, 0.02166422462463379, 0.02158777618408203, 0.021834943771362306, 0.02431830406188965, 0.02178611183166504, 0.02158243179321289, 0.021534175872802735, 0.02164374351501465, 0.021731584548950196, 0.02194246482849121, 0.021828704833984375, 0.021805984497070312, 0.021892160415649415, 0.021904512405395506, 0.02184582328796387]",tokens/s,45.80469961388636,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1845.264384,2861.498368,0.0,2466.250752,2401.696256,s,1,9.3064697265625,9.3064697265625,0.0,9.3064697265625,9.3064697265625,9.3064697265625,9.3064697265625,[9.3064697265625],,kWh,6.920842063327655e-05,7.62660962710615e-06,2.5094464520009074e-05,0.00010192949478039177,,MB,1639.661568,3087.990784,0.0,2680.160256,2582.173696,s,10,5.461781799316406,0.5461781799316406,0.002053754413714127,0.5455082397460937,0.548137109375,0.5496025207519531,0.5507748498535157,"[0.5510679321289063, 0.544267578125, 0.54442578125, 0.5445333251953125, 0.5478114624023438, 0.5469923095703125, 0.5443823852539063, 0.5457959594726562, 0.5472845458984374, 0.5452205200195313]",tokens/s,468.711511016498,kWh,1.6342655354862e-05,1.8015006276936102e-06,1.0818573469665674e-05,2.8962729452221288e-05,tokens/kWh,8838945.943348104,MB,1639.661568,3090.087936,0.0,2680.160256,2582.176256,s,10,15.202343749999999,1.5202343749999998,0.008446818794560296,1.5178092041015625,1.5296287109375,1.5350193725585939,1.5393319018554688,"[1.5284307861328126, 1.5095377197265625, 1.51311474609375, 1.521512451171875, 1.5233001708984375, 1.51756640625, 1.5404100341796876, 1.5158154296875, 1.51460400390625, 1.518052001953125]",tokens/s,41.440978467547154,kWh,4.3751790035144206e-05,4.8266266312004895e-06,2.3022925825737693e-05,7.160134249208238e-05,tokens/kWh,879871.7706580221,,s,630,15.200313385009773,0.024127481563507566,0.00042550805216144027,0.024015567779541017,0.02454982967376709,0.02477015514373779,0.025282242221832275,"[0.02515177536010742, 0.02418659210205078, 0.0241232967376709, 0.024244319915771483, 0.024126752853393555, 0.02398886489868164, 0.024005792617797853, 0.02413814353942871, 0.024048160552978516, 0.02402444839477539, 0.024124031066894532, 0.024070144653320313, 0.024057119369506837, 0.024205375671386718, 0.023984607696533204, 0.023998655319213868, 0.023975231170654296, 0.023982112884521484, 0.02395952033996582, 0.023952064514160157, 0.023975711822509765, 0.024025312423706056, 0.02411929512023926, 0.024197120666503907, 0.024344575881958007, 0.024393728256225586, 0.024444927215576173, 0.024451072692871095, 0.024434688568115235, 0.024427967071533205, 0.024650304794311524, 0.02460198402404785, 0.024660383224487305, 0.024381248474121094, 0.024301376342773438, 0.024125471115112304, 0.024123071670532226, 0.024127872467041015, 0.024166208267211914, 0.024955583572387696, 0.024201215744018553, 0.02417411231994629, 0.024117727279663086, 0.02409062385559082, 0.024473600387573242, 0.02579631996154785, 0.025287967681884765, 0.024222240447998047, 0.024191455841064455, 0.024169984817504882, 0.02427724838256836, 0.024950944900512695, 0.024669599533081055, 0.02446406364440918, 0.024188928604125977, 0.023963647842407225, 0.02394460868835449, 0.02395372772216797, 0.024072479248046875, 0.023948959350585938, 0.02390265655517578, 0.023964704513549803, 0.02393587112426758, 0.025022464752197264, 0.024383487701416014, 0.024134912490844727, 0.024048383712768556, 0.023867168426513673, 0.023865568161010743, 0.02390425682067871, 0.024207328796386717, 0.023920671463012695, 0.023932928085327147, 0.023889759063720702, 0.023875743865966796, 0.023821504592895507, 0.02380803108215332, 0.023849760055541992, 0.02384886360168457, 0.02378329658508301, 0.024049888610839842, 0.02384828758239746, 0.02392540740966797, 0.023901824951171876, 0.02374220848083496, 0.023847551345825196, 0.023816192626953125, 0.023767040252685546, 0.023837696075439452, 0.02386227226257324, 0.023778783798217774, 0.023888416290283203, 0.023887264251708985, 0.023954015731811523, 0.02406399917602539, 0.023828479766845705, 0.023922016143798828, 0.023868064880371093, 0.023901472091674803, 0.023834623336791993, 0.023919328689575196, 0.02393926429748535, 0.023914239883422853, 0.023930944442749024, 0.023994367599487306, 0.024022815704345703, 0.0239815673828125, 0.023909088134765624, 0.023934207916259765, 0.02390912055969238, 0.023949312210083007, 0.023982080459594726, 0.02423526382446289, 0.023995136260986327, 0.024016895294189454, 0.02400985527038574, 0.024099103927612303, 0.02404582405090332, 0.023992671966552734, 0.024068191528320314, 0.02403721618652344, 0.024014911651611327, 0.023895072937011718, 0.023910911560058593, 0.02397795104980469, 0.02397235107421875, 0.025036800384521486, 0.024236032485961914, 0.024029151916503906, 0.023980064392089842, 0.023962976455688477, 0.024015520095825197, 0.024012479782104492, 0.024004928588867186, 0.023990272521972656, 0.023971839904785155, 0.02391244888305664, 0.02388787269592285, 0.023777055740356445, 0.02388400077819824, 0.02393497657775879, 0.02390732765197754, 0.02394153594970703, 0.023818656921386717, 0.023896255493164063, 0.02386124801635742, 0.02381337547302246, 0.023912576675415038, 0.023932832717895508, 0.023908639907836916, 0.023769088745117187, 0.023869152069091796, 0.023892831802368165, 0.023918464660644533, 0.023897855758666993, 0.02402124786376953, 0.023870752334594725, 0.023866048812866213, 0.023966880798339845, 0.02393907165527344, 0.023961503982543944, 0.02392697525024414, 0.02387171173095703, 0.024017471313476563, 0.02405171203613281, 0.02409219169616699, 0.024381568908691406, 0.02411087989807129, 0.02405958366394043, 0.02459123229980469, 0.02418057632446289, 0.023957344055175783, 0.0242238712310791, 0.023959264755249024, 0.023957984924316406, 0.024258560180664062, 0.024065280914306642, 0.02394598388671875, 0.02390220832824707, 0.024432031631469727, 0.02386390495300293, 0.023985727310180664, 0.02395180892944336, 0.0240762882232666, 0.023999616622924803, 0.024478591918945313, 0.02395136070251465, 0.023988224029541014, 0.024025087356567384, 0.02523574447631836, 0.024180736541748047, 0.024270048141479494, 0.024027711868286134, 0.024013248443603516, 0.023899616241455077, 0.023982175827026365, 0.0239550724029541, 0.023888479232788085, 0.023863296508789062, 0.023875072479248048, 0.023798271179199217, 0.023946943283081053, 0.02394144058227539, 0.023910400390625, 0.02384396743774414, 0.023865440368652343, 0.02390505599975586, 0.023767040252685546, 0.023971839904785155, 0.02389401626586914, 0.02388902473449707, 0.0239270076751709, 0.023982751846313478, 0.0241727352142334, 0.023971647262573243, 0.023996448516845702, 0.02393235206604004, 0.025938623428344725, 0.03129737663269043, 0.024203264236450195, 0.0240797119140625, 0.024040096282958983, 0.02405171203613281, 0.024016895294189454, 0.024041471481323243, 0.02432204818725586, 0.02409881591796875, 0.02388172721862793, 0.02395039939880371, 0.02413043212890625, 0.024162368774414064, 0.024020288467407228, 0.024008384704589845, 0.023972864151000976, 0.024771839141845702, 0.024029951095581054, 0.024035232543945313, 0.02398627281188965, 0.024049184799194337, 0.02389616012573242, 0.02400422477722168, 0.02392755126953125, 0.02388787269592285, 0.023879680633544922, 0.023819711685180663, 0.023806432723999023, 0.023890016555786132, 0.023881824493408203, 0.023822240829467774, 0.02384486389160156, 0.023875583648681642, 0.023781343460083006, 0.024661951065063477, 0.02415523147583008, 0.023929664611816406, 0.023981760025024414, 0.023880159378051758, 0.02388172721862793, 0.023914239883422853, 0.023894271850585937, 0.025006080627441408, 0.025540607452392578, 0.024194623947143553, 0.024285631179809572, 0.024288448333740234, 0.024421184539794923, 0.024290464401245118, 0.024243040084838866, 0.024319999694824217, 0.024401920318603516, 0.024444223403930664, 0.024357376098632814, 0.024331743240356446, 0.024316448211669922, 0.024284351348876954, 0.02413590431213379, 0.024032032012939453, 0.02403705596923828, 0.02401545524597168, 0.024106719970703124, 0.0242992000579834, 0.02433875274658203, 0.024200864791870117, 0.024187231063842775, 0.02405097579956055, 0.024062623977661134, 0.024021055221557615, 0.02398988723754883, 0.02405526351928711, 0.02404630470275879, 0.024106496810913085, 0.024039327621459963, 0.024050464630126955, 0.024444032669067382, 0.024861312866210936, 0.024549055099487304, 0.024170495986938476, 0.024344768524169922, 0.02402137565612793, 0.024030847549438475, 0.023945791244506836, 0.02398188781738281, 0.02409881591796875, 0.023868511199951172, 0.023910879135131836, 0.023968191146850587, 0.023910175323486327, 0.0238635196685791, 0.023843008041381834, 0.023881248474121094, 0.02388969612121582, 0.02396342468261719, 0.023967679977416993, 0.024373823165893555, 0.024400096893310547, 0.025037599563598634, 0.02434252738952637, 0.024104543685913086, 0.02399273681640625, 0.023805952072143553, 0.023795711517333985, 0.023838720321655273, 0.0238450870513916, 0.02375235176086426, 0.023771263122558593, 0.023748607635498048, 0.023893056869506837, 0.023845823287963867, 0.02390425682067871, 0.023867391586303712, 0.023910400390625, 0.023887519836425782, 0.023904672622680666, 0.02374550437927246, 0.023865728378295897, 0.0238372802734375, 0.023820287704467775, 0.023797760009765623, 0.023871488571166992, 0.023776512145996093, 0.02387830352783203, 0.02386089515686035, 0.02386966323852539, 0.02387990379333496, 0.023883775711059572, 0.023811296463012697, 0.02408732795715332, 0.02489753532409668, 0.024792224884033202, 0.02400489616394043, 0.024012928009033204, 0.0238353271484375, 0.024009504318237306, 0.024314847946166993, 0.024082111358642577, 0.023961151123046875, 0.024070911407470703, 0.024004608154296874, 0.024139776229858398, 0.023994367599487306, 0.02401033592224121, 0.024017311096191405, 0.02453651237487793, 0.024072416305541994, 0.02426915168762207, 0.025136415481567382, 0.024829856872558592, 0.024108928680419923, 0.024215583801269532, 0.02411123275756836, 0.024134431838989258, 0.02395683288574219, 0.024076959609985352, 0.023994367599487306, 0.024573951721191405, 0.023928672790527343, 0.024082592010498047, 0.02595840072631836, 0.025231231689453126, 0.02427587127685547, 0.024047264099121092, 0.023974239349365236, 0.02423583984375, 0.024006752014160155, 0.024269920349121094, 0.024208383560180666, 0.02425177574157715, 0.024348608016967775, 0.024535743713378907, 0.024481119155883788, 0.024506208419799804, 0.02461574363708496, 0.024510719299316405, 0.024512256622314453, 0.024389184951782227, 0.024426944732666017, 0.02427494430541992, 0.02423753547668457, 0.024297727584838866, 0.02437353515625, 0.02457804870605469, 0.024576000213623047, 0.024567808151245117, 0.024627199172973634, 0.024646751403808592, 0.024564544677734376, 0.024727039337158203, 0.02465564727783203, 0.02464851188659668, 0.02467430305480957, 0.024573951721191405, 0.02463942337036133, 0.024691808700561525, 0.025084896087646483, 0.024813568115234375, 0.0248623046875, 0.024753984451293946, 0.02459231948852539, 0.024691167831420897, 0.02471340751647949, 0.024805343627929688, 0.02457792091369629, 0.024616191864013672, 0.024382080078125, 0.024395103454589843, 0.024213855743408202, 0.024283744812011718, 0.024172544479370117, 0.024020992279052734, 0.024066047668457033, 0.024190975189208985, 0.024426496505737305, 0.024451040267944337, 0.02479865646362305, 0.024723455429077147, 0.024154720306396486, 0.02404915237426758, 0.023955968856811522, 0.024006015777587892, 0.02395724868774414, 0.02426969528198242, 0.025402496337890625, 0.024168703079223634, 0.024023456573486326, 0.023957727432250976, 0.02384931182861328, 0.0239835205078125, 0.024041631698608398, 0.024195167541503908, 0.023952991485595702, 0.02410655975341797, 0.023951583862304688, 0.02404416084289551, 0.02392268753051758, 0.02404153633117676, 0.024074176788330078, 0.024033279418945314, 0.023920064926147462, 0.024110912322998047, 0.02396236801147461, 0.024135679244995118, 0.024161663055419922, 0.02421824073791504, 0.024194656372070314, 0.024410175323486327, 0.024400224685668947, 0.024221696853637696, 0.024352767944335937, 0.02407164764404297, 0.024017055511474608, 0.024070016860961913, 0.023910560607910157, 0.02393516731262207, 0.024032512664794923, 0.024064287185668946, 0.023923328399658203, 0.024276992797851563, 0.024056831359863282, 0.024015615463256836, 0.024041887283325195, 0.024042720794677733, 0.02396633529663086, 0.02392185592651367, 0.023935712814331055, 0.02396771240234375, 0.02388800048828125, 0.023958816528320312, 0.024183519363403322, 0.023971839904785155, 0.023919872283935547, 0.02403753662109375, 0.023839328765869142, 0.023860544204711915, 0.023945920944213866, 0.024556800842285155, 0.024009471893310548, 0.023821727752685547, 0.02391481590270996, 0.023910688400268554, 0.023889759063720702, 0.02394268798828125, 0.02393948745727539, 0.024004831314086914, 0.02393017578125, 0.025045440673828124, 0.024164352416992187, 0.023965055465698243, 0.023824544906616212, 0.02384492874145508, 0.023867807388305663, 0.023881759643554688, 0.023895360946655272, 0.023855775833129884, 0.023842111587524414, 0.023841407775878905, 0.023791679382324217, 0.023770496368408202, 0.023788192749023437, 0.023853023529052733, 0.023788576126098634, 0.024034271240234373, 0.02399967956542969, 0.024130559921264647, 0.024102720260620117, 0.023945215225219727, 0.02400636863708496, 0.023875360488891603, 0.023956064224243165, 0.024002464294433593, 0.023928831100463867, 0.02430156707763672, 0.023930303573608397, 0.023829055786132813, 0.02386944007873535, 0.023868640899658202, 0.02385590362548828, 0.0238919677734375, 0.02386124801635742, 0.0238590087890625, 0.02391878318786621, 0.023826271057128905, 0.02403139114379883, 0.02386124801635742, 0.02385478401184082, 0.023795488357543946, 0.023883552551269532, 0.023878400802612304, 0.023971839904785155, 0.023998016357421874, 0.024240575790405273, 0.024137727737426756, 0.0241582088470459, 0.02418611145019531, 0.024285503387451172, 0.024474048614501955, 0.024768096923828125, 0.024366624832153322, 0.024578336715698243, 0.024262752532958985, 0.024297824859619142, 0.02407030487060547, 0.024270847320556642, 0.024152063369750978, 0.024129535675048826, 0.024141216278076173, 0.024322080612182616, 0.02425814437866211, 0.025225215911865235, 0.02427891159057617, 0.02406208038330078, 0.024195072174072265, 0.024039424896240235, 0.024052864074707032, 0.02407084846496582, 0.02406991958618164, 0.02401321601867676, 0.023971839904785155, 0.024002559661865236, 0.024012384414672853, 0.024060159683227538, 0.02419318389892578, 0.024059551239013672, 0.024103263854980468, 0.02405580711364746, 0.02406809616088867, 0.02408166313171387, 0.02398409652709961, 0.024009599685668945, 0.024022815704345703, 0.02404774475097656, 0.02395955276489258, 0.02400611114501953, 0.024010976791381835, 0.023981632232666014, 0.023974655151367187, 0.025268224716186522, 0.024215551376342775, 0.024002559661865236, 0.024027135848999022, 0.02394726371765137, 0.02400230407714844, 0.024015104293823242, 0.024199167251586915, 0.0241213436126709, 0.024051456451416015, 0.024036895751953124, 0.023890655517578126, 0.023997440338134765, 0.024024063110351563, 0.023977632522583007, 0.023982431411743162, 0.023926528930664062, 0.024018943786621092, 0.023971391677856446, 0.024080287933349608, 0.024869279861450197, 0.0242159366607666, 0.02393440055847168, 0.023945791244506836, 0.02486409568786621, 0.024090400695800783, 0.0239968318939209, 0.023941600799560547, 0.0239554557800293, 0.023963552474975586, 0.023971744537353516, 0.02395359992980957, 0.02393087959289551, 0.02397123146057129, 0.023879711151123046]",tokens/s,41.446513900252405,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,1,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,870.170624,617.545728,0.0,222.298112,199.93344,s,1,8.03411474609375,8.03411474609375,0.0,8.03411474609375,8.03411474609375,8.03411474609375,8.03411474609375,[8.03411474609375],,kWh,2.450999462917025e-05,2.69318842136144e-06,7.537506029997587e-06,3.474068908052927e-05,,MB,1211.572224,669.974528,0.0,262.144,220.881408,s,16,0.5669669075012207,0.035435431718826294,0.0010705433270812425,0.035248767852783205,0.03584315109252929,0.036712830543518066,0.03876446552276611,"[0.03927737426757812, 0.03503222274780273, 0.03528144073486328, 0.034322624206542966, 0.03507164764404297, 0.03585798263549805, 0.03582831954956055, 0.035390846252441405, 0.03559091186523437, 0.035216094970703125, 0.035463329315185546, 0.03534169769287109, 0.03486675262451172, 0.034920894622802734, 0.03455049514770508, 0.03495427322387695]",tokens/s,7224.407537385559,kWh,1.2342303559247405e-06,1.3611274231380592e-07,8.165512861518986e-07,2.186894384390445e-06,tokens/kWh,117060980.09454402,MB,1245.51168,684.654592,0.0,276.824064,220.883968,s,16,9.711272521972656,0.606954532623291,0.008585633852691918,0.6104183959960938,0.6147964477539063,0.6172033386230469,0.6172731384277343,"[0.5949608154296875, 0.58690283203125, 0.5928167114257813, 0.6113905639648437, 0.6124186401367188, 0.6122392578125, 0.6092311401367188, 0.6065200805664063, 0.6100323486328125, 0.6172905883789063, 0.6171742553710937, 0.610804443359375, 0.6068065795898437, 0.6120545043945312, 0.6117877807617188, 0.5988419799804687]",tokens/s,103.79690176743638,kWh,1.7168328469075072e-05,1.8933499684787224e-06,6.887743071723286e-06,2.594942150927708e-05,tokens/kWh,2427799.7864991752,,s,1008,9.702431542396553,0.009625428117456892,0.00026042284172965544,0.009625679969787599,0.009851100540161133,0.009908850908279419,0.01048333580970764,"[0.009447872161865234, 0.009883872032165527, 0.009592991828918457, 0.009680928230285645, 0.009769023895263672, 0.009662367820739747, 0.00961945629119873, 0.009543904304504395, 0.009411968231201173, 0.009510944366455078, 0.009316127777099609, 0.00942959976196289, 0.009633536338806153, 0.009889151573181152, 0.009661343574523926, 0.00935308837890625, 0.009390175819396973, 0.009297056198120117, 0.009340800285339356, 0.009318400382995605, 0.009380031585693359, 0.009345824241638184, 0.009279232025146484, 0.009408767700195313, 0.009334303855895997, 0.009332480430603028, 0.009302751541137695, 0.009256159782409668, 0.009249567985534668, 0.009538687705993653, 0.009282431602478027, 0.009288800239562989, 0.009294816017150879, 0.009297856330871582, 0.009354432106018067, 0.009312640190124512, 0.009260671615600586, 0.00941487979888916, 0.009314144134521484, 0.009253631591796875, 0.00919961643218994, 0.00927948760986328, 0.00922323226928711, 0.009255871772766114, 0.009237919807434082, 0.009220383644104003, 0.009230655670166016, 0.009252256393432617, 0.00960905647277832, 0.009428959846496582, 0.009360063552856445, 0.009312543869018555, 0.00996127986907959, 0.009870559692382812, 0.010033632278442383, 0.009816384315490723, 0.009544832229614258, 0.009667455673217773, 0.009404416084289552, 0.009328255653381348, 0.009349504470825195, 0.009328160285949707, 0.009316831588745118, 0.009525343894958497, 0.009607104301452637, 0.009431232452392578, 0.009420255661010742, 0.009408512115478516, 0.009412351608276368, 0.009341919898986816, 0.009263839721679688, 0.009296863555908204, 0.009326239585876465, 0.009684800148010254, 0.009287327766418456, 0.009227359771728515, 0.009227807998657227, 0.009212160110473632, 0.009289183616638184, 0.009218720436096192, 0.0091278076171875, 0.009142271995544434, 0.00932863998413086, 0.009203776359558106, 0.009279135704040528, 0.009216032028198241, 0.00928384017944336, 0.009214271545410156, 0.009309887886047363, 0.009207776069641113, 0.009220128059387206, 0.009185152053833008, 0.00923408031463623, 0.009265631675720214, 0.009371135711669922, 0.00929635238647461, 0.009306240081787109, 0.009285504341125489, 0.009322527885437012, 0.00929587173461914, 0.009286751747131347, 0.009294848442077636, 0.009238431930541992, 0.009256095886230468, 0.009261919975280761, 0.009430591583251952, 0.009429439544677735, 0.009507871627807618, 0.009454784393310546, 0.009492256164550782, 0.009390080451965332, 0.00931552028656006, 0.009323328018188477, 0.009362848281860351, 0.009230912208557129, 0.009256735801696777, 0.00925107192993164, 0.00942899227142334, 0.009298015594482421, 0.009239935874938964, 0.009159199714660645, 0.009328512191772461, 0.009303232192993165, 0.009180095672607422, 0.009190591812133789, 0.009220064163208008, 0.009267071723937988, 0.00943561553955078, 0.00954643154144287, 0.009376735687255859, 0.009472224235534668, 0.009382335662841797, 0.009531392097473144, 0.009433024406433106, 0.009355680465698242, 0.009446335792541503, 0.009403103828430176, 0.009461759567260742, 0.009308095932006835, 0.009351231575012208, 0.009328479766845704, 0.009283743858337402, 0.00929097557067871, 0.00927619171142578, 0.009274911880493165, 0.009390432357788086, 0.009349247932434083, 0.009261055946350098, 0.009506239891052246, 0.009422464370727539, 0.009231040000915527, 0.00928998374938965, 0.009314016342163086, 0.009215583801269531, 0.009251744270324706, 0.00929155158996582, 0.009220095634460449, 0.009334272384643554, 0.009203328132629394, 0.009170975685119629, 0.009195648193359374, 0.009214688301086425, 0.009293791770935058, 0.009230367660522461, 0.00923852825164795, 0.00927948760986328, 0.00937936019897461, 0.009452159881591798, 0.009427840232849122, 0.009343968391418458, 0.009350848197937012, 0.009389408111572265, 0.009313152313232423, 0.009361311912536622, 0.009377984046936036, 0.009385215759277343, 0.00963167953491211, 0.00966329574584961, 0.009664544105529785, 0.009652192115783692, 0.009612480163574218, 0.009548319816589355, 0.00981430435180664, 0.009943039894104003, 0.009624768257141113, 0.009614144325256348, 0.009523200035095216, 0.009523200035095216, 0.009533439636230469, 0.009478624343872071, 0.009737215995788574, 0.009750528335571289, 0.00968832015991211, 0.009748224258422852, 0.009639936447143555, 0.009650176048278808, 0.009576448440551758, 0.009660415649414063, 0.009524736404418945, 0.009575967788696288, 0.009534720420837403, 0.009532416343688965, 0.009501407623291016, 0.009538816452026368, 0.009546496391296387, 0.009613311767578125, 0.009697343826293945, 0.009756608009338379, 0.009654272079467773, 0.0097259521484375, 0.009699328422546387, 0.009706720352172852, 0.009686112403869629, 0.00983420753479004, 0.009747584342956543, 0.009736063957214356, 0.009894880294799804, 0.009772959709167481, 0.009694656372070313, 0.009681568145751954, 0.00976460838317871, 0.009732352256774903, 0.009707200050354003, 0.009885536193847657, 0.009771391868591309, 0.009736384391784668, 0.009703328132629394, 0.009703424453735352, 0.009764415740966797, 0.009817983627319336, 0.009847359657287598, 0.009705471992492675, 0.009762080192565917, 0.009691871643066407, 0.009650176048278808, 0.009794560432434082, 0.00993177604675293, 0.009791744232177735, 0.009791199684143066, 0.00973417568206787, 0.009665920257568359, 0.009630335807800292, 0.009607135772705078, 0.009599007606506348, 0.009582592010498046, 0.009631744384765625, 0.009594335556030273, 0.009579039573669434, 0.00962179183959961, 0.009606207847595215, 0.009966367721557618, 0.009864224433898925, 0.009728768348693847, 0.009780991554260253, 0.009867584228515625, 0.009820032119750976, 0.009682815551757812, 0.009683072090148926, 0.00966659164428711, 0.009727775573730469, 0.009598496437072754, 0.009607839584350585, 0.00975158405303955, 0.009662848472595215, 0.009642144203186036, 0.009896384239196778, 0.009752351760864258, 0.009782976150512695, 0.009587231636047364, 0.009539584159851074, 0.009578207969665527, 0.009640064239501953, 0.0097609281539917, 0.009832448005676269, 0.009678848266601562, 0.009674143791198731, 0.009722463607788086, 0.009668352127075196, 0.00967091178894043, 0.00967676830291748, 0.009576479911804199, 0.009721440315246582, 0.009687456130981445, 0.009823583602905273, 0.009754847526550293, 0.009850496292114258, 0.009798175811767578, 0.009890080451965332, 0.00964310359954834, 0.00962384033203125, 0.009656959533691407, 0.009654080390930176, 0.00962390422821045, 0.009645503997802734, 0.009607808113098144, 0.009596128463745118, 0.009746912002563477, 0.009666655540466309, 0.009727935791015625, 0.009897791862487793, 0.009782655715942383, 0.010109855651855468, 0.009703392028808594, 0.00971776008605957, 0.009703424453735352, 0.009557567596435546, 0.00975648021697998, 0.009560288429260254, 0.009544351577758788, 0.009524991989135742, 0.009566399574279785, 0.009518912315368652, 0.010208415985107421, 0.009755647659301758, 0.009953120231628417, 0.009440064430236816, 0.009772319793701172, 0.009629504203796387, 0.00984768009185791, 0.009584287643432617, 0.009733728408813477, 0.009590975761413574, 0.009642720222473145, 0.009518688201904296, 0.00959494400024414, 0.009668319702148438, 0.009644512176513673, 0.009582240104675293, 0.009628000259399415, 0.009624896049499511, 0.00961196804046631, 0.009709280014038086, 0.009664799690246582, 0.009739999771118163, 0.009668512344360352, 0.009656512260437011, 0.00959280014038086, 0.009611392021179199, 0.009633888244628906, 0.009545727729797364, 0.009674752235412597, 0.009637920379638672, 0.00965833568572998, 0.009666463851928712, 0.00971292781829834, 0.00967683219909668, 0.009672896385192871, 0.010103391647338868, 0.009857088088989259, 0.009727295875549316, 0.00967347240447998, 0.00975004768371582, 0.009642335891723633, 0.009559295654296874, 0.010099103927612305, 0.010545503616333008, 0.010921983718872071, 0.010004480361938477, 0.00971776008605957, 0.009684831619262695, 0.009625760078430176, 0.009709695816040038, 0.009615232467651367, 0.009707200050354003, 0.009793536186218262, 0.009676223754882812, 0.00973299217224121, 0.00988479995727539, 0.009646719932556152, 0.009568511962890625, 0.009506815910339356, 0.009621503829956055, 0.00954543972015381, 0.009546079635620118, 0.009545663833618163, 0.009560064315795898, 0.009516863822937012, 0.009967807769775391, 0.009519776344299317, 0.009664511680603028, 0.0096112642288208, 0.010798912048339843, 0.012132543563842774, 0.009691455841064452, 0.009690912246704101, 0.009564224243164063, 0.009650015830993652, 0.009531328201293945, 0.009705535888671875, 0.009598496437072754, 0.009545472145080567, 0.009618368148803711, 0.009574175834655762, 0.009661888122558594, 0.009585151672363281, 0.009570367813110351, 0.009614527702331543, 0.009620287895202636, 0.009600223541259766, 0.009645024299621581, 0.009611071586608887, 0.009715392112731933, 0.009715423583984376, 0.009599583625793457, 0.009594592094421386, 0.009596384048461915, 0.009569087982177734, 0.00952457618713379, 0.009534367561340332, 0.009571423530578613, 0.009612959861755371, 0.00976319980621338, 0.009531744003295898, 0.009554176330566407, 0.009568287849426269, 0.009592831611633301, 0.009661472320556641, 0.009608160018920899, 0.009574399948120118, 0.009570303916931153, 0.00958182430267334, 0.009619263648986817, 0.009495231628417969, 0.00952575969696045, 0.009541376113891602, 0.00971776008605957, 0.009644031524658203, 0.0095862398147583, 0.009658559799194336, 0.009570143699645995, 0.009585344314575195, 0.0095250883102417, 0.009553119659423829, 0.009638079643249512, 0.009620032310485839, 0.009583871841430664, 0.00970748805999756, 0.009548480033874512, 0.00951734447479248, 0.009573760032653809, 0.009591039657592773, 0.009543680191040039, 0.00970956802368164, 0.009793248176574707, 0.010053919792175294, 0.009944448471069336, 0.00968569564819336, 0.009922623634338379, 0.009713536262512206, 0.009567392349243165, 0.009699616432189941, 0.009695743560791016, 0.009801792144775391, 0.00960102367401123, 0.009666432380676269, 0.009656671524047852, 0.009633888244628906, 0.009620256423950195, 0.00968387222290039, 0.009682656288146973, 0.009783583641052246, 0.009635583877563476, 0.009656479835510254, 0.009541279792785644, 0.009611712455749512, 0.009586688041687011, 0.009625311851501464, 0.009572640419006347, 0.009660415649414063, 0.009564160346984863, 0.009552895545959473, 0.00953990364074707, 0.00965283203125, 0.009556384086608886, 0.009648927688598633, 0.009597855567932129, 0.009565728187561035, 0.009482815742492675, 0.00955401611328125, 0.009505951881408691, 0.009536160469055175, 0.00951910400390625, 0.009491840362548827, 0.009509568214416503, 0.009482111930847168, 0.009525376319885254, 0.00952086353302002, 0.009519328117370605, 0.00951910400390625, 0.009558015823364258, 0.009566176414489745, 0.009531295776367188, 0.009578623771667481, 0.009557663917541504, 0.009505120277404785, 0.009473983764648437, 0.009491552352905273, 0.009593184471130372, 0.009656096458435058, 0.009792351722717285, 0.009664511680603028, 0.009590784072875976, 0.009630816459655762, 0.009575615882873536, 0.009500320434570313, 0.009721152305603028, 0.009724703788757325, 0.009793536186218262, 0.00961740779876709, 0.00970537567138672, 0.009683296203613282, 0.009624959945678711, 0.00970911979675293, 0.009576671600341797, 0.009623647689819336, 0.009660479545593261, 0.009537631988525391, 0.009567935943603516, 0.009535967826843262, 0.00962559986114502, 0.009785247802734374, 0.009834464073181153, 0.009742655754089356, 0.009690688133239747, 0.009636287689208984, 0.009652095794677734, 0.009656448364257813, 0.009643296241760253, 0.009556703567504883, 0.009797632217407226, 0.009616543769836425, 0.009657183647155761, 0.009613311767578125, 0.009576448440551758, 0.009585760116577148, 0.009597855567932129, 0.00960048007965088, 0.009615903854370117, 0.009562111854553223, 0.009595135688781738, 0.009604063987731934, 0.009709856033325196, 0.009588768005371093, 0.009638208389282227, 0.009702560424804687, 0.009933440208435059, 0.009869215965270996, 0.009767231941223144, 0.009719615936279296, 0.009688575744628907, 0.009654879570007324, 0.00981606388092041, 0.00961945629119873, 0.009653696060180664, 0.009635711669921875, 0.009644991874694824, 0.009623552322387695, 0.009682080268859864, 0.009741151809692383, 0.009689087867736817, 0.009725664138793945, 0.00982806396484375, 0.009787967681884766, 0.009786944389343261, 0.009730496406555177, 0.00972390365600586, 0.009674976348876953, 0.00956287956237793, 0.009942815780639649, 0.00992899227142334, 0.010126272201538085, 0.010017056465148926, 0.009740544319152831, 0.00986460781097412, 0.00966649627685547, 0.009731040000915528, 0.009623488426208497, 0.00972390365600586, 0.00966864013671875, 0.009692768096923828, 0.009736576080322266, 0.009637887954711915, 0.009670656204223632, 0.009690943717956544, 0.009646464347839355, 0.00976467227935791, 0.010141695976257324, 0.009758720397949219, 0.009764863967895507, 0.009704607963562012, 0.009705984115600585, 0.009733983993530274, 0.009763487815856934, 0.009772480010986328, 0.009728416442871094, 0.009709856033325196, 0.009919391632080079, 0.009644576072692871, 0.00964031982421875, 0.009716799736022949, 0.00967353630065918, 0.00975376033782959, 0.009779775619506836, 0.009780768394470216, 0.00971673583984375, 0.009682720184326171, 0.009842880249023437, 0.010102720260620117, 0.009794943809509278, 0.009883999824523926, 0.009904671669006347, 0.009862784385681153, 0.009801728248596191, 0.00979155158996582, 0.009877440452575683, 0.009789440155029297, 0.00987110424041748, 0.009892352104187012, 0.00977894401550293, 0.009884960174560547, 0.009851615905761718, 0.009746527671813965, 0.009746335983276367, 0.009867551803588868, 0.00989087963104248, 0.009789728164672851, 0.009808256149291992, 0.009758015632629395, 0.009848640441894532, 0.009743231773376465, 0.009546112060546875, 0.009905983924865722, 0.009841312408447265, 0.009899423599243165, 0.009830080032348632, 0.009835200309753418, 0.009769503593444823, 0.009813055992126465, 0.00956287956237793, 0.009850655555725098, 0.009641823768615722, 0.009960960388183594, 0.009818943977355957, 0.00989132785797119, 0.009744256019592286, 0.009803647994995118, 0.009762847900390625, 0.009863519668579101, 0.00974448013305664, 0.009750335693359376, 0.009734560012817382, 0.009750528335571289, 0.009672639846801757, 0.009715456008911133, 0.009760640144348144, 0.009836480140686036, 0.009687423706054688, 0.009940223693847656, 0.00968291187286377, 0.009702431678771972, 0.009680000305175782, 0.009728768348693847, 0.009756704330444337, 0.00978940773010254, 0.009752575874328612, 0.00971116828918457, 0.00971731185913086, 0.009851776123046874, 0.00975692844390869, 0.009752320289611817, 0.009836799621582031, 0.00984447956085205, 0.00993660831451416, 0.009849120140075683, 0.009821279525756836, 0.009927488327026368, 0.009797727584838867, 0.00990835189819336, 0.00979475212097168, 0.009806879997253418, 0.009747551918029786, 0.009840288162231446, 0.009851807594299317, 0.009901856422424317, 0.00985324764251709, 0.009880576133728027, 0.009890527725219726, 0.009778495788574218, 0.009818431854248046, 0.009611840248107911, 0.009684991836547852, 0.00968892765045166, 0.009718048095703124, 0.00955571174621582, 0.010055935859680176, 0.009907839775085449, 0.009775456428527831, 0.00981123161315918, 0.009704192161560059, 0.009858783721923827, 0.009752863883972169, 0.009902239799499512, 0.01001046371459961, 0.009804032325744628, 0.009832127571105957, 0.00969324779510498, 0.009701215744018554, 0.009683327674865723, 0.009790623664855957, 0.009710495948791503, 0.009688384056091308, 0.009709983825683595, 0.009578783988952637, 0.009635552406311036, 0.009619135856628418, 0.009787712097167969, 0.009596927642822266, 0.009591103553771973, 0.009625280380249023, 0.009691360473632813, 0.009634655952453613, 0.009638079643249512, 0.009618176460266113, 0.009539135932922363, 0.009559935569763183, 0.010627103805541991, 0.009675295829772949, 0.009717887878417968, 0.009706687927246094, 0.009699999809265137, 0.009649727821350098, 0.009739871978759765, 0.00962060832977295, 0.009592576026916504, 0.009613311767578125, 0.009567328453063965, 0.009614239692687989, 0.009545408248901367, 0.009650495529174805, 0.009576255798339844, 0.00950819206237793, 0.00986182403564453, 0.009847999572753906, 0.009671648025512696, 0.009841952323913574, 0.00962435245513916, 0.00966211223602295, 0.009511199951171875, 0.009535264015197754, 0.009478367805480958, 0.009441280364990234, 0.009464863777160645, 0.009521216392517089, 0.009451583862304687, 0.009549920082092284, 0.009616127967834473, 0.00946793556213379, 0.00981049633026123, 0.009789055824279785, 0.009763520240783691, 0.00969696044921875, 0.009662336349487304, 0.010201215744018555, 0.01009059238433838, 0.009874367713928222, 0.009585023880004883, 0.009709407806396484, 0.009634655952453613, 0.009622688293457032, 0.009713408470153808, 0.009616512298583985, 0.009625472068786622, 0.009566335678100587, 0.00954150390625, 0.009667712211608887, 0.009665696144104004, 0.009688799858093261, 0.009669952392578125, 0.009570207595825195, 0.009588607788085938, 0.009584832191467284, 0.009519424438476563, 0.00958505630493164, 0.009573760032653809, 0.009697055816650391, 0.009556832313537598, 0.00960102367401123, 0.009555968284606933, 0.009668288230895996, 0.009566816329956054, 0.0096212158203125, 0.009615039825439453, 0.009517472267150879, 0.009524127960205079, 0.009548800468444824, 0.0094967041015625, 0.009534560203552245, 0.00950761604309082, 0.009504768371582031, 0.009524543762207032, 0.009503487586975098, 0.009540896415710449, 0.009869983673095703, 0.009561152458190917, 0.00952620792388916, 0.00961945629119873, 0.009545727729797364, 0.009526816368103028, 0.009525823593139648, 0.0095763521194458, 0.009511967658996583, 0.00963811206817627, 0.009548416137695312, 0.009611071586608887, 0.009576416015625, 0.009695775985717773, 0.00957420825958252, 0.009566240310668946, 0.009603039741516113, 0.00952768039703369, 0.009738719940185547, 0.009861120223999023, 0.009818112373352051, 0.009744383811950684, 0.00962559986114502, 0.009682239532470703, 0.00958672046661377, 0.009664959907531738, 0.009726335525512695, 0.009683903694152831, 0.009587807655334473, 0.009512672424316406, 0.009642080307006836, 0.009613311767578125, 0.009621503829956055, 0.009776255607604981, 0.009474368095397949, 0.009523776054382324, 0.009672736167907715, 0.009891807556152344, 0.009889792442321778, 0.009649984359741211, 0.009689279556274414, 0.009746432304382324, 0.009663711547851563, 0.009643808364868164, 0.009675775527954102, 0.009602656364440918, 0.009586943626403808, 0.009820320129394532, 0.009684703826904296, 0.009560288429260254, 0.009746560096740723, 0.009849023818969727, 0.009887264251708985, 0.0097193603515625, 0.009724575996398926, 0.009797792434692382, 0.009717599868774414, 0.009834495544433594, 0.009754143714904785, 0.009795104026794434, 0.009737152099609375, 0.009885567665100098, 0.009764991760253906, 0.0097609281539917, 0.009752415657043456, 0.009756352424621582, 0.009711487770080567, 0.009693792343139648, 0.009682784080505371, 0.00974847984313965, 0.009674688339233399, 0.009621888160705567, 0.00977683162689209, 0.009731679916381837, 0.009638015747070312, 0.009693375587463379, 0.009630047798156738, 0.00963161563873291, 0.00981174373626709, 0.00974847984313965, 0.00959718418121338, 0.009842720031738281, 0.009944992065429687, 0.009842687606811524, 0.009942432403564454, 0.009909119606018067, 0.009858688354492188, 0.009818207740783692, 0.009850879669189454, 0.009889535903930664, 0.00989414405822754, 0.009795871734619141, 0.010454527854919434, 0.010950655937194824, 0.010774751663208008, 0.01005568027496338, 0.01006175994873047, 0.009836607933044434, 0.009920512199401856, 0.009981951713562011, 0.009844736099243164, 0.009824352264404297, 0.009877408027648926, 0.009795583724975587, 0.009773216247558593, 0.009760607719421386, 0.009688639640808105, 0.009607872009277344, 0.009682687759399414, 0.009739359855651856, 0.009507871627807618, 0.009461503982543945, 0.009545856475830079, 0.009477343559265137, 0.009415488243103027, 0.009435104370117188, 0.009422656059265136, 0.009462271690368652, 0.009432191848754884, 0.009552448272705079, 0.009549247741699218, 0.009648927688598633, 0.009645855903625489, 0.00963584041595459, 0.009586432456970214, 0.009662943840026856, 0.009497695922851563, 0.00947270393371582, 0.009524255752563477, 0.009488991737365723, 0.009594592094421386, 0.009447263717651368, 0.009540703773498535, 0.009443008422851562, 0.009438976287841797, 0.009400320053100587, 0.009470239639282226, 0.009656319618225098, 0.009555968284606933, 0.00939417552947998, 0.009334783554077148, 0.00944547176361084, 0.009328543663024902, 0.009187328338623046, 0.009523200035095216, 0.009322784423828124, 0.009346495628356934, 0.009333215713500976, 0.009364319801330566, 0.00942409610748291, 0.009347071647644043, 0.00953916835784912, 0.009448863983154298, 0.009579263687133788, 0.009508864402770996, 0.01324169635772705, 0.010420576095581054, 0.010524991989135743, 0.010485504150390625, 0.009701631546020507, 0.009324543952941895, 0.009377216339111328, 0.009525823593139648, 0.00939417552947998, 0.00928553581237793, 0.009527392387390136, 0.00930611228942871, 0.009223648071289062, 0.009226911544799805, 0.009233311653137207, 0.009210047721862792, 0.009220864295959473, 0.009220128059387206, 0.009224320411682129, 0.00934284782409668, 0.009584639549255371, 0.009237855911254882, 0.009636480331420899, 0.009351200103759765, 0.009354720115661621, 0.00937564754486084, 0.009327232360839844, 0.00932585620880127, 0.009260831832885742, 0.00932755184173584, 0.009272480010986328, 0.0093274564743042, 0.009308159828186035, 0.009255231857299805, 0.009359040260314942, 0.00935321617126465, 0.009248543739318847, 0.009302240371704101, 0.00936297607421875, 0.010651712417602538, 0.009429408073425292, 0.00934928035736084, 0.009588031768798829, 0.009392767906188966, 0.009439167976379394, 0.009336064338684082, 0.009327327728271485, 0.009303872108459472, 0.009316512107849121, 0.009338560104370118, 0.009338368415832519]",tokens/s,103.8914828303977,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4260.343808,4725.866496,0.0,4347.396096,4328.833024,s,1,10.356025390625,10.356025390625,0.0,10.356025390625,10.356025390625,10.356025390625,10.356025390625,[10.356025390625],,kWh,9.157256575830767e-05,1.0093641610077022e-05,3.0194468599975544e-05,0.00013186067596836025,,MB,1385.377792,5220.794368,0.0,4812.96384,4756.928512,s,10,3.487279876708985,0.3487279876708984,0.0018193859702139725,0.3491572570800781,0.3502679779052734,0.3507423263549805,0.3511218051147461,"[0.3448252868652344, 0.348823486328125, 0.34959466552734375, 0.34949102783203123, 0.3461040649414063, 0.3501625671386719, 0.3486292724609375, 0.3485438537597656, 0.34988897705078126, 0.3512166748046875]",tokens/s,734.0965137607261,kWh,1.0162405058045626e-05,1.1207336724002644e-06,6.739708456896455e-06,1.8022847187342344e-05,tokens/kWh,14204193.007850157,MB,1385.377792,5325.651968,0.0,4917.82144,4876.091904,s,10,20.489777587890625,2.0489777587890625,0.008619845570690534,2.0503623046875,2.05667080078125,2.0608374755859376,2.0641708154296876,"[2.03930029296875, 2.045240234375, 2.053543212890625, 2.049490478515625, 2.0404039306640627, 2.055744873046875, 2.05490625, 2.051234130859375, 2.065004150390625, 2.0349100341796875]",tokens/s,30.74703945895086,kWh,5.958664198445075e-05,6.572193137807543e-06,3.9495856309304044e-05,0.00010565469143156238,tokens/kWh,596282.0878693127,,s,630,20.486324972152726,0.03251797614627414,0.0007612176222767933,0.03237366485595703,0.03315205459594727,0.03342379589080811,0.03613603931427002,"[0.035639007568359374, 0.03324729537963867, 0.03235619354248047, 0.0322116813659668, 0.031927967071533205, 0.03177846336364746, 0.03207443237304688, 0.038157535552978517, 0.03178780746459961, 0.032153472900390626, 0.03230047988891602, 0.03191263961791992, 0.031971168518066403, 0.03236470413208008, 0.032218753814697264, 0.032127361297607425, 0.03240140914916992, 0.03218227386474609, 0.03200579071044922, 0.03203855895996094, 0.03194540786743164, 0.032053249359130856, 0.032290241241455075, 0.03203129577636719, 0.031835487365722656, 0.03188751983642578, 0.03196678352355957, 0.031987808227539063, 0.03245142364501953, 0.032950271606445314, 0.032727039337158204, 0.03289907073974609, 0.032745216369628904, 0.032471294403076174, 0.03255839920043945, 0.03265811157226563, 0.03229062271118164, 0.03224588775634766, 0.031983680725097656, 0.031940160751342775, 0.03215385437011719, 0.03196323204040527, 0.0320096321105957, 0.03179737663269043, 0.03211443328857422, 0.03211663818359375, 0.03205974578857422, 0.0320181770324707, 0.031938655853271485, 0.031979936599731446, 0.031891008377075196, 0.0320103988647461, 0.03204092788696289, 0.032134944915771485, 0.032299488067626954, 0.03216569519042969, 0.03223961639404297, 0.032168510437011716, 0.03228579330444336, 0.03259894561767578, 0.03264681625366211, 0.03253692626953125, 0.033017856597900394, 0.035425792694091796, 0.03315126419067383, 0.032007518768310546, 0.031842527389526365, 0.03201500701904297, 0.031959039688110355, 0.0318791675567627, 0.03249894332885742, 0.03259878540039062, 0.03222732925415039, 0.03187302398681641, 0.03175820732116699, 0.03207385635375976, 0.03207916641235352, 0.03220755386352539, 0.031894559860229495, 0.03235939025878906, 0.03206134414672852, 0.03217622375488281, 0.03223952102661133, 0.03214140701293945, 0.03227347183227539, 0.03219488143920898, 0.03218691253662109, 0.03237836837768555, 0.03206310272216797, 0.032183040618896486, 0.03219068908691406, 0.03306905746459961, 0.03334348678588867, 0.03278643035888672, 0.03265945434570312, 0.032716384887695314, 0.03251446533203125, 0.03231692886352539, 0.032395713806152346, 0.032114273071289064, 0.032653568267822265, 0.0324876480102539, 0.032493057250976565, 0.03252191925048828, 0.032672542572021485, 0.03240127944946289, 0.032382495880126955, 0.03235084915161133, 0.03223484802246094, 0.03234064102172852, 0.03281919860839844, 0.03272268676757813, 0.03274361419677734, 0.03265478515625, 0.03228860855102539, 0.03286505508422852, 0.03277113723754883, 0.03234624099731445, 0.032176959991455076, 0.03237459182739258, 0.03283779144287109, 0.03331689453125, 0.03285385513305664, 0.032729248046875, 0.03273519897460937, 0.03229699325561523, 0.036216320037841795, 0.03335014343261719, 0.032449569702148434, 0.03188607978820801, 0.03203299331665039, 0.03206553649902344, 0.0321638412475586, 0.03231462478637695, 0.0322790412902832, 0.032076030731201174, 0.032081504821777344, 0.03213059234619141, 0.034144382476806644, 0.03181849670410156, 0.03162521553039551, 0.031813503265380856, 0.03175436782836914, 0.03195199966430664, 0.032059391021728514, 0.03244940948486328, 0.032314910888671874, 0.03220323181152344, 0.032518207550048826, 0.032355422973632815, 0.03227267074584961, 0.03243040084838867, 0.0326003189086914, 0.03226947021484375, 0.03257987213134766, 0.03322118377685547, 0.033147998809814457, 0.033108318328857425, 0.03282796859741211, 0.032640033721923825, 0.032588768005371097, 0.032571392059326174, 0.03272000122070313, 0.032647998809814456, 0.032760929107666016, 0.03252691268920899, 0.032745376586914066, 0.03246950531005859, 0.032217086791992186, 0.032408737182617185, 0.03241363143920899, 0.03241257476806641, 0.03234521484375, 0.03267264175415039, 0.032806495666503906, 0.032814849853515626, 0.032473758697509764, 0.03246633529663086, 0.032422496795654294, 0.03221651077270508, 0.03238355255126953, 0.033067008972167966, 0.03248537445068359, 0.03236422348022461, 0.032701759338378905, 0.032608543395996094, 0.0327154541015625, 0.03258780670166016, 0.03640019226074219, 0.03617670440673828, 0.03307014465332031, 0.032084926605224606, 0.031891359329223636, 0.03199337577819824, 0.03230352020263672, 0.032064865112304684, 0.032035648345947264, 0.032059200286865236, 0.03207187271118164, 0.03202252960205078, 0.032086017608642575, 0.032378814697265626, 0.03218406295776367, 0.03213036727905273, 0.03215631866455078, 0.03211043167114258, 0.03228313446044922, 0.03362748718261719, 0.03225872039794922, 0.031915456771850585, 0.03183059120178223, 0.03186892890930176, 0.03211801528930664, 0.032494335174560546, 0.03249478530883789, 0.032358623504638674, 0.03208195114135742, 0.03270265579223633, 0.033196414947509766, 0.033159168243408206, 0.032916576385498046, 0.03274844741821289, 0.032651329040527345, 0.03278841781616211, 0.032356128692626954, 0.03229718399047852, 0.03240755081176758, 0.03256057739257812, 0.032459327697753906, 0.03224496078491211, 0.03219475173950195, 0.032292865753173826, 0.03252694320678711, 0.032286113739013675, 0.03212758255004883, 0.03240719985961914, 0.03268262481689453, 0.03280831909179688, 0.032513729095458986, 0.03250038528442383, 0.03233116912841797, 0.03254742431640625, 0.03260009765625, 0.03280892944335938, 0.03297894287109375, 0.03275161743164062, 0.03304009628295899, 0.032790817260742185, 0.03301990509033203, 0.03304652786254883, 0.03303628921508789, 0.03317929458618164, 0.036000831604003906, 0.03352876663208008, 0.032307201385498044, 0.031748096466064454, 0.031510528564453126, 0.03144655990600586, 0.03179772758483887, 0.03117465591430664, 0.03122790336608887, 0.032984447479248044, 0.03163587188720703, 0.03195721626281738, 0.03164329528808594, 0.031582143783569334, 0.03153686332702637, 0.032276798248291015, 0.03213750457763672, 0.03157923126220703, 0.03129644775390625, 0.031760448455810546, 0.031336448669433595, 0.031862367630004884, 0.03181814384460449, 0.031987712860107424, 0.03219971084594726, 0.03261539077758789, 0.03251609420776367, 0.03260822296142578, 0.03261648178100586, 0.032952320098876955, 0.03275980758666992, 0.032702049255371096, 0.033231231689453126, 0.03488771057128906, 0.03258163070678711, 0.03257753753662109, 0.03253609466552734, 0.032317920684814455, 0.03211209487915039, 0.032112255096435546, 0.03261942291259766, 0.032712703704833986, 0.0324956169128418, 0.03257872009277344, 0.03242448043823242, 0.03235667037963867, 0.03214745712280274, 0.03205734252929687, 0.03197664070129395, 0.03206979370117188, 0.03225462341308594, 0.032086017608642575, 0.03261644744873047, 0.032546817779541014, 0.033538047790527346, 0.03271254348754883, 0.0328287353515625, 0.03276476669311523, 0.03270041656494141, 0.03247228622436524, 0.03309648132324219, 0.03280691146850586, 0.03278643035888672, 0.03558729553222656, 0.03322959899902344, 0.032276481628417966, 0.032389118194580076, 0.03228057479858398, 0.032091392517089846, 0.031989952087402344, 0.03200467300415039, 0.032389118194580076, 0.03226217651367187, 0.03208182525634766, 0.03210655975341797, 0.03250790405273438, 0.03230515289306641, 0.03210627365112305, 0.03242825698852539, 0.032421886444091795, 0.03213840103149414, 0.032282623291015625, 0.03243708801269531, 0.0331960334777832, 0.0320654411315918, 0.03222127914428711, 0.03213721466064453, 0.03237273788452148, 0.03253247833251953, 0.032847713470458985, 0.032379039764404295, 0.03313449478149414, 0.033359329223632814, 0.03326828765869141, 0.03303430557250977, 0.03283955383300781, 0.032718177795410155, 0.03332995223999023, 0.03215145492553711, 0.03215574264526367, 0.03267776107788086, 0.03277987289428711, 0.032562942504882814, 0.032344863891601565, 0.03238092803955078, 0.03248892974853516, 0.03265337753295899, 0.032647647857666016, 0.03263078308105469, 0.03280441665649414, 0.032544670104980467, 0.03239369583129883, 0.03253868865966797, 0.032689407348632814, 0.03297267150878906, 0.03238307189941406, 0.03244521713256836, 0.0322979850769043, 0.032484352111816404, 0.03284377670288086, 0.03269126510620117, 0.032938270568847655, 0.03301161575317383, 0.033407390594482424, 0.033417247772216795, 0.03325164794921875, 0.03603647994995117, 0.033675392150878905, 0.03233078384399414, 0.0320296630859375, 0.03201827239990234, 0.03215980911254883, 0.0320533447265625, 0.032003265380859375, 0.032578369140625, 0.03212492752075195, 0.03206470489501953, 0.03218719863891602, 0.03214950561523437, 0.03206553649902344, 0.032196094512939456, 0.032133632659912106, 0.0318831672668457, 0.032153182983398435, 0.03222169494628906, 0.03230825424194336, 0.03218940734863281, 0.03252019119262695, 0.032108062744140624, 0.03200048065185547, 0.03249094390869141, 0.03220912170410156, 0.03221286392211914, 0.03247296142578125, 0.033043006896972656, 0.03310163116455078, 0.03328841781616211, 0.034219551086425784, 0.03278211212158203, 0.03285676956176758, 0.03281100845336914, 0.03689471817016601, 0.03159859275817871, 0.03201836776733399, 0.032237632751464844, 0.032538623809814454, 0.03201958465576172, 0.03214134216308594, 0.03227939224243164, 0.03226828765869141, 0.03219660949707031, 0.032464897155761716, 0.03248332977294922, 0.032470497131347656, 0.0324101448059082, 0.03241104125976563, 0.03248393630981445, 0.032158912658691405, 0.0325715217590332, 0.03323875045776367, 0.032532543182373044, 0.032838558197021486, 0.0329543685913086, 0.03269222259521484, 0.03298476791381836, 0.033280223846435544, 0.03324518585205078, 0.033285343170166015, 0.03314499282836914, 0.03595043182373047, 0.033975105285644534, 0.03211673736572265, 0.03188121604919433, 0.031885120391845705, 0.03181366348266602, 0.031919872283935544, 0.03194083213806152, 0.03191007995605469, 0.03197068786621094, 0.032258689880371096, 0.03245619201660156, 0.032136959075927736, 0.032025344848632814, 0.032008190155029294, 0.03201638412475586, 0.03205734252929687, 0.032056671142578125, 0.03242598342895508, 0.032258655548095705, 0.033429153442382814, 0.03236495971679688, 0.03224166488647461, 0.03215359878540039, 0.03224099349975586, 0.032199329376220706, 0.03246457672119141, 0.032385345458984374, 0.0332410888671875, 0.03327721786499024, 0.034013919830322266, 0.03264716720581055, 0.032620449066162106, 0.03257676696777344, 0.03237974548339844, 0.03216588973999023, 0.03238489532470703, 0.03193187141418457, 0.03222390365600586, 0.03212895965576172, 0.03298720169067383, 0.03287859344482422, 0.03230681610107422, 0.032319774627685545, 0.03237827301025391, 0.03240620803833008, 0.03277004623413086, 0.032868350982666016, 0.03258752059936523, 0.03271500778198242, 0.03221820831298828, 0.032310176849365234, 0.032806495666503906, 0.032717216491699216, 0.03235977554321289, 0.03237542343139648, 0.032653343200683596, 0.03313663864135742, 0.032927745819091796, 0.03323904037475586, 0.033529857635498046, 0.03327164840698242, 0.03298073577880859, 0.03645027160644531, 0.0337955207824707, 0.032489601135253905, 0.03205286407470703, 0.03594118499755859, 0.037631999969482424, 0.03216793441772461, 0.033726463317871096, 0.033009536743164064, 0.03182985687255859, 0.031929664611816407, 0.03188412857055664, 0.03199193572998047, 0.03199180793762207, 0.03234739303588867, 0.0321984977722168, 0.032046142578125, 0.03237257766723633, 0.032304161071777346, 0.03211363220214844, 0.03212076950073242, 0.03345187377929688, 0.03211862564086914, 0.032371070861816405, 0.03244646453857422, 0.03209817504882812, 0.03226022338867188, 0.03344384002685547, 0.03246694564819336, 0.03518463897705078, 0.03279443359375, 0.032869792938232424, 0.03311286544799805, 0.03338649749755859, 0.032950271606445314, 0.03263033676147461, 0.03286675262451172, 0.03241984176635742, 0.032339744567871094, 0.032400638580322265, 0.03249151992797852, 0.03252080154418945, 0.03222566223144531, 0.03215151977539062, 0.03223759841918945, 0.03264716720581055, 0.03210841751098633, 0.03207539367675781, 0.0325923843383789, 0.032307201385498044, 0.032059391021728514, 0.03211673736572265, 0.03227033615112305, 0.0324073600769043, 0.03273747253417969, 0.032580894470214845, 0.03304726409912109, 0.03306480026245117, 0.03275788879394531, 0.03262262344360352, 0.0329273910522461, 0.03348515319824219, 0.03319919967651367, 0.035547359466552735, 0.03317132949829102, 0.032169952392578124, 0.0315221118927002, 0.0316376953125, 0.03196345520019531, 0.031922527313232425, 0.0319069766998291, 0.0316711368560791, 0.03173990440368652, 0.031639520645141604, 0.0315248966217041, 0.032161792755126956, 0.031906911849975586, 0.031734687805175785, 0.03183350372314453, 0.031590688705444334, 0.031833728790283206, 0.031697343826293946, 0.03163158416748047, 0.031862815856933596, 0.031882400512695315, 0.031881664276123045, 0.03183568000793457, 0.032314239501953126, 0.03250358581542969, 0.03237887954711914, 0.03252374267578125, 0.032979713439941404, 0.03331043243408203, 0.03328028869628906, 0.03300556945800781, 0.03278438568115234, 0.03269830322265625, 0.03214956665039063, 0.03203420639038086, 0.032610912322998044, 0.0324128303527832, 0.03209833526611328, 0.03172435188293457, 0.03209577560424805, 0.0321929931640625, 0.032221023559570315, 0.03263283157348633, 0.032438209533691406, 0.03223984146118164, 0.032282623291015625, 0.03219180679321289, 0.03225017547607422, 0.032180160522460935, 0.032311744689941406, 0.032200702667236326, 0.03219993591308594, 0.03231615829467773, 0.03236044692993164, 0.03227558517456055, 0.0324754867553711, 0.03280131149291992, 0.03289424133300781, 0.032798561096191406, 0.032682880401611325, 0.032785472869873045, 0.03274233627319336]",tokens/s,30.75222134064387,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1301.409792,1085.210624,0.0,706.740224,675.13344,s,1,8.295859375,8.295859375,0.0,8.295859375,8.295859375,8.295859375,8.295859375,[8.295859375],,kWh,3.7229761741673427e-05,4.099526151185418e-06,1.2231954230029762e-05,5.3561242122888603e-05,,MB,1362.333696,1403.977728,0.0,996.1472,942.731264,s,10,0.34586979675292967,0.034586979675292964,0.00027697715567381,0.03452019309997559,0.03478697814941406,0.035076289367675775,0.03530773834228515,"[0.0353656005859375, 0.03451801681518555, 0.03472268676757812, 0.034364673614501955, 0.034372161865234375, 0.034527904510498045, 0.03444595336914062, 0.03446873474121094, 0.03456169509887695, 0.034522369384765626]",tokens/s,7401.629237457595,kWh,1.0847740577603184e-06,1.1963046159972966e-07,7.20151340267668e-07,1.9245558596277156e-06,tokens/kWh,133017703.13360529,MB,1391.06304,1412.366336,0.0,1004.535808,942.733824,s,10,15.406380859375,1.5406380859375,0.00800279734248737,1.5395208740234376,1.5534908325195311,1.5542131286621095,1.554790965576172,"[1.546219482421875, 1.53852490234375, 1.553330322265625, 1.5330682373046876, 1.540516845703125, 1.533992919921875, 1.541059814453125, 1.5330107421875, 1.5549354248046876, 1.53172216796875]",tokens/s,40.89214759458813,kWh,4.483736391431977e-05,4.9451845287065935e-06,1.852830850293145e-05,6.831085694595781e-05,tokens/kWh,922254.5700142607,,s,630,15.40428860473634,0.02445125175354973,0.0004281711990985916,0.024329071998596193,0.02480334358215332,0.025083141994476314,0.02647872211456299,"[0.024335296630859374, 0.024533056259155275, 0.024349920272827147, 0.024212032318115233, 0.02451068878173828, 0.024208768844604493, 0.024222335815429687, 0.024162303924560546, 0.024139776229858398, 0.024217599868774413, 0.024276992797851563, 0.02467635154724121, 0.02411017608642578, 0.02521999931335449, 0.024410303115844727, 0.024257408142089844, 0.024265663146972656, 0.024383487701416014, 0.024696832656860353, 0.026015775680541992, 0.024395776748657227, 0.02436911964416504, 0.024882463455200194, 0.02453987121582031, 0.024286687850952147, 0.024384031295776366, 0.024383487701416014, 0.027291648864746092, 0.024642911911010743, 0.024470176696777344, 0.02450227165222168, 0.02457804870605469, 0.02433228874206543, 0.024303392410278322, 0.02434623908996582, 0.024631999969482423, 0.02830121612548828, 0.02460585594177246, 0.024445791244506836, 0.024629024505615233, 0.02432793617248535, 0.024920543670654296, 0.02435481643676758, 0.024175840377807616, 0.024875808715820312, 0.02430681610107422, 0.024295743942260743, 0.02430828857421875, 0.024426496505737305, 0.024205087661743164, 0.024407712936401368, 0.024449600219726562, 0.024186880111694335, 0.025133056640625, 0.02426838493347168, 0.02416681671142578, 0.0240611515045166, 0.024311872482299806, 0.02442313575744629, 0.024182783126831055, 0.024162303924560546, 0.024174591064453126, 0.024371200561523438, 0.02557267189025879, 0.024203968048095704, 0.024205631256103515, 0.024169696807861327, 0.024218080520629882, 0.024662015914916992, 0.0257392635345459, 0.024762367248535155, 0.024414207458496092, 0.024262208938598633, 0.024197536468505858, 0.024231744766235352, 0.02432431983947754, 0.024501632690429688, 0.024314495086669923, 0.024241504669189454, 0.024279712677001953, 0.02414918327331543, 0.024205312728881836, 0.024116031646728514, 0.02422127914428711, 0.02434908866882324, 0.02422163200378418, 0.024154176712036134, 0.024424448013305664, 0.024147968292236328, 0.024319999694824217, 0.02461075210571289, 0.02447273635864258, 0.02435369682312012, 0.025304927825927734, 0.024316064834594725, 0.02435686492919922, 0.02435465621948242, 0.02418636894226074, 0.02453356742858887, 0.024372543334960937, 0.024245023727416992, 0.024426496505737305, 0.02431551933288574, 0.024454944610595702, 0.02428169631958008, 0.02447769546508789, 0.02432204818725586, 0.024327999114990236, 0.024332479476928712, 0.024213504791259766, 0.02444697570800781, 0.024280351638793947, 0.024296159744262694, 0.02421116828918457, 0.02428303909301758, 0.024203647613525392, 0.024229440689086914, 0.024526687622070314, 0.024381824493408203, 0.024398208618164063, 0.024670047760009764, 0.024418272018432618, 0.024376480102539063, 0.02447654342651367, 0.02525526428222656, 0.024973983764648437, 0.02489139175415039, 0.024595712661743162, 0.024406784057617186, 0.02476032066345215, 0.02440950393676758, 0.02445782470703125, 0.024561119079589844, 0.02576361656188965, 0.024677120208740234, 0.0247459831237793, 0.024821632385253905, 0.024973247528076174, 0.02473798370361328, 0.02471731185913086, 0.024677696228027343, 0.028435136795043944, 0.02498150444030762, 0.024713216781616212, 0.024604671478271483, 0.024606719970703125, 0.02485606384277344, 0.024556032180786135, 0.024358495712280274, 0.0244553279876709, 0.02470528030395508, 0.024421407699584962, 0.024262912750244142, 0.024349407196044923, 0.024154111862182616, 0.024303615570068358, 0.024439807891845702, 0.02440083122253418, 0.024507551193237304, 0.025123136520385742, 0.02470694351196289, 0.02448588752746582, 0.02477743911743164, 0.024803327560424804, 0.0246824951171875, 0.0247293758392334, 0.02473187255859375, 0.024688640594482423, 0.024778751373291014, 0.024635391235351564, 0.024762367248535155, 0.02492144012451172, 0.024468128204345702, 0.025327680587768554, 0.024537248611450194, 0.024469440460205077, 0.024364896774291992, 0.024565759658813476, 0.02451456069946289, 0.02414543914794922, 0.02432252883911133, 0.024202367782592774, 0.02433113670349121, 0.024327680587768553, 0.0243023681640625, 0.024254112243652343, 0.024300928115844726, 0.024239904403686525, 0.024286111831665038, 0.02472483253479004, 0.02448044776916504, 0.024252384185791016, 0.025021663665771486, 0.024191776275634767, 0.02426470375061035, 0.024225791931152343, 0.02427903938293457, 0.024204448699951173, 0.024122047424316406, 0.024178848266601563, 0.02439936065673828, 0.024268991470336915, 0.024248159408569336, 0.024280767440795898, 0.024164255142211915, 0.024283903121948242, 0.024191104888916015, 0.024211456298828125, 0.024383071899414063, 0.02430544090270996, 0.024377119064331054, 0.024212032318115233, 0.024322336196899413, 0.025277856826782227, 0.02432195281982422, 0.024197824478149416, 0.024246271133422852, 0.02457708740234375, 0.02425040054321289, 0.02437603187561035, 0.024256704330444336, 0.02434377670288086, 0.024297887802124024, 0.024265087127685547, 0.024242176055908202, 0.02424831962585449, 0.02425651168823242, 0.02427289581298828, 0.024122655868530272, 0.024472288131713867, 0.02475609588623047, 0.024369279861450197, 0.024224895477294922, 0.024271743774414063, 0.024236032485961914, 0.02417990493774414, 0.024178592681884766, 0.0245134391784668, 0.024321983337402344, 0.024324159622192382, 0.02411929512023926, 0.024487295150756837, 0.025303680419921874, 0.024196704864501952, 0.024170047760009767, 0.024150815963745118, 0.02432931137084961, 0.02412544059753418, 0.02417967987060547, 0.024176639556884767, 0.024371200561523438, 0.024262655258178712, 0.024358047485351562, 0.02532643127441406, 0.024162559509277343, 0.02421855926513672, 0.02480415916442871, 0.02448918342590332, 0.024435232162475586, 0.02465203285217285, 0.024553247451782226, 0.02450444793701172, 0.024332256317138673, 0.024205440521240233, 0.024238079071044923, 0.0248152961730957, 0.02413759994506836, 0.024428991317749022, 0.024219648361206055, 0.024205055236816406, 0.0242193603515625, 0.024293920516967774, 0.024294559478759765, 0.024196992874145506, 0.024249120712280272, 0.02416249656677246, 0.02435686492919922, 0.02432204818725586, 0.024498176574707032, 0.024387840270996095, 0.024612607955932616, 0.02490108871459961, 0.024601119995117188, 0.02474393653869629, 0.024795135498046874, 0.024465408325195313, 0.024541088104248047, 0.02435286331176758, 0.024352767944335937, 0.024375295639038085, 0.02456166458129883, 0.0242872314453125, 0.024389631271362306, 0.02429465675354004, 0.02553891181945801, 0.024344959259033204, 0.024414016723632814, 0.024407424926757813, 0.025303903579711913, 0.024456607818603517, 0.02444758415222168, 0.02452889633178711, 0.02445919990539551, 0.024446271896362306, 0.02424515151977539, 0.02432534408569336, 0.024267295837402343, 0.024307039260864256, 0.02423222351074219, 0.024230367660522462, 0.024373247146606446, 0.024594432830810548, 0.02432953643798828, 0.024304319381713867, 0.024418304443359375, 0.024374847412109376, 0.024222143173217775, 0.02447065544128418, 0.02455567932128906, 0.02420195198059082, 0.024116640090942384, 0.02422649574279785, 0.0242457275390625, 0.0240928955078125, 0.024112768173217773, 0.02401545524597168, 0.024105024337768555, 0.024227136611938475, 0.024173183441162108, 0.024188159942626953, 0.024328832626342774, 0.024375104904174806, 0.024179264068603514, 0.024198911666870118, 0.024597503662109374, 0.0242445125579834, 0.02415279960632324, 0.024225791931152343, 0.02450227165222168, 0.0242457275390625, 0.0242806396484375, 0.024216415405273438, 0.024311935424804688, 0.024206783294677733, 0.02428166389465332, 0.024160287857055665, 0.024450464248657225, 0.024410688400268554, 0.024227840423583984, 0.02427903938293457, 0.02429747200012207, 0.02434377670288086, 0.026665279388427734, 0.024537055969238282, 0.02427136039733887, 0.02425651168823242, 0.024264415740966796, 0.024270559310913088, 0.024308191299438477, 0.024366975784301758, 0.024325759887695312, 0.024363616943359374, 0.024375295639038085, 0.024888736724853516, 0.024244735717773438, 0.024234079360961915, 0.024170495986938476, 0.024169759750366213, 0.024926944732666014, 0.02422684860229492, 0.024219648361206055, 0.024241119384765624, 0.02434169578552246, 0.024202047348022462, 0.02420240020751953, 0.025342687606811524, 0.024272991180419923, 0.02425551986694336, 0.02429974365234375, 0.024338111877441407, 0.024187040328979493, 0.024318111419677733, 0.02415350341796875, 0.02423222351074219, 0.02448953628540039, 0.024339168548583985, 0.02429955291748047, 0.024303455352783204, 0.024190656661987303, 0.024389120101928712, 0.025127904891967773, 0.024526847839355468, 0.02437238311767578, 0.024251136779785156, 0.0243438720703125, 0.02431161689758301, 0.024253536224365234, 0.024323680877685546, 0.02424860763549805, 0.024360960006713867, 0.02429305648803711, 0.02431622314453125, 0.024276992797851563, 0.024623104095458984, 0.024313312530517578, 0.024308256149291992, 0.024270847320556642, 0.02428927993774414, 0.02428108787536621, 0.024272991180419923, 0.024500127792358398, 0.024493440628051758, 0.024241952896118163, 0.024312671661376954, 0.025188352584838865, 0.024395776748657227, 0.024487743377685545, 0.024186912536621093, 0.024160415649414062, 0.027264352798461913, 0.024944320678710937, 0.024601247787475584, 0.024396064758300782, 0.024363040924072266, 0.024343616485595704, 0.024265056610107423, 0.024359519958496095, 0.024325471878051758, 0.024425119400024415, 0.024868864059448242, 0.024373247146606446, 0.0245166072845459, 0.024334272384643554, 0.02437126350402832, 0.02433228874206543, 0.02448307228088379, 0.02435148811340332, 0.02424831962585449, 0.02591948890686035, 0.024584192276000977, 0.02429132843017578, 0.02447657585144043, 0.024200576782226563, 0.0242346248626709, 0.024129535675048826, 0.02412748718261719, 0.024234048843383788, 0.024327680587768553, 0.024209760665893556, 0.02414806365966797, 0.02453913688659668, 0.024203264236450195, 0.024188928604125977, 0.0241810245513916, 0.024234815597534178, 0.024124319076538087, 0.024233503341674803, 0.024230367660522462, 0.024413440704345705, 0.024322656631469725, 0.02413590431213379, 0.024327104568481445, 0.024302560806274413, 0.02437721633911133, 0.024103071212768553, 0.024190975189208985, 0.024227840423583984, 0.024137184143066405, 0.024270847320556642, 0.024240671157836916, 0.025655296325683592, 0.02467430305480957, 0.0242872314453125, 0.024214752197265627, 0.024363008499145508, 0.024298271179199218, 0.0241778564453125, 0.02431862449645996, 0.024227487564086915, 0.024098304748535155, 0.024215999603271483, 0.02425913619995117, 0.024365055084228517, 0.024553216934204102, 0.02457961654663086, 0.024138015747070314, 0.024688735961914062, 0.02422400093078613, 0.02440755271911621, 0.024123104095458984, 0.024491104125976562, 0.02463417625427246, 0.02510710334777832, 0.024312128067016603, 0.02433945655822754, 0.024300128936767577, 0.024299936294555666, 0.02422524833679199, 0.024311904907226563, 0.024359359741210937, 0.024385311126708983, 0.024231967926025392, 0.02441206359863281, 0.024447263717651366, 0.0249466552734375, 0.0258023681640625, 0.02491644859313965, 0.025419200897216797, 0.02480348777770996, 0.02490131187438965, 0.024811807632446288, 0.02480073547363281, 0.02462175941467285, 0.02475859260559082, 0.024332128524780273, 0.024293535232543944, 0.024356191635131835, 0.02448246383666992, 0.024455039978027344, 0.02431398391723633, 0.024215551376342775, 0.024337823867797852, 0.024521312713623046, 0.025024511337280272, 0.02482348823547363, 0.025467199325561525, 0.02430771255493164, 0.024387584686279298, 0.024172544479370117, 0.02436848068237305, 0.025053855895996093, 0.026398815155029298, 0.02465337562561035, 0.024547679901123047, 0.0242872314453125, 0.024190240859985352, 0.02424275207519531, 0.02421798324584961, 0.024562976837158204, 0.024328096389770508, 0.024357471466064453, 0.02476851272583008, 0.02651136016845703, 0.024284671783447266, 0.02445158386230469, 0.02437446403503418, 0.02478163146972656, 0.024710432052612304, 0.024936479568481447, 0.024817695617675783, 0.025018463134765623, 0.02654982376098633, 0.024695871353149414, 0.02468217658996582, 0.024594240188598633, 0.024533439636230468, 0.024673376083374023, 0.024413087844848632, 0.024382495880126955, 0.02424496078491211, 0.02432758331298828, 0.024453983306884766, 0.024467456817626954, 0.02441334342956543, 0.024395711898803712, 0.02436102485656738, 0.0243986873626709, 0.024561695098876953, 0.024307680130004884, 0.024580127716064454, 0.024856351852416993, 0.0242587833404541, 0.024610815048217775, 0.02465177536010742, 0.02450022315979004, 0.024300735473632814, 0.02417251205444336, 0.024215967178344726, 0.024320159912109375, 0.024277280807495118, 0.02414918327331543, 0.02427587127685547, 0.02426652717590332, 0.024214784622192384, 0.024156864166259766, 0.024190271377563476, 0.024185728073120118, 0.024203264236450195, 0.024366207122802734, 0.024318527221679688, 0.024262975692749024, 0.024186880111694335, 0.024245664596557616, 0.024238208770751953, 0.024238208770751953, 0.024217504501342774, 0.024221376419067384, 0.02429567909240723, 0.0242359676361084, 0.024339008331298827, 0.02435481643676758, 0.024328384399414062, 0.024479551315307616, 0.024391679763793944, 0.0244715518951416, 0.0243240966796875, 0.024170495986938476, 0.024201215744018553, 0.024387392044067382, 0.024236223220825196, 0.02449612808227539, 0.024887296676635744, 0.024442880630493165, 0.024375295639038085, 0.02430156707763672, 0.024215551376342775, 0.024225791931152343, 0.024242431640625, 0.024082176208496092, 0.024518655776977538, 0.024147968292236328, 0.024311071395874025, 0.024314592361450196, 0.024219648361206055, 0.024176639556884767, 0.02419286346435547, 0.024172704696655275, 0.024143871307373048, 0.024135679244995118, 0.024161632537841798]",tokens/s,40.89770168330234,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3722.653696,4371.447808,0.0,3992.977408,3875.045888,s,1,9.69694921875,9.69694921875,0.0,9.69694921875,9.69694921875,9.69694921875,9.69694921875,[9.69694921875],,kWh,7.936803180833749e-05,8.743291083237318e-06,2.7014743834000154e-05,0.00011512606672557496,,MB,1954.742272,4595.843072,0.0,4188.012544,4099.587072,s,10,3.046258911132813,0.30462589111328126,0.0025102218270380726,0.30477621459960935,0.3076242462158203,0.3078436019897461,0.3080190866088867,"[0.2989281311035156, 0.3049571838378906, 0.30459524536132815, 0.3032857666015625, 0.30345806884765625, 0.30550613403320315, 0.30324420166015625, 0.3080629577636719, 0.30757550048828125, 0.30664572143554686]",tokens/s,840.3750550041108,kWh,8.844502419949459e-06,9.753858381427302e-07,5.889070367818137e-06,1.5708958625910327e-05,tokens/kWh,16296433.525374118,MB,1963.761664,4700.700672,0.0,4292.870144,4197.763584,s,10,22.24019262695312,2.224019262695312,0.013936060472504645,2.222382568359375,2.2395323730468752,2.245690869140625,2.250617666015625,"[2.212475830078125, 2.20854931640625, 2.238163818359375, 2.220225341796875, 2.21496533203125, 2.205560302734375, 2.23595654296875, 2.227906982421875, 2.224539794921875, 2.251849365234375]",tokens/s,28.32709278050481,kWh,6.514719932088363e-05,7.185694272140852e-06,4.2981329839581884e-05,0.00011531422343260636,tokens/kWh,546333.2980499096,,s,630,22.23730520248411,0.035297309845212906,0.0004976233514561854,0.03523739242553711,0.035796012878417965,0.03608641014099121,0.03681490692138672,"[0.03545292663574219, 0.03514486312866211, 0.034976608276367185, 0.034837696075439455, 0.035070304870605466, 0.03481238555908203, 0.034871295928955076, 0.03514531326293945, 0.034859424591064454, 0.03491193771362305, 0.03500678253173828, 0.03504489517211914, 0.03498566436767578, 0.03528169631958008, 0.03475167846679687, 0.034859039306640624, 0.03480656051635742, 0.035128833770751954, 0.03493529510498047, 0.035416065216064455, 0.03500543975830078, 0.03509145736694336, 0.034936832427978515, 0.035309089660644534, 0.03483081436157227, 0.034912254333496096, 0.03482352066040039, 0.03494355010986328, 0.034760032653808594, 0.03491507339477539, 0.035334144592285156, 0.035530750274658206, 0.03545209503173828, 0.03553363037109375, 0.03545702362060547, 0.03541196823120117, 0.035522079467773436, 0.03609852981567383, 0.035501953125, 0.03554828643798828, 0.035371391296386716, 0.03516476821899414, 0.03552259063720703, 0.03531980895996094, 0.035110912322998046, 0.03514291381835938, 0.034960128784179687, 0.03520716857910156, 0.034904064178466795, 0.03493622589111328, 0.034981857299804686, 0.03550886535644531, 0.03535871887207031, 0.03513753509521484, 0.035077407836914064, 0.03481631851196289, 0.034984096527099606, 0.03472000122070312, 0.03484652709960938, 0.03513158416748047, 0.03496345520019531, 0.03492454528808594, 0.03490374374389649, 0.03561062240600586, 0.03522864151000977, 0.03503513717651367, 0.034920448303222655, 0.03504294586181641, 0.03482457733154297, 0.03504742431640625, 0.034920318603515625, 0.03493219375610351, 0.034915199279785154, 0.03477187347412109, 0.03473702239990235, 0.03483225631713867, 0.034847904205322265, 0.03475500869750976, 0.034867710113525394, 0.034582561492919925, 0.03467059326171875, 0.034719585418701175, 0.034871295928955076, 0.034683361053466796, 0.03461907196044922, 0.0348812141418457, 0.03545119857788086, 0.03478489685058594, 0.03511251068115234, 0.034673118591308595, 0.03482790374755859, 0.036085952758789064, 0.03491279983520508, 0.03479062271118164, 0.03503376007080078, 0.03535788726806641, 0.03496646499633789, 0.03487238311767578, 0.03511123275756836, 0.03488832092285156, 0.03499008178710938, 0.03488915252685547, 0.03497824096679687, 0.03533631896972656, 0.03506175994873047, 0.0349793930053711, 0.036022689819335936, 0.03498713684082031, 0.03490057754516602, 0.03483679962158203, 0.03477840042114258, 0.0349703369140625, 0.035373054504394534, 0.03530342483520508, 0.035381248474121094, 0.035250175476074216, 0.035520511627197264, 0.03567001724243164, 0.0353259506225586, 0.035184513092041014, 0.035152000427246095, 0.035127296447753906, 0.03529449462890625, 0.0352918701171875, 0.035206401824951175, 0.03528985595703125, 0.0358570556640625, 0.035509727478027345, 0.03552719879150391, 0.035620864868164064, 0.03548364639282227, 0.0354317741394043, 0.036676254272460934, 0.03561062240600586, 0.035846145629882815, 0.035281982421875, 0.03558291244506836, 0.03546102523803711, 0.03530556869506836, 0.03501670455932617, 0.03539475250244141, 0.0350052490234375, 0.03500435256958008, 0.03495328140258789, 0.035104351043701174, 0.035682720184326173, 0.03566339111328125, 0.0353546257019043, 0.035232223510742185, 0.03508224105834961, 0.03531545639038086, 0.03539174270629883, 0.03533004760742187, 0.03554079818725586, 0.03543008041381836, 0.035526336669921874, 0.035563743591308594, 0.035570270538330076, 0.03548915100097656, 0.03533824157714844, 0.03517504119873047, 0.03540991973876953, 0.03550207901000976, 0.03539763259887695, 0.035299327850341795, 0.03522710418701172, 0.035138080596923825, 0.035149120330810545, 0.03523244857788086, 0.03494022369384766, 0.03508070373535156, 0.03590163040161133, 0.03742051315307617, 0.03576886367797852, 0.035676158905029294, 0.03567001724243164, 0.03574784088134766, 0.035454975128173825, 0.035672065734863284, 0.03543040084838867, 0.03564287948608398, 0.03529369735717774, 0.0352542724609375, 0.03501875305175781, 0.039215103149414066, 0.03560195159912109, 0.0357743034362793, 0.035394336700439455, 0.035202560424804685, 0.03569907379150391, 0.03517747116088867, 0.03496857452392578, 0.034944286346435545, 0.03503792190551758, 0.03485203170776367, 0.03534112167358398, 0.03491635131835937, 0.034869247436523435, 0.034884830474853516, 0.03479632186889649, 0.034844001770019534, 0.034835105895996095, 0.03467059326171875, 0.03468492889404297, 0.035672065734863284, 0.03465830230712891, 0.03475225448608398, 0.03486745452880859, 0.034764801025390625, 0.03499187088012695, 0.03499647903442383, 0.03486684799194336, 0.03486547088623047, 0.03494297790527344, 0.03489590454101563, 0.03508019256591797, 0.0352542724609375, 0.03544374465942383, 0.03936355209350586, 0.035767425537109376, 0.035494785308837894, 0.035438591003417966, 0.035659584045410156, 0.03550636672973633, 0.035700382232666014, 0.03523004913330078, 0.03546112060546875, 0.03521478271484375, 0.03535520172119141, 0.03515760040283203, 0.034982078552246096, 0.03499030303955078, 0.03492428970336914, 0.035643646240234375, 0.034973056793212894, 0.035289119720458985, 0.03535436630249023, 0.03515068817138672, 0.036234848022460936, 0.03504579162597656, 0.034885887145996095, 0.034933502197265626, 0.034854080200195314, 0.03534652709960937, 0.035122913360595705, 0.03525174331665039, 0.03535488128662109, 0.03533846282958984, 0.03539734268188476, 0.03541376113891601, 0.03607196807861328, 0.03545702362060547, 0.03643801498413086, 0.03532185745239258, 0.035186302185058596, 0.03571340942382813, 0.035389438629150394, 0.03516416168212891, 0.03514883041381836, 0.03497264099121094, 0.03492659378051758, 0.035073856353759765, 0.03471379089355469, 0.03480335998535156, 0.034574687957763674, 0.034968929290771486, 0.03492665481567383, 0.034951774597167966, 0.03494063949584961, 0.03482038497924805, 0.03483443069458008, 0.03500851058959961, 0.034887680053710936, 0.03470064163208008, 0.03468764877319336, 0.03466649627685547, 0.03485625457763672, 0.034982593536376956, 0.03466854476928711, 0.034797569274902344, 0.034778656005859374, 0.03574012756347656, 0.035448833465576174, 0.03494297790527344, 0.03493030548095703, 0.03471807861328125, 0.03506332778930664, 0.035402496337890624, 0.03485795211791992, 0.0347770881652832, 0.03516460800170899, 0.03495532989501953, 0.03507020950317383, 0.0353704948425293, 0.035460609436035156, 0.035500961303710936, 0.03539344024658203, 0.03555715179443359, 0.03549020767211914, 0.035573760986328126, 0.035646751403808595, 0.035214046478271484, 0.035092479705810545, 0.03542425537109375, 0.03506995010375977, 0.035178497314453126, 0.03507382583618164, 0.03500566482543945, 0.036086784362792966, 0.035844097137451174, 0.034950496673583985, 0.03520783996582031, 0.03550003051757813, 0.035514366149902346, 0.03550185775756836, 0.03628998565673828, 0.035383872985839844, 0.03612876892089844, 0.035788734436035155, 0.035295295715332034, 0.035135486602783206, 0.03503827285766602, 0.03497465515136719, 0.03482624053955078, 0.03502684783935547, 0.03521750259399414, 0.03516211318969727, 0.03492659378051758, 0.03482828903198242, 0.035775646209716794, 0.03483513641357422, 0.034875232696533205, 0.034533695220947264, 0.03461529541015625, 0.03471769714355469, 0.0357130241394043, 0.034721569061279295, 0.034945247650146484, 0.03507199859619141, 0.03469311904907227, 0.03591167831420899, 0.0345945930480957, 0.034652385711669925, 0.03483356857299805, 0.03470761489868164, 0.0348842887878418, 0.0347380142211914, 0.034842784881591794, 0.035094528198242186, 0.03489756774902344, 0.034795295715332034, 0.034896446228027345, 0.03473721694946289, 0.0347309455871582, 0.03478937530517578, 0.03666124725341797, 0.03503308868408203, 0.035106143951416015, 0.03573417663574219, 0.03476025772094726, 0.03482022476196289, 0.034793792724609376, 0.03476643371582031, 0.034789505004882815, 0.03473027038574219, 0.034701057434082035, 0.034763008117675784, 0.0348221435546875, 0.03477088165283203, 0.03476486587524414, 0.03475235366821289, 0.03479334259033203, 0.0347977294921875, 0.034707584381103516, 0.034955265045166016, 0.03473583984375, 0.03490639877319336, 0.035004417419433595, 0.03623481750488281, 0.03502739334106445, 0.03476224136352539, 0.03468566513061523, 0.03477481460571289, 0.03539286422729492, 0.03538995361328125, 0.03538095855712891, 0.03508268737792969, 0.035148799896240236, 0.03522457504272461, 0.03536185455322265, 0.03539616012573242, 0.0354758415222168, 0.03549321746826172, 0.03536553573608398, 0.035356670379638674, 0.035342334747314456, 0.035563518524169925, 0.035804767608642575, 0.035387809753417966, 0.035261566162109376, 0.035494785308837894, 0.03555433654785156, 0.0349988784790039, 0.03512153625488281, 0.03498758316040039, 0.03526291275024414, 0.03605871963500976, 0.035510688781738284, 0.035547134399414065, 0.03566796875, 0.035755168914794924, 0.03567292785644531, 0.03559219360351563, 0.03565465545654297, 0.03573030471801758, 0.03590351867675781, 0.03568239974975586, 0.03589897537231445, 0.035516830444335935, 0.035573760986328126, 0.03569427108764649, 0.03682892990112305, 0.03611091232299805, 0.03592739105224609, 0.035334815979003904, 0.03569868850708008, 0.03515142440795899, 0.03517715072631836, 0.03544409561157227, 0.0353012809753418, 0.03521379089355469, 0.03549900817871094, 0.035531230926513675, 0.03678057479858399, 0.036896446228027346, 0.03538358306884765, 0.035084320068359376, 0.035272705078125, 0.03503308868408203, 0.035225601196289064, 0.03500233459472656, 0.035856449127197265, 0.035598175048828125, 0.03556556701660156, 0.03605100631713867, 0.035741470336914063, 0.03559088134765625, 0.03526006317138672, 0.03515631866455078, 0.035447967529296874, 0.03536124801635742, 0.03524236679077149, 0.035185791015625, 0.035016830444335935, 0.03594931030273438, 0.035082527160644535, 0.03524323272705078, 0.03525795364379883, 0.03524291229248047, 0.03540377426147461, 0.03542611312866211, 0.03494924926757813, 0.03505942535400391, 0.03496502304077148, 0.03477791976928711, 0.034742111206054686, 0.03490409469604492, 0.03472198486328125, 0.037865089416503905, 0.03513286590576172, 0.035187583923339844, 0.03531161499023437, 0.03535257720947266, 0.03535436630249023, 0.03520127868652344, 0.03504272079467773, 0.035207775115966795, 0.03549385452270508, 0.03549520111083984, 0.035242336273193356, 0.03513753509521484, 0.03499836730957031, 0.03482246398925781, 0.03577043151855469, 0.035076000213623046, 0.03484374237060547, 0.03529619216918945, 0.03574784088134766, 0.035899391174316404, 0.03495529556274414, 0.03511497497558594, 0.03507523345947266, 0.03528508758544922, 0.03541648101806641, 0.035409793853759766, 0.03540425491333008, 0.035440254211425784, 0.03558233642578125, 0.03550207901000976, 0.035467456817626954, 0.035571521759033206, 0.03559158325195313, 0.036123233795166014, 0.035388607025146485, 0.03594044876098633, 0.03558860778808594, 0.035274879455566406, 0.03612044906616211, 0.035102718353271486, 0.03538534545898438, 0.035200542449951175, 0.03529366302490235, 0.035778560638427735, 0.03569631958007812, 0.03572137451171875, 0.03577052688598633, 0.035506175994873046, 0.03578243255615234, 0.03556169509887695, 0.03570687866210937, 0.035606529235839846, 0.03558399963378906, 0.03570278549194336, 0.03555942535400391, 0.03537715148925781, 0.03547500610351562, 0.03531411361694336, 0.035332096099853515, 0.0353259506225586, 0.03544268798828125, 0.03610403060913086, 0.03527423858642578, 0.03518940734863281, 0.03542425537109375, 0.03568377685546875, 0.036299327850341796, 0.03547955322265625, 0.03545087814331055, 0.03527475357055664, 0.03525836944580078, 0.03505881500244141, 0.035089279174804686, 0.0348671989440918, 0.034961185455322265, 0.03492047882080078, 0.034876991271972656, 0.03492108917236328, 0.03491430282592774, 0.03487744140625, 0.0351539192199707, 0.03501567840576172, 0.034923519134521484, 0.0349917106628418, 0.034892192840576174, 0.03485696029663086, 0.0349224967956543, 0.03478732681274414, 0.034994174957275394, 0.034834144592285156, 0.03588534545898438, 0.0348590087890625, 0.03521033477783203, 0.03499827194213867, 0.035172382354736326, 0.034827136993408205, 0.034928638458251955, 0.0349306869506836, 0.036214656829833984, 0.035624897003173825, 0.03568582534790039, 0.035453697204589844, 0.03509436798095703, 0.03489807891845703, 0.03489785766601562, 0.03476076889038086, 0.03500851058959961, 0.03525836944580078, 0.03530547332763672, 0.03535171127319336, 0.036683998107910155, 0.03518323135375977, 0.03533004760742187, 0.036251903533935544, 0.035923168182373046, 0.035674655914306644, 0.03565785598754883, 0.03567174530029297, 0.03627027130126953, 0.036773792266845705, 0.035690593719482425, 0.0357130241394043, 0.03582156753540039, 0.03590956878662109, 0.03588227081298828, 0.03616438293457031, 0.036003551483154296, 0.03576860809326172, 0.03604889678955078, 0.035813377380371096, 0.03566748809814453, 0.03534236907958985, 0.03526496124267578, 0.03547676849365235, 0.036526817321777344, 0.035624576568603517, 0.03564287948608398, 0.03608870315551758, 0.03662416076660156, 0.03751715087890625, 0.03652851104736328, 0.03592396926879883, 0.03569036865234375, 0.035795040130615234, 0.03571036911010742, 0.03565145492553711, 0.03589321517944336, 0.035844894409179685, 0.03555670547485352, 0.03538191986083984, 0.035332000732421875, 0.03602851104736328, 0.03552025604248047, 0.03565798568725586, 0.035850238800048825, 0.03546931076049804, 0.03572326278686523, 0.035659774780273434, 0.03567776107788086, 0.03549638366699219, 0.035542015075683595]",tokens/s,28.33077093935027,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3798, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ self.model = DeciCoderModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ self.self_attn = DeciCoderAttention(config=config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ self._init_rope() File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1709, in __getattr__ raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2169.253888,2486.10816,0.0,2107.63776,1984.899072,s,1,8.8144833984375,8.8144833984375,0.0,8.8144833984375,8.8144833984375,8.8144833984375,8.8144833984375,[8.8144833984375],,kWh,5.424428525838417e-05,5.9763018791294576e-06,1.7273902708003774e-05,7.749448984551741e-05,,MB,2199.273472,2823.749632,0.0,2415.919104,2256.887808,s,10,1.735695556640625,0.17356955566406249,0.0004596212890163055,0.17345442962646485,0.17400710449218748,0.17439400177001954,0.17470351959228517,"[0.17478089904785157, 0.17301289367675782, 0.17341500854492187, 0.17358131408691407, 0.17349385070800782, 0.17329776000976563, 0.17332313537597657, 0.17392112731933593, 0.1735089874267578, 0.17336058044433594]",tokens/s,1474.9130342620604,kWh,5.1528663482451505e-06,5.678961177005749e-07,3.42743938912333e-06,9.148201855069055e-06,tokens/kWh,27983641.3817377,MB,2214.412288,2865.692672,0.0,2457.862144,2339.93984,s,10,19.12052990722656,1.9120529907226562,0.012359207432674373,1.9121944580078125,1.9239779296875001,1.9306148803710939,1.9359244409179688,"[1.908995361328125, 1.9225030517578126, 1.9153935546875, 1.8950416259765626, 1.9156470947265625, 1.92163134765625, 1.905275146484375, 1.9004888916015625, 1.898302001953125, 1.9372518310546876]",tokens/s,32.94887762299375,kWh,5.549004872341853e-05,6.120737917687838e-06,3.159161640387469e-05,9.320240304498108e-05,tokens/kWh,675948.2367594655,,s,630,19.117521875381463,0.030345272818065826,0.0005584700915796066,0.030308752059936522,0.03076296901702881,0.031010713958740232,0.0323111268234253,"[0.03106768035888672, 0.03059779167175293, 0.030469600677490234, 0.030671743392944335, 0.030220287322998047, 0.03019699287414551, 0.0302476806640625, 0.030318143844604493, 0.03017568016052246, 0.030491968154907227, 0.030253471374511717, 0.030986528396606445, 0.031847776412963866, 0.030677663803100587, 0.030050304412841795, 0.030253055572509766, 0.03034316825866699, 0.030317983627319335, 0.030644800186157225, 0.030861343383789062, 0.029888063430786132, 0.02978019142150879, 0.029738367080688475, 0.029812639236450195, 0.029677824020385744, 0.029803007125854493, 0.029860031127929686, 0.029831167221069335, 0.029740032196044923, 0.030026752471923827, 0.029859935760498047, 0.029895967483520507, 0.029973119735717774, 0.030443519592285157, 0.030228479385375977, 0.030320640563964843, 0.030607295989990235, 0.030521408081054687, 0.030511104583740234, 0.030736383438110353, 0.03043961524963379, 0.030275392532348632, 0.029995008468627928, 0.030060543060302734, 0.030054399490356445, 0.030417951583862304, 0.030307104110717773, 0.030138080596923827, 0.030042591094970702, 0.02997248077392578, 0.029957279205322266, 0.029903711318969725, 0.02999465560913086, 0.030027519226074217, 0.030222944259643555, 0.03068832015991211, 0.03089299201965332, 0.030279680252075194, 0.030504127502441407, 0.030594976425170898, 0.0308822078704834, 0.030503423690795898, 0.03059872055053711, 0.03075071907043457, 0.030143840789794922, 0.030057119369506835, 0.030042112350463866, 0.030015487670898438, 0.030063871383666993, 0.030118528366088866, 0.03034124755859375, 0.030256160736083983, 0.030419008255004883, 0.03050092887878418, 0.030472671508789063, 0.030411136627197265, 0.03057459259033203, 0.03042505645751953, 0.030414880752563475, 0.03085945510864258, 0.030623327255249022, 0.030361888885498046, 0.030528831481933593, 0.030496608734130858, 0.030992864608764648, 0.03054240036010742, 0.03065011215209961, 0.030727167129516602, 0.030485504150390624, 0.030267391204833984, 0.030320640563964843, 0.030344383239746094, 0.03066963195800781, 0.031498239517211916, 0.030906368255615234, 0.03066044807434082, 0.03067510414123535, 0.03058633613586426, 0.030691392898559572, 0.03059350395202637, 0.030734336853027344, 0.03069523239135742, 0.030820032119750977, 0.03051366424560547, 0.030332704544067383, 0.030240991592407226, 0.03039023971557617, 0.030498815536499024, 0.030380064010620118, 0.03029529571533203, 0.030381919860839844, 0.030384767532348634, 0.030316831588745118, 0.030027776718139648, 0.029892416000366212, 0.03024505615234375, 0.03014409637451172, 0.030536096572875978, 0.03146249580383301, 0.030487455368041993, 0.0309403190612793, 0.03069219207763672, 0.030762943267822265, 0.03068934440612793, 0.030760480880737306, 0.031062623977661134, 0.030873760223388672, 0.03057868766784668, 0.030142112731933592, 0.030048351287841796, 0.02993984031677246, 0.030175008773803712, 0.029862495422363283, 0.032608192443847654, 0.03075926399230957, 0.030394367218017578, 0.03040185546875, 0.03061199951171875, 0.030374048233032226, 0.03123200035095215, 0.030545120239257813, 0.030532384872436525, 0.030379295349121094, 0.030554847717285158, 0.03032268714904785, 0.03119033622741699, 0.030450368881225588, 0.030631231307983398, 0.030009056091308595, 0.030264287948608398, 0.030267711639404296, 0.03027452850341797, 0.02997324752807617, 0.030122976303100586, 0.029897760391235352, 0.030027423858642578, 0.029873855590820314, 0.030009727478027343, 0.029802240371704102, 0.030245376586914063, 0.03029372787475586, 0.030582592010498046, 0.03034886360168457, 0.030623935699462892, 0.030394432067871093, 0.030662912368774414, 0.030484735488891603, 0.03056844711303711, 0.0305328311920166, 0.030532543182373046, 0.03037593650817871, 0.030270719528198244, 0.03019388771057129, 0.030943775177001955, 0.03278643035888672, 0.03042918395996094, 0.030047712326049806, 0.030093856811523437, 0.029995008468627928, 0.02996019172668457, 0.03012601661682129, 0.030238784790039063, 0.03051907157897949, 0.02989619255065918, 0.030005983352661133, 0.03040460777282715, 0.030320032119750977, 0.03021676826477051, 0.029904224395751952, 0.03020364761352539, 0.030183679580688478, 0.03019980812072754, 0.030353408813476562, 0.030274879455566405, 0.030382783889770507, 0.03029167938232422, 0.0304169921875, 0.030231935501098633, 0.030270015716552735, 0.030513408660888672, 0.030234624862670898, 0.029981727600097655, 0.03005539131164551, 0.02999295997619629, 0.029908992767333983, 0.029727968215942382, 0.029864736557006836, 0.02970419120788574, 0.029739007949829102, 0.02968780708312988, 0.029800447463989257, 0.029648895263671874, 0.029968128204345704, 0.030216575622558594, 0.03022220802307129, 0.029833215713500977, 0.029995008468627928, 0.029784416198730467, 0.02976028823852539, 0.029627264022827147, 0.03001753616333008, 0.02980803108215332, 0.029860448837280274, 0.032159744262695314, 0.03015235137939453, 0.029860191345214844, 0.02991641616821289, 0.02970086479187012, 0.029814783096313476, 0.029673471450805664, 0.029899904251098633, 0.029672128677368164, 0.02980473518371582, 0.02976153564453125, 0.029847551345825195, 0.02972572708129883, 0.029754335403442383, 0.02975334358215332, 0.029701696395874024, 0.029845951080322265, 0.029822240829467772, 0.030121824264526368, 0.029835487365722658, 0.031031967163085938, 0.030066688537597655, 0.030265344619750976, 0.03012403106689453, 0.030043487548828126, 0.03162297630310058, 0.030636896133422853, 0.030930944442749023, 0.030428384780883787, 0.030752832412719727, 0.030445247650146483, 0.030435359954833985, 0.03039983940124512, 0.030259647369384766, 0.029974720001220704, 0.02989641571044922, 0.029966848373413086, 0.03059052848815918, 0.03016534423828125, 0.03004662322998047, 0.029908767700195314, 0.030182783126831054, 0.030543487548828126, 0.03022105598449707, 0.03012019157409668, 0.030250240325927734, 0.029940479278564452, 0.029786336898803712, 0.029853471755981444, 0.029834272384643555, 0.029946847915649413, 0.02995199966430664, 0.030251007080078125, 0.030705663681030275, 0.031197183609008788, 0.03059507179260254, 0.0306212158203125, 0.030625919342041015, 0.030832799911499023, 0.03059712028503418, 0.030763200759887695, 0.030435327529907227, 0.030352575302124023, 0.03010767936706543, 0.03019990348815918, 0.03033888053894043, 0.030558719635009765, 0.030306175231933595, 0.030213888168334962, 0.02987718391418457, 0.03011974334716797, 0.029843456268310548, 0.02979635238647461, 0.02972604751586914, 0.029976800918579103, 0.029905344009399416, 0.030466047286987305, 0.03055411148071289, 0.03071107292175293, 0.030456159591674806, 0.0305579833984375, 0.03444591903686523, 0.032032703399658205, 0.0308155517578125, 0.030655231475830078, 0.03044112014770508, 0.030906719207763673, 0.031497440338134765, 0.030011999130249024, 0.0300382080078125, 0.030248224258422853, 0.03010633659362793, 0.030722015380859374, 0.030260063171386718, 0.030464000701904297, 0.03061555290222168, 0.03055820846557617, 0.03196457672119141, 0.0316135368347168, 0.03049616050720215, 0.030408992767333984, 0.0306998405456543, 0.030414527893066406, 0.0305996150970459, 0.030336896896362306, 0.030552064895629883, 0.030111743927001954, 0.030489919662475586, 0.03070636749267578, 0.030676992416381835, 0.03049283218383789, 0.030613344192504884, 0.030533632278442382, 0.03100057601928711, 0.03186796760559082, 0.03174905586242676, 0.030958911895751954, 0.030806432723999022, 0.030691616058349608, 0.030838783264160157, 0.030752767562866212, 0.030732160568237306, 0.030756992340087892, 0.030894079208374024, 0.030570240020751954, 0.030877952575683595, 0.030617536544799803, 0.030761024475097657, 0.030496320724487304, 0.030726591110229493, 0.030439424514770507, 0.030543071746826172, 0.030262048721313477, 0.030295839309692384, 0.030247135162353514, 0.03019366455078125, 0.029879840850830078, 0.02992323112487793, 0.029930047988891602, 0.029910943984985353, 0.030000288009643553, 0.02992633628845215, 0.029890560150146486, 0.03010313606262207, 0.03064054489135742, 0.03077120018005371, 0.030693056106567383, 0.030335296630859376, 0.029879776000976563, 0.029950496673583984, 0.029718175888061523, 0.029749824523925782, 0.02995792007446289, 0.029848960876464842, 0.029847232818603516, 0.030042816162109375, 0.030727968215942383, 0.030746912002563475, 0.029821887969970703, 0.029843807220458984, 0.029786239624023436, 0.02982147216796875, 0.030070783615112305, 0.03003392028808594, 0.030208000183105467, 0.030346912384033205, 0.030402912139892577, 0.030183744430541993, 0.030352319717407227, 0.030276351928710938, 0.030474239349365235, 0.030218143463134766, 0.030054496765136718, 0.030040063858032227, 0.03009721565246582, 0.0299520320892334, 0.03015670394897461, 0.029883808135986328, 0.02981340789794922, 0.02978220748901367, 0.030006624221801757, 0.029798208236694337, 0.029837919235229493, 0.029843711853027345, 0.029802303314208984, 0.0301296329498291, 0.030621919631958008, 0.030607872009277344, 0.03064784049987793, 0.03028630447387695, 0.03064793586730957, 0.030277599334716798, 0.030910720825195314, 0.03055836868286133, 0.03069715118408203, 0.03041926383972168, 0.030670848846435547, 0.030111743927001954, 0.03031590461730957, 0.030073471069335937, 0.03049180793762207, 0.03035251235961914, 0.030412511825561525, 0.030164512634277343, 0.029987295150756835, 0.029863935470581054, 0.030131744384765624, 0.030226943969726562, 0.03019977569580078, 0.030029823303222656, 0.03063590431213379, 0.03032048034667969, 0.03049497604370117, 0.030435039520263673, 0.030362239837646486, 0.03037343978881836, 0.03053171157836914, 0.030480384826660156, 0.030975648880004883, 0.03029007911682129, 0.03014179229736328, 0.030372447967529297, 0.030071008682250978, 0.029945695877075195, 0.029693952560424806, 0.02978540802001953, 0.02996281623840332, 0.02982310485839844, 0.02998886489868164, 0.030762624740600587, 0.03055859184265137, 0.030826496124267577, 0.030538911819458007, 0.030468896865844728, 0.030362848281860352, 0.030353279113769532, 0.03010972785949707, 0.0301331844329834, 0.030021247863769532, 0.030275039672851563, 0.030425504684448244, 0.030414655685424806, 0.030228992462158204, 0.030093503952026368, 0.029997055053710937, 0.030068031311035158, 0.029885120391845703, 0.03023052787780762, 0.03013612747192383, 0.030378175735473634, 0.030212448120117186, 0.029867679595947265, 0.029724159240722657, 0.02988697624206543, 0.02978825569152832, 0.029822879791259766, 0.029642847061157225, 0.029894432067871093, 0.029740352630615235, 0.030024192810058595, 0.02983897590637207, 0.030065343856811522, 0.030312448501586913, 0.03058073616027832, 0.030453760147094725, 0.03051091194152832, 0.030334911346435546, 0.030556415557861327, 0.030555871963500975, 0.0302893123626709, 0.030086015701293944, 0.030310400009155275, 0.029911039352416992, 0.030156736373901368, 0.030316608428955078, 0.030162784576416017, 0.02997817611694336, 0.029960735321044922, 0.029847232818603516, 0.03036115264892578, 0.02970707130432129, 0.030621728897094726, 0.030691295623779296, 0.030470176696777342, 0.030401920318603514, 0.030310848236083984, 0.03041708755493164, 0.03019878387451172, 0.030218591690063478, 0.029910783767700195, 0.030446495056152344, 0.030090944290161133, 0.030380319595336915, 0.0300152645111084, 0.030101760864257813, 0.029839359283447265, 0.02990825653076172, 0.029747840881347656, 0.02981692886352539, 0.029796480178833008, 0.029849472045898436, 0.03101900863647461, 0.03278438568115234, 0.030514495849609375, 0.030436031341552733, 0.030320287704467774, 0.030415199279785157, 0.03026915168762207, 0.030365983963012697, 0.030469888687133788, 0.030376192092895507, 0.030240608215332032, 0.030111072540283203, 0.029743936538696288, 0.029726720809936522, 0.02976742362976074, 0.030028032302856444, 0.02996767997741699, 0.02991379165649414, 0.02979216003417969, 0.029716575622558594, 0.029483007431030273, 0.029832576751708983, 0.029604576110839845, 0.030408607482910157, 0.030457183837890624, 0.031844383239746095, 0.030589599609375, 0.03022972869873047, 0.029993728637695314, 0.02975334358215332, 0.029645952224731445, 0.02979724884033203, 0.02969772720336914, 0.029808351516723633, 0.029803104400634765, 0.02959564781188965, 0.029681312561035157, 0.029628768920898437, 0.02958336067199707, 0.029883615493774413, 0.02978416061401367, 0.02990764808654785, 0.029797439575195313, 0.030674943923950194, 0.030463199615478515, 0.030424863815307616, 0.030471168518066406, 0.030374080657958984, 0.03057833671569824, 0.030378143310546876, 0.030822399139404297, 0.030971904754638672, 0.030719999313354493, 0.030455104827880858, 0.030644927978515625, 0.03066009521484375, 0.030629568099975586, 0.030493024826049805, 0.030591455459594727, 0.030482431411743165, 0.030619647979736327, 0.030518976211547852, 0.030867776870727538, 0.03054182434082031, 0.03068441581726074, 0.03053366470336914, 0.03064291191101074, 0.03142860794067383, 0.030697471618652345, 0.030176576614379884, 0.030406879425048827, 0.030344863891601563, 0.030558271408081053, 0.030513919830322266, 0.03047603225708008, 0.03054172706604004, 0.030609760284423828, 0.030547967910766603, 0.030687231063842774, 0.03097804832458496, 0.030519296646118164, 0.035633056640625, 0.03513935852050781, 0.031072576522827147, 0.03138355255126953, 0.03037151908874512, 0.030582719802856446, 0.030427520751953124, 0.030449663162231445, 0.030184736251831056, 0.030114143371582032, 0.029906784057617188, 0.029671968460083006, 0.02982707214355469, 0.029750368118286134, 0.0312957763671875, 0.031915807723999025, 0.030251871109008788, 0.03013983917236328, 0.03040937614440918, 0.030409664154052735, 0.030264223098754883, 0.03048886489868164, 0.030806848526000977, 0.03237295913696289, 0.031296255111694336]",tokens/s,32.9540619389211,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,7393.939456,7808.679936,0.0,7430.209536,7414.23104,s,1,11.698818359375,11.698818359375,0.0,11.698818359375,11.698818359375,11.698818359375,11.698818359375,[11.698818359375],,kWh,0.00013635895708749785,1.503068300448296e-05,4.451781339201433e-05,0.00019590745348399514,,MB,1706.82368,8584.626176,0.0,8176.795648,8052.041728,s,10,7.289791870117187,0.7289791870117187,0.0011240875603674401,0.7291243591308594,0.7299213500976562,0.7300348937988281,0.7301257287597657,"[0.7261265258789062, 0.7281160888671875, 0.7288939819335938, 0.7288358154296875, 0.7288203125, 0.729354736328125, 0.729755859375, 0.729843994140625, 0.7301484375, 0.7298961181640625]",tokens/s,351.1760068890481,kWh,2.1263119047915778e-05,2.3449535359086695e-06,1.421509073714296e-05,3.782316332096741e-05,tokens/kWh,6768339.227144586,MB,1711.788032,8731.426816,0.0,8323.596288,8263.496192,s,10,33.483260986328126,3.3483260986328127,0.006480873101779639,3.347747680664063,3.3554736572265624,3.356179089355469,3.356743435058594,"[3.334534912109375, 3.34156396484375, 3.347009765625, 3.3464755859375, 3.348485595703125, 3.346180419921875, 3.353955078125, 3.352854248046875, 3.35531689453125, 3.356884521484375]",tokens/s,18.81537166458313,kWh,9.774259868624956e-05,1.0781551038561054e-05,6.486063919005557e-05,0.0001733847889148662,tokens/kWh,363353.6736082061,,s,630,33.47922916412354,0.05314163359384689,0.0006883491317176238,0.05311716842651368,0.053848936462402346,0.05402839469909668,0.05579953453063965,"[0.05493145751953125, 0.052621311187744144, 0.052209663391113284, 0.05240342330932617, 0.05208067321777344, 0.05200563049316406, 0.05291337585449219, 0.052299808502197266, 0.05230259323120117, 0.052393985748291017, 0.05225471878051758, 0.052217281341552735, 0.05221638488769531, 0.05261062240600586, 0.052351425170898434, 0.052332542419433595, 0.05280767822265625, 0.0530882568359375, 0.05320908737182617, 0.05304729461669922, 0.052792415618896485, 0.05274307250976563, 0.05238784027099609, 0.05242630386352539, 0.0529117431640625, 0.05258528137207031, 0.052744190216064454, 0.05286038589477539, 0.05258607864379883, 0.05267516708374023, 0.05284694290161133, 0.05271142578125, 0.05303615951538086, 0.05292428970336914, 0.05300735855102539, 0.0530873908996582, 0.05343318557739258, 0.05355724716186523, 0.05349903869628906, 0.05301129531860352, 0.05314121627807617, 0.052979999542236325, 0.05258966445922852, 0.05317110443115235, 0.052940479278564455, 0.05301484680175781, 0.05323129653930664, 0.052940128326416015, 0.05268278503417969, 0.05318137741088867, 0.05358732986450195, 0.05342643356323242, 0.053371265411376954, 0.05345894241333008, 0.053269985198974606, 0.05354038238525391, 0.05348659133911133, 0.05336265563964844, 0.05361600112915039, 0.05357635116577148, 0.05342771148681641, 0.05302479934692383, 0.052967166900634764, 0.0562625617980957, 0.05240182495117188, 0.05183139038085938, 0.052199104309082034, 0.05209542465209961, 0.05272937774658203, 0.052341217041015624, 0.05176115036010742, 0.052086784362792966, 0.051943103790283204, 0.05230326461791992, 0.055758750915527344, 0.051555328369140625, 0.052628894805908204, 0.05269952011108398, 0.052386016845703126, 0.052703231811523435, 0.05356710433959961, 0.054382080078125, 0.05368921661376953, 0.05243699264526367, 0.052754432678222656, 0.05243494415283203, 0.052531200408935545, 0.05286656188964844, 0.05230339050292969, 0.05253014373779297, 0.05227695846557617, 0.05305168151855469, 0.05287859344482422, 0.05251148986816406, 0.05288755035400391, 0.05275852966308594, 0.05283020782470703, 0.05355708694458008, 0.053591873168945314, 0.05387299346923828, 0.05370169448852539, 0.053125438690185545, 0.053336704254150394, 0.052746238708496096, 0.053067649841308594, 0.053213310241699216, 0.052915679931640626, 0.05280188751220703, 0.05282835388183594, 0.05291417694091797, 0.05292639923095703, 0.05269916915893555, 0.05342006301879883, 0.053284191131591795, 0.053448768615722654, 0.05364582443237305, 0.05393139266967773, 0.05372726440429688, 0.05354905700683594, 0.05347398376464844, 0.05366579055786133, 0.05342950439453125, 0.05333606338500976, 0.05310950469970703, 0.053362144470214846, 0.05407798385620117, 0.05581619262695312, 0.0528238410949707, 0.05232662582397461, 0.05263683319091797, 0.052087646484375, 0.053483455657958985, 0.051933025360107424, 0.051953887939453124, 0.0524758415222168, 0.05196294403076172, 0.05230640029907226, 0.05290835189819336, 0.05311510467529297, 0.052923969268798825, 0.05240467071533203, 0.052836353302001954, 0.052822017669677736, 0.053766143798828124, 0.05369651031494141, 0.053356544494628906, 0.052910079956054686, 0.05284828948974609, 0.05255936050415039, 0.05259964752197266, 0.05247795104980469, 0.05276662445068359, 0.05279862213134766, 0.052813793182373045, 0.05274652862548828, 0.052811870574951174, 0.052934558868408206, 0.05315449523925781, 0.05305702209472656, 0.05329318237304687, 0.053556961059570314, 0.053501758575439456, 0.053547103881835936, 0.05353478240966797, 0.05322412872314453, 0.05323980712890625, 0.05323763275146484, 0.05356294250488281, 0.052996673583984376, 0.05302272033691406, 0.05289574432373047, 0.05362278366088867, 0.05313740921020508, 0.05314559936523437, 0.053207038879394535, 0.05348556900024414, 0.05316182327270508, 0.05363916778564453, 0.0538658561706543, 0.053555999755859375, 0.0539607048034668, 0.05378867340087891, 0.05361395263671875, 0.05360403060913086, 0.05344268798828125, 0.05324473571777344, 0.05320435333251953, 0.05347734451293945, 0.05374224090576172, 0.055185375213623045, 0.05253283309936523, 0.05254019165039062, 0.05204592132568359, 0.052381694793701174, 0.052514816284179686, 0.05263504028320312, 0.05229529571533203, 0.052253662109375, 0.052744190216064454, 0.05252710342407227, 0.05256508636474609, 0.05278543853759766, 0.05269977569580078, 0.05282598495483398, 0.05220512008666992, 0.052716064453125, 0.053847358703613284, 0.05342281723022461, 0.053556705474853514, 0.053015071868896486, 0.05296870422363281, 0.05281049728393555, 0.05262905502319336, 0.05266476821899414, 0.05292998504638672, 0.052528831481933595, 0.052888446807861325, 0.05273948669433594, 0.053224033355712894, 0.05314252853393555, 0.052980735778808595, 0.05342313766479492, 0.05330425643920898, 0.05301046371459961, 0.053233665466308595, 0.05353267288208008, 0.05370675277709961, 0.05351833724975586, 0.05379398345947266, 0.05338390350341797, 0.052787296295166014, 0.05303091049194336, 0.05342172622680664, 0.05324137496948242, 0.053050174713134765, 0.05305753707885742, 0.053579776763916016, 0.053515743255615235, 0.0535700798034668, 0.0530780143737793, 0.05344870376586914, 0.05387468719482422, 0.054368255615234375, 0.05363622283935547, 0.05344659042358398, 0.05354927825927734, 0.05397699356079102, 0.05351507186889649, 0.05328486251831055, 0.05285007858276367, 0.05295513534545898, 0.05321788787841797, 0.055826431274414064, 0.05284454345703125, 0.052217857360839844, 0.052176895141601565, 0.052258815765380856, 0.05227315139770508, 0.05202329635620117, 0.0524901123046875, 0.05209715270996094, 0.052373504638671874, 0.05297478485107422, 0.052902366638183596, 0.05253968048095703, 0.05296511840820312, 0.053399871826171875, 0.05318608093261719, 0.05256035232543945, 0.053464927673339845, 0.053674144744873045, 0.05337427139282227, 0.05307577514648437, 0.05361958312988281, 0.05301763153076172, 0.052783649444580076, 0.05261151885986328, 0.05239603042602539, 0.053348350524902347, 0.05244432067871094, 0.052654945373535156, 0.052950687408447265, 0.05270767974853516, 0.05285859298706055, 0.0534879035949707, 0.053079967498779294, 0.05400092697143555, 0.053256801605224606, 0.053722686767578125, 0.053652065277099606, 0.05321852874755859, 0.05324812698364258, 0.053535457611083984, 0.05325209426879883, 0.05333724975585938, 0.05340041732788086, 0.052888961791992185, 0.0528752326965332, 0.05268239974975586, 0.05306412887573242, 0.053123550415039064, 0.05314704132080078, 0.053451038360595705, 0.05346063995361328, 0.0532732162475586, 0.05395059204101563, 0.05392793655395508, 0.05399715042114258, 0.0539785270690918, 0.05428022384643555, 0.05346607971191406, 0.05337811279296875, 0.05321209716796875, 0.0534950065612793, 0.053153633117675785, 0.05600652694702148, 0.05279414367675781, 0.052055999755859374, 0.05234899139404297, 0.05182793426513672, 0.052577056884765626, 0.052359169006347656, 0.0521965446472168, 0.0523702392578125, 0.05218918228149414, 0.05254492950439453, 0.05206249618530273, 0.05236905670166016, 0.053332382202148435, 0.05291443252563476, 0.05314355087280274, 0.053352481842041014, 0.053528224945068356, 0.053825408935546874, 0.05348191833496094, 0.05333401489257812, 0.053059585571289064, 0.052319263458251955, 0.052759521484375, 0.05257839965820312, 0.05220547103881836, 0.05246297454833984, 0.05235980987548828, 0.05268473434448242, 0.05243283081054687, 0.0524453125, 0.05306777572631836, 0.05288140869140625, 0.053411838531494144, 0.05330124664306641, 0.05346057510375977, 0.053770561218261716, 0.05398537445068359, 0.053524673461914064, 0.053200382232666016, 0.05339123153686524, 0.05335065460205078, 0.05334364700317383, 0.0531033935546875, 0.053364704132080075, 0.05305142211914062, 0.05300656127929688, 0.0526616325378418, 0.05293641662597656, 0.052678913116455076, 0.05339123153686524, 0.053567264556884764, 0.05351299285888672, 0.05355705642700195, 0.05402848052978516, 0.05429043197631836, 0.05393923187255859, 0.05455971145629883, 0.05292851257324219, 0.053553150177001956, 0.0533831672668457, 0.05315379333496094, 0.05447782516479492, 0.055576545715332035, 0.05284864044189453, 0.05203510284423828, 0.05224905776977539, 0.05198361587524414, 0.05374233627319336, 0.052342784881591796, 0.05249407958984375, 0.05284048080444336, 0.052525279998779296, 0.052555553436279295, 0.05289299011230469, 0.052894622802734374, 0.05281689453125, 0.052782081604003904, 0.053579776763916016, 0.053184513092041016, 0.053818527221679686, 0.0537237434387207, 0.053403903961181644, 0.0526798095703125, 0.052806526184082034, 0.052711071014404295, 0.0525142707824707, 0.05263449478149414, 0.05306777572631836, 0.052813278198242185, 0.05298640060424804, 0.052653247833251954, 0.05239072036743164, 0.05313494491577148, 0.053057758331298825, 0.05281811141967773, 0.05361385726928711, 0.0538631362915039, 0.05439827346801758, 0.05378937530517578, 0.05356748962402344, 0.05382291030883789, 0.05374585723876953, 0.05305996704101563, 0.05279286575317383, 0.053105056762695314, 0.05342585754394531, 0.05337945556640625, 0.0531467514038086, 0.0529826545715332, 0.0527215690612793, 0.053195999145507815, 0.05324185562133789, 0.05323798370361328, 0.0540228157043457, 0.05395455932617187, 0.05390713500976563, 0.053948734283447264, 0.054072479248046874, 0.053911582946777344, 0.05361721420288086, 0.05346739196777344, 0.0532209587097168, 0.05400822448730469, 0.05402828979492187, 0.05370598220825195, 0.05594275283813477, 0.05277942276000976, 0.05242675018310547, 0.052160511016845705, 0.05207606506347656, 0.052744670867919924, 0.052621311187744144, 0.052787200927734375, 0.051950817108154294, 0.05259107208251953, 0.052674880981445314, 0.052176895141601565, 0.05203353500366211, 0.05243084716796875, 0.05292572784423828, 0.053243934631347654, 0.052932544708251955, 0.05363759994506836, 0.054151233673095704, 0.05362646484375, 0.05333260726928711, 0.053225345611572265, 0.052922431945800784, 0.05281523132324219, 0.052615135192871094, 0.052982494354248046, 0.05284249496459961, 0.05249001693725586, 0.05311923217773438, 0.05274803161621094, 0.05269712066650391, 0.05246380615234375, 0.05294720077514648, 0.053577022552490236, 0.05373110580444336, 0.0536910400390625, 0.05415935897827148, 0.054075199127197264, 0.053712257385253905, 0.05348406219482422, 0.05341136169433594, 0.05282643127441406, 0.053563232421875, 0.053486175537109375, 0.052833633422851564, 0.052929183959960935, 0.05337001419067383, 0.05312182235717773, 0.0533966064453125, 0.05364012908935547, 0.05350137710571289, 0.05349228668212891, 0.05347865676879883, 0.053615360260009765, 0.05383913421630859, 0.05374003219604492, 0.05371023941040039, 0.05404345703125, 0.05364940643310547, 0.05397401428222656, 0.05393920135498047, 0.05357068634033203, 0.05372198486328125, 0.05515468978881836, 0.05246566390991211, 0.05217279815673828, 0.051975872039794924, 0.051951648712158204, 0.052652320861816405, 0.052530719757080076, 0.052273632049560544, 0.05228300857543945, 0.05311222457885742, 0.052810718536376956, 0.05252703857421875, 0.05240019226074219, 0.0533072624206543, 0.05325222396850586, 0.053300960540771485, 0.05331372833251953, 0.05382092666625977, 0.05386710357666016, 0.053354496002197264, 0.05282592010498047, 0.05412268829345703, 0.053215232849121094, 0.0526250228881836, 0.052578113555908204, 0.05295468902587891, 0.05300089645385742, 0.05315001678466797, 0.053526527404785154, 0.0530513916015625, 0.05293260955810547, 0.053473281860351565, 0.053333248138427734, 0.053252864837646484, 0.05345024108886719, 0.05319631958007812, 0.05397808074951172, 0.053544960021972655, 0.053528575897216796, 0.05331558227539063, 0.05355110549926758, 0.05299814224243164, 0.05299135971069336, 0.05271590423583984, 0.053474559783935546, 0.05366886520385742, 0.05406515121459961, 0.053872638702392575, 0.053815105438232425, 0.05305567932128906, 0.05336064147949219, 0.053520126342773436, 0.05347081756591797, 0.05386921691894531, 0.05418393707275391, 0.054013950347900394, 0.05402624130249024, 0.05353993606567383, 0.05343939208984375, 0.05345183944702148, 0.05323052978515625, 0.05346918487548828, 0.05356748962402344, 0.05700185775756836, 0.058417919158935544, 0.051490943908691404, 0.052127742767333986, 0.052142078399658204, 0.05259190368652344, 0.05243363189697266, 0.05227724838256836, 0.052028446197509765, 0.05258134460449219, 0.05239622497558594, 0.05247078323364258, 0.052564449310302734, 0.05321295928955078, 0.05318531036376953, 0.05319795227050781, 0.052966239929199216, 0.05363481521606445, 0.054828094482421874, 0.05377705764770508, 0.05297187042236328, 0.05234044647216797, 0.052399391174316405, 0.0529332160949707, 0.053010143280029294, 0.05254419326782227, 0.05277692794799805, 0.0529815673828125, 0.052977535247802736, 0.05285209655761719, 0.05329983901977539, 0.053364128112792966, 0.0541357421875, 0.05361372756958008, 0.053412704467773436, 0.05355110549926758, 0.05398070526123047, 0.05361916732788086, 0.053319679260253904, 0.053431903839111325, 0.0534224967956543, 0.05310873413085938, 0.05303209686279297, 0.053220062255859374, 0.05318870544433594, 0.052921409606933596, 0.05282096099853516, 0.05281382369995117, 0.05282137680053711, 0.05352640151977539, 0.05353958511352539, 0.05324556732177734, 0.05349619293212891, 0.053661697387695315, 0.05348873519897461, 0.053690784454345705, 0.0541308479309082, 0.05346953582763672, 0.05354086303710937, 0.053477375030517575, 0.053460990905761716, 0.0535654411315918, 0.0539986572265625]",tokens/s,18.817637554066216,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,8197.681152,11080.237056,0.0,10701.766656,10468.923392,s,1,13.2618525390625,13.2618525390625,0.0,13.2618525390625,13.2618525390625,13.2618525390625,13.2618525390625,[13.2618525390625],,kWh,0.00018762591863748335,2.0689324372002834e-05,6.120893785599435e-05,0.0002695241808654805,,MB,3997.712384,11570.970624,0.0,11163.140096,10923.364352,s,10,7.304953063964844,0.7304953063964844,0.002164870958327621,0.7312315673828125,0.7322960510253906,0.7326491241455078,0.7329315826416015,"[0.7250454711914063, 0.7301515502929687, 0.7313854370117188, 0.72986328125, 0.7315874633789062, 0.7287291870117187, 0.7318931884765625, 0.733002197265625, 0.7310776977539063, 0.7322175903320313]",tokens/s,350.4471524435137,kWh,2.1300824997918728e-05,2.3490922201286624e-06,1.4110725574286878e-05,3.7760642792334264e-05,tokens/kWh,6779545.60805226,MB,4002.03776,11573.067776,0.0,11165.237248,10923.366912,s,10,34.147839843750006,3.414783984375,0.005954342253382606,3.4163756103515626,3.4219365234374997,3.42217822265625,3.42237158203125,"[3.404217041015625, 3.408157958984375, 3.4100400390625, 3.410267333984375, 3.41900244140625, 3.414662109375, 3.418089111328125, 3.41910107421875, 3.422419921875, 3.4218828125]",tokens/s,18.449190428521568,kWh,0.00010015927759166512,1.104791151118098e-05,6.658178342411099e-05,0.00017778897252695712,tokens/kWh,354352.68624688,,s,630,34.1449394950867,0.05419831665886773,0.0008717218262891059,0.054098608016967774,0.05497198028564453,0.05529226551055908,0.0588721688079834,"[0.058912704467773434, 0.05374745559692383, 0.05303049468994141, 0.05383411026000977, 0.05320127868652344, 0.053007904052734374, 0.05378879928588867, 0.05372518539428711, 0.05328736114501953, 0.05348953628540039, 0.052907806396484375, 0.05330697631835937, 0.053582462310791015, 0.05273955154418945, 0.05393052673339844, 0.053833728790283204, 0.053700641632080076, 0.054357662200927734, 0.05469734573364258, 0.05388911819458008, 0.0541124496459961, 0.05401615905761719, 0.05353932952880859, 0.05382064056396484, 0.05440182495117187, 0.053671070098876957, 0.05362511825561524, 0.05384431838989258, 0.0531541748046875, 0.053620574951171875, 0.053876190185546874, 0.053762462615966795, 0.05390348815917969, 0.05429353713989258, 0.053787071228027346, 0.054445919036865235, 0.05604422378540039, 0.055314430236816405, 0.05389046478271484, 0.05391215896606445, 0.05425689697265625, 0.05386111831665039, 0.053884895324707034, 0.054284320831298825, 0.05343641662597656, 0.05402009582519531, 0.05435391998291016, 0.05342351913452149, 0.053999713897705075, 0.05416806411743164, 0.0539504623413086, 0.05399465560913086, 0.05442361450195313, 0.05445302581787109, 0.054099071502685545, 0.054674015045166016, 0.05449552154541016, 0.054300670623779294, 0.05418739318847656, 0.05455116653442383, 0.05406860733032227, 0.054517631530761716, 0.05453286361694336, 0.059359745025634764, 0.053949600219726564, 0.05318127822875977, 0.053311134338378904, 0.05312124633789062, 0.05278323364257813, 0.05350358581542969, 0.05320745468139648, 0.053476577758789064, 0.053381824493408205, 0.05306585693359375, 0.053928001403808594, 0.05418384170532226, 0.05351225662231445, 0.05430793762207031, 0.053991870880126955, 0.05392243194580078, 0.05623580932617187, 0.05609804916381836, 0.055782142639160155, 0.05396889495849609, 0.05352566528320313, 0.05322633743286133, 0.052937889099121095, 0.053522880554199216, 0.05347369766235351, 0.053217281341552736, 0.05351424026489258, 0.05342403030395508, 0.05407344055175781, 0.053840961456298825, 0.05401696014404297, 0.05386444854736328, 0.05429452896118164, 0.0547729606628418, 0.05405344009399414, 0.055820510864257815, 0.05496207809448242, 0.05411849594116211, 0.05346713638305664, 0.05389648056030273, 0.054086368560791014, 0.053394912719726566, 0.05381584167480469, 0.053808673858642575, 0.05351881790161133, 0.054335136413574216, 0.05401430511474609, 0.053561214447021485, 0.054986495971679684, 0.054745471954345704, 0.054142112731933596, 0.054557537078857424, 0.05500723266601563, 0.05447884750366211, 0.05493350219726562, 0.0546602897644043, 0.05416633605957031, 0.054085792541503905, 0.053671775817871095, 0.053944255828857424, 0.05356345748901367, 0.05412659072875976, 0.05954150390625, 0.05396835327148437, 0.05328131103515625, 0.05309027099609375, 0.05330243301391602, 0.05285315322875977, 0.05304947280883789, 0.05320671844482422, 0.05288003158569336, 0.0533043212890625, 0.05348863983154297, 0.05351542282104492, 0.05453500747680664, 0.054470497131347655, 0.053963966369628906, 0.05402710342407226, 0.05407712173461914, 0.0556036491394043, 0.055894016265869144, 0.05419417572021484, 0.05345075225830078, 0.05343231964111328, 0.05320697784423828, 0.05344668960571289, 0.0534600944519043, 0.053214111328125, 0.053143520355224606, 0.05354828643798828, 0.053590656280517575, 0.053665950775146486, 0.05412249755859375, 0.054736351013183596, 0.054198814392089845, 0.05436972808837891, 0.054958049774169924, 0.05567343902587891, 0.05393203353881836, 0.05496329498291016, 0.054535072326660154, 0.053792896270751955, 0.05355507278442383, 0.05439897537231445, 0.05386582565307617, 0.05376476669311524, 0.053940223693847655, 0.05385363388061523, 0.053994049072265626, 0.0539504623413086, 0.05465292739868164, 0.05458905410766601, 0.05444230270385742, 0.0550503044128418, 0.05508729553222656, 0.055305568695068356, 0.05503392028808594, 0.05446902465820312, 0.054451969146728514, 0.05412451171875, 0.05449091339111328, 0.053622753143310546, 0.05394803237915039, 0.053911872863769535, 0.05355785751342773, 0.05910028839111328, 0.053558143615722656, 0.05332787322998047, 0.05288345718383789, 0.053430145263671874, 0.05327171325683594, 0.05264499282836914, 0.0534628791809082, 0.05339750289916992, 0.053278785705566406, 0.0539068489074707, 0.05454083251953125, 0.053564800262451175, 0.05398585510253906, 0.054146720886230466, 0.05356380844116211, 0.05424089431762695, 0.0550516471862793, 0.05475753784179688, 0.05419913482666015, 0.05403238296508789, 0.05357904052734375, 0.053553760528564455, 0.05366182327270508, 0.05359552001953125, 0.05350668716430664, 0.0541712646484375, 0.05384368133544922, 0.053659904479980466, 0.05419440078735351, 0.05412384033203125, 0.05375065612792969, 0.054363872528076174, 0.05438006210327148, 0.0542031364440918, 0.05463859176635742, 0.054849281311035156, 0.054690048217773436, 0.053811294555664066, 0.05401795196533203, 0.054261760711669924, 0.05386380767822266, 0.05377654266357422, 0.05429296112060547, 0.053884929656982425, 0.05438032150268555, 0.05414524841308594, 0.053544960021972655, 0.054301887512207034, 0.054303550720214845, 0.05436620712280273, 0.05455241775512695, 0.054562271118164064, 0.05504697418212891, 0.054392704010009764, 0.054458366394042966, 0.05479401779174805, 0.054290462493896484, 0.05450156784057617, 0.0541921272277832, 0.05415484619140625, 0.05426217651367188, 0.054666400909423825, 0.059961856842041014, 0.05433375930786133, 0.053098495483398435, 0.05336576080322265, 0.0533265266418457, 0.05286848068237305, 0.05323462295532227, 0.05332787322998047, 0.05348777770996094, 0.05357542419433594, 0.0539257926940918, 0.05396003341674805, 0.05468793487548828, 0.054233760833740235, 0.053905441284179685, 0.053972766876220706, 0.054026432037353515, 0.05496627044677734, 0.05482700729370117, 0.05414096069335937, 0.05411836624145508, 0.05413475036621094, 0.05379244613647461, 0.05353871917724609, 0.053129024505615234, 0.05357017517089844, 0.053710849761962894, 0.05391974258422851, 0.054281471252441406, 0.054226753234863284, 0.05436524963378906, 0.05457052612304687, 0.0542006721496582, 0.05401190567016602, 0.055113601684570315, 0.05528793716430664, 0.054875839233398435, 0.05382099151611328, 0.054917793273925784, 0.05421884918212891, 0.05374771118164062, 0.053833728790283204, 0.05436415863037109, 0.05361648178100586, 0.053913761138916015, 0.05458060836791992, 0.05366041564941406, 0.054499137878417966, 0.05434579086303711, 0.05401152038574219, 0.054661502838134764, 0.055093246459960936, 0.05503180694580078, 0.05451571273803711, 0.05507276916503906, 0.054970367431640625, 0.05446656036376953, 0.05434502410888672, 0.054346431732177736, 0.055459712982177736, 0.05410940933227539, 0.05424390411376953, 0.05480483245849609, 0.05947939300537109, 0.053601119995117186, 0.05336288070678711, 0.05308601760864258, 0.05309868621826172, 0.05319456100463867, 0.05349504089355469, 0.05345356750488281, 0.05353577423095703, 0.05387964630126953, 0.053467262268066404, 0.05391769790649414, 0.05455276870727539, 0.05381439971923828, 0.05412457656860352, 0.054010528564453125, 0.05400092697143555, 0.0553045768737793, 0.05467171096801758, 0.054234302520751954, 0.0539799690246582, 0.05370470428466797, 0.05351001739501953, 0.05346112060546875, 0.054077438354492184, 0.05378047943115234, 0.0533135986328125, 0.05384969711303711, 0.05409814453125, 0.05395059204101563, 0.054755199432373045, 0.05432252883911133, 0.05350051116943359, 0.05466681671142578, 0.05584550476074219, 0.05523251342773437, 0.05424662399291992, 0.05419702529907226, 0.0538787841796875, 0.053808895111083985, 0.05419638442993164, 0.05421596908569336, 0.053743488311767576, 0.053898174285888674, 0.05429967880249023, 0.05416016006469727, 0.05388102340698242, 0.0549150390625, 0.054574718475341795, 0.05431132888793945, 0.05403763198852539, 0.05468044662475586, 0.0544351692199707, 0.056000415802001956, 0.05517388916015625, 0.05433958435058594, 0.05394204711914063, 0.054499359130859376, 0.054143169403076175, 0.05351628875732422, 0.05455270385742188, 0.054429569244384766, 0.053972160339355466, 0.058772926330566404, 0.053666656494140624, 0.05321712112426758, 0.0528416633605957, 0.053312480926513674, 0.05324537658691406, 0.05285257720947266, 0.05370675277709961, 0.05401264190673828, 0.053340160369873046, 0.05450096130371094, 0.05402582550048828, 0.054395423889160154, 0.053967041015625, 0.054098014831542966, 0.05406281661987305, 0.05387424087524414, 0.05482700729370117, 0.05479497528076172, 0.05413888168334961, 0.05398454284667969, 0.05349004745483398, 0.05347689437866211, 0.05342816162109375, 0.05342201614379883, 0.05376467132568359, 0.05387203216552734, 0.054196800231933594, 0.054006145477294924, 0.05450140762329102, 0.0548752326965332, 0.05456943893432617, 0.05396934509277344, 0.05504000091552735, 0.05532467269897461, 0.05433724975585937, 0.05436361694335937, 0.05466156768798828, 0.0542314224243164, 0.05371686553955078, 0.053897342681884765, 0.05376176071166992, 0.05383168029785156, 0.05451395034790039, 0.054317054748535154, 0.05412822341918945, 0.054319232940673826, 0.05486211013793945, 0.054494239807128905, 0.05454947280883789, 0.05506867218017578, 0.05449852752685547, 0.05453084945678711, 0.0551629753112793, 0.05479414367675781, 0.054370304107666016, 0.05588281631469726, 0.05455353546142578, 0.05425551986694336, 0.05413817596435547, 0.05429328155517578, 0.05457676696777344, 0.0541082878112793, 0.05835190582275391, 0.053450782775878905, 0.053184513092041016, 0.0531578254699707, 0.05326787185668945, 0.05318060684204102, 0.05335871887207031, 0.05339078521728516, 0.053908382415771484, 0.05402009582519531, 0.05396438217163086, 0.05452560043334961, 0.05352495956420898, 0.05409222412109375, 0.054273887634277346, 0.0534015998840332, 0.05475708770751953, 0.055295806884765625, 0.054903263092041014, 0.05426995086669922, 0.05382963180541992, 0.05423865509033203, 0.05362540817260742, 0.053766143798828124, 0.05368435287475586, 0.05433536148071289, 0.05417331314086914, 0.05389936065673828, 0.05422835159301758, 0.05403535842895508, 0.054199295043945314, 0.053907711029052736, 0.054698753356933597, 0.055260990142822264, 0.05483292770385742, 0.054505889892578124, 0.05528579330444336, 0.054693599700927735, 0.05399577713012695, 0.05425539016723633, 0.05402796936035156, 0.053508159637451175, 0.054335968017578125, 0.05392745590209961, 0.05432163238525391, 0.05439078521728516, 0.053884929656982425, 0.05481369781494141, 0.05404569625854492, 0.05443791961669922, 0.054426593780517576, 0.0541736946105957, 0.05548339080810547, 0.05486524963378906, 0.055231136322021486, 0.054417407989501954, 0.05450758361816406, 0.05512524795532227, 0.05412524795532227, 0.05403647994995117, 0.05449728012084961, 0.054489086151123044, 0.05404022216796875, 0.058654720306396485, 0.054024192810058595, 0.05357468795776367, 0.052875328063964847, 0.053297950744628904, 0.05327065658569336, 0.05304288101196289, 0.053655872344970705, 0.0538419189453125, 0.05377433776855469, 0.054063102722167966, 0.054542335510253906, 0.05402153778076172, 0.05564604949951172, 0.054986495971679684, 0.05351068878173828, 0.05411888122558594, 0.05566463851928711, 0.05509312057495117, 0.05405913543701172, 0.05351216125488281, 0.053593982696533204, 0.05327385711669922, 0.054053375244140625, 0.05402256011962891, 0.053720191955566404, 0.05390835189819336, 0.0541921272277832, 0.05385171127319336, 0.05434400177001953, 0.05522819137573242, 0.054450431823730466, 0.05432710266113281, 0.054507808685302736, 0.05499059295654297, 0.054917057037353514, 0.0543721923828125, 0.05539683151245117, 0.05400073623657226, 0.05385683059692383, 0.05431331253051758, 0.05384966278076172, 0.0544277114868164, 0.05406278228759766, 0.05337363052368164, 0.054443809509277345, 0.054296321868896484, 0.054344158172607425, 0.054566913604736325, 0.054844768524169925, 0.05551785659790039, 0.0542023696899414, 0.0551893424987793, 0.0547465934753418, 0.05512851333618164, 0.054581504821777344, 0.0542658576965332, 0.05458848190307617, 0.05460838317871094, 0.05412704086303711, 0.05456281661987305, 0.053866497039794924, 0.05394800186157227, 0.05939129638671875, 0.053627582550048826, 0.05331513595581055, 0.05310508728027344, 0.05358313751220703, 0.0533240966796875, 0.05340147018432617, 0.05342057418823242, 0.05391155242919922, 0.05346028900146484, 0.05396140670776367, 0.05395251083374023, 0.05431628799438477, 0.0541396484375, 0.05458451080322266, 0.05399951934814453, 0.05457193756103516, 0.0549450569152832, 0.05433216094970703, 0.05409584045410156, 0.05395878219604492, 0.053571071624755856, 0.053682334899902345, 0.054145153045654294, 0.05420041656494141, 0.05389311981201172, 0.05388035202026367, 0.053639583587646485, 0.054185535430908205, 0.05405952072143555, 0.05455257415771484, 0.05467136001586914, 0.05459267044067383, 0.0545145263671875, 0.05528284835815429, 0.05444489669799805, 0.054926559448242186, 0.054854015350341796, 0.054304512023925784, 0.053932510375976565, 0.054612159729003906, 0.053833728790283204, 0.053921791076660154, 0.054454113006591795, 0.054327457427978516, 0.053804512023925784, 0.05437807846069336, 0.05437916946411133, 0.05446684646606445, 0.0551464958190918, 0.054679550170898435, 0.05440441513061523, 0.05502588653564453, 0.05489286422729492, 0.054423583984375, 0.054343841552734376, 0.05505020904541016, 0.05473459243774414, 0.05401011276245117, 0.054236705780029294, 0.05481929779052734, 0.05461510467529297, 0.054268863677978514]",tokens/s,18.450757544632776,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,791.773184,537.853952,0.0,159.383552,141.760512,s,1,7.665677734375,7.665677734375,0.0,7.665677734375,7.665677734375,7.665677734375,7.665677734375,[7.665677734375],,kWh,1.739781461252126e-05,1.911688780291314e-06,5.643060069976391e-06,2.4952563462788968e-05,,MB,1160.425472,630.12864,0.0,222.298112,184.771584,s,24,0.19141417598724367,0.007975590666135152,6.836737458771063e-05,0.007971375942230224,0.008055766487121582,0.008078140878677368,0.008162211923599243,"[0.007951168060302735, 0.008186271667480469, 0.007896448135375976, 0.007883999824523926, 0.007929376125335693, 0.007905183792114258, 0.007932703971862793, 0.007983871936798096, 0.007885727882385254, 0.00795904016494751, 0.007985151767730713, 0.007983551979064942, 0.0079234881401062, 0.007959199905395507, 0.00805014419555664, 0.008058176040649415, 0.007984896183013917, 0.007949503898620605, 0.007914175987243652, 0.007996895790100097, 0.008027520179748535, 0.007984992027282714, 0.008001024246215821, 0.008081664085388183]",tokens/s,32097.93615499749,kWh,2.3439026773762346e-07,2.5849197541382268e-08,1.3309247595179246e-07,3.933319412307982e-07,tokens/kWh,650849761.1430571,MB,1194.758144,630.12864,0.0,222.298112,184.774144,s,24,9.903674468994138,0.4126531028747558,0.0038970866275595925,0.4117662048339844,0.41466705017089844,0.4213199203491211,0.42564024017333985,"[0.42658013916015625, 0.41120654296875, 0.4101144409179687, 0.41051300048828127, 0.4092403869628906, 0.4224936218261719, 0.41035195922851564, 0.4108104248046875, 0.41123602294921874, 0.4097622375488281, 0.4129798889160156, 0.41197479248046875, 0.4123426208496094, 0.4128636169433594, 0.41315121459960935, 0.4113250427246094, 0.4105936889648438, 0.41165512084960937, 0.4091098937988281, 0.41213641357421876, 0.41466262817382815, 0.4146689453125, 0.4118772888183594, 0.4120245361328125]",tokens/s,152.67060773591496,kWh,1.1901169499595156e-05,1.3124532521006592e-06,4.383788915328133e-06,1.7597411667023947e-05,tokens/kWh,3580071.955585187,,s,1512,9.89292088890076,0.006542937095833834,0.00020275533170362382,0.006507872104644775,0.006608620595932007,0.006682943940162659,0.00737621145725251,"[0.006587135791778565, 0.007636703968048096, 0.00865443229675293, 0.008525823593139649, 0.00865283203125, 0.00781379222869873, 0.006614208221435547, 0.006588223934173584, 0.0065976958274841305, 0.0066061439514160155, 0.007475840091705322, 0.006551551818847656, 0.006547679901123047, 0.006542175769805908, 0.006562911987304688, 0.006509791851043701, 0.006513279914855957, 0.006706975936889649, 0.006613215923309326, 0.006574336051940918, 0.006596352100372314, 0.006586143970489502, 0.006591743946075439, 0.006601471900939942, 0.006570208072662354, 0.006537568092346191, 0.006555168151855468, 0.006514815807342529, 0.006569983959197998, 0.006529248237609863, 0.006581471920013428, 0.006617663860321045, 0.006567935943603515, 0.006555424213409424, 0.006621407985687256, 0.006614143848419189, 0.006610112190246582, 0.006618112087249756, 0.006645472049713135, 0.006672832012176513, 0.006723999977111816, 0.006668416023254395, 0.006645343780517578, 0.0066769919395446775, 0.006643968105316162, 0.006655648231506348, 0.006668288230895996, 0.006762495994567871, 0.006616703987121582, 0.006668831825256348, 0.006622464179992676, 0.006660704135894775, 0.006670239925384521, 0.00668236780166626, 0.006638175964355468, 0.006663936138153076, 0.0066620478630065915, 0.006674752235412597, 0.006706975936889649, 0.006692863941192627, 0.006649856090545654, 0.006696991920471192, 0.006704927921295166, 0.006633471965789795, 0.0066488637924194335, 0.006638815879821777, 0.006605696201324463, 0.0065606398582458494, 0.006555647850036621, 0.006555647850036621, 0.006714879989624023, 0.006576159954071045, 0.0065428800582885745, 0.006560224056243897, 0.00658457612991333, 0.006541535854339599, 0.006583295822143555, 0.006566559791564942, 0.006558047771453857, 0.0065474557876586915, 0.00656547212600708, 0.006551008224487304, 0.006538176059722901, 0.0065491518974304195, 0.006569632053375244, 0.006580192089080811, 0.00653385591506958, 0.0065474557876586915, 0.006497663974761963, 0.006492800235748291, 0.006463263988494873, 0.006485631942749023, 0.006561984062194824, 0.006480288028717041, 0.0064650559425354, 0.006480127811431885, 0.006452799797058106, 0.006474400043487549, 0.0065146880149841305, 0.006479743957519531, 0.006498688220977783, 0.006490079879760742, 0.006479328155517578, 0.006504447937011719, 0.006455615997314453, 0.006572127819061279, 0.0064774718284606935, 0.006445280075073242, 0.006516416072845459, 0.006479296207427978, 0.0064701118469238286, 0.006468031883239746, 0.006463488101959228, 0.006498655796051025, 0.006487455844879151, 0.006467840194702148, 0.006475776195526123, 0.006476831912994385, 0.006456639766693115, 0.006452896118164062, 0.006494207859039307, 0.006511680126190186, 0.006466495990753174, 0.0064471039772033695, 0.006487840175628662, 0.006457151889801026, 0.006372384071350098, 0.006488863945007324, 0.006446656227111817, 0.006480512142181397, 0.006469632148742676, 0.006444255828857422, 0.006480415821075439, 0.0064924159049987796, 0.006495327949523926, 0.00658735990524292, 0.0064982399940490725, 0.006526976108551025, 0.0065064959526062015, 0.006468607902526856, 0.006537280082702637, 0.006509632110595703, 0.0064691839218139644, 0.006479519844055176, 0.006480000019073486, 0.006488031864166259, 0.006475808143615723, 0.006514944076538086, 0.0064980158805847165, 0.006459455966949463, 0.006476287841796875, 0.006512639999389648, 0.006475776195526123, 0.006516672134399414, 0.006501760005950928, 0.006480576038360596, 0.006524928092956543, 0.006565023899078369, 0.0064951682090759275, 0.006539167881011963, 0.00738483190536499, 0.006539519786834717, 0.006543360233306885, 0.006489471912384033, 0.006501279830932617, 0.00651958417892456, 0.006466752052307129, 0.006552768230438232, 0.0065214080810546875, 0.0064943361282348636, 0.006537087917327881, 0.006551743984222412, 0.006464384078979492, 0.006470848083496094, 0.006444799900054932, 0.0064921917915344235, 0.006457664012908935, 0.006454783916473389, 0.006494592189788819, 0.006427807807922363, 0.006481728076934814, 0.006478623867034912, 0.006478079795837402, 0.006448927879333496, 0.0064633598327636715, 0.006420608043670654, 0.006453536033630371, 0.006419519901275635, 0.006466047763824463, 0.00634611177444458, 0.0071617598533630375, 0.006481632232666016, 0.006487008094787598, 0.0064650559425354, 0.0064774398803710935, 0.006479968070983886, 0.006462240219116211, 0.00647603178024292, 0.006491903781890869, 0.0064778242111206055, 0.006449344158172607, 0.00648524808883667, 0.006486591815948487, 0.006486015796661377, 0.006454304218292236, 0.006448095798492432, 0.006492159843444824, 0.0064737281799316405, 0.006446271896362305, 0.006457695960998535, 0.006458111763000488, 0.006815455913543701, 0.0066847681999206545, 0.0066967358589172365, 0.006506624221801757, 0.006452640056610107, 0.006478432178497314, 0.006448959827423096, 0.006463679790496826, 0.006463488101959228, 0.006464735984802246, 0.006484543800354004, 0.006461311817169189, 0.00649455976486206, 0.006446623802185058, 0.0064512319564819336, 0.0064926080703735355, 0.006479072093963623, 0.0064560317993164066, 0.0064609599113464354, 0.006478367805480957, 0.006549503803253174, 0.006495456218719483, 0.006460192203521728, 0.006479872226715088, 0.006473919868469238, 0.006739264011383057, 0.006465343952178955, 0.006441664218902588, 0.006481919765472412, 0.006485119819641113, 0.006455679893493652, 0.006530687808990478, 0.00653331184387207, 0.006447455883026123, 0.006463232040405273, 0.006476384162902832, 0.006731776237487793, 0.006544864177703858, 0.006664735794067383, 0.006565887928009034, 0.006465536117553711, 0.0063688640594482425, 0.006464288234710693, 0.00646284818649292, 0.00646998405456543, 0.006483839988708496, 0.0064512319564819336, 0.006480000019073486, 0.006457215785980225, 0.006450911998748779, 0.00644927978515625, 0.006426559925079346, 0.006482336044311523, 0.006551487922668457, 0.0065064959526062015, 0.006582560062408448, 0.006475327968597412, 0.006479135990142822, 0.006468480110168457, 0.0064924159049987796, 0.006479199886322022, 0.006457759857177735, 0.006475776195526123, 0.006483168125152588, 0.006476480007171631, 0.006459487915039063, 0.006529056072235108, 0.006447072029113769, 0.00652288007736206, 0.006498271942138672, 0.0064982399940490725, 0.006578271865844727, 0.006486015796661377, 0.006445055961608887, 0.006477280139923096, 0.0064655041694641115, 0.0065326719284057614, 0.006470335960388184, 0.006455615997314453, 0.006494207859039307, 0.006643455982208252, 0.006457600116729736, 0.006793216228485107, 0.006454368114471435, 0.006598656177520752, 0.00648035192489624, 0.006491648197174072, 0.0064802560806274416, 0.00646611213684082, 0.006464672088623047, 0.006445919990539551, 0.006440959930419922, 0.006467584133148193, 0.0064646081924438475, 0.006448031902313233, 0.006486015796661377, 0.006459392070770263, 0.006465536117553711, 0.006457119941711426, 0.006537439823150634, 0.006510591983795166, 0.0065146880149841305, 0.006510816097259522, 0.006520607948303223, 0.006460383892059326, 0.006473760128021241, 0.00649948787689209, 0.00643507194519043, 0.006522528171539307, 0.006701119899749756, 0.007994080066680908, 0.007450751781463623, 0.007647264003753662, 0.010730976104736329, 0.008782015800476075, 0.006592864036560058, 0.006547103881835938, 0.006536960124969482, 0.006568543910980225, 0.006553599834442139, 0.0065476160049438474, 0.006539103984832764, 0.006522111892700196, 0.00652675199508667, 0.006994847774505615, 0.006544896125793457, 0.006552127838134766, 0.006567935943603515, 0.006490111827850342, 0.00658022403717041, 0.006498303890228272, 0.0064839677810668945, 0.006540512084960938, 0.006541728019714355, 0.006515071868896484, 0.006999807834625244, 0.006545631885528564, 0.006501599788665772, 0.00658457612991333, 0.00651142406463623, 0.00651852798461914, 0.006505856037139893, 0.0065829439163208004, 0.006561376094818115, 0.006494592189788819, 0.006510591983795166, 0.006500351905822754, 0.006496255874633789, 0.006490111827850342, 0.0065268478393554685, 0.00646889591217041, 0.006490975856781006, 0.00653110408782959, 0.00651475191116333, 0.006516416072845459, 0.00648524808883667, 0.006493152141571045, 0.006492159843444824, 0.006502399921417237, 0.0065409598350524905, 0.006649631977081299, 0.0065277438163757326, 0.006528831958770752, 0.006534848213195801, 0.006512351989746094, 0.006510528087615966, 0.006476448059082031, 0.0063788161277770995, 0.006484096050262451, 0.0064579200744628905, 0.006480991840362549, 0.00644598388671875, 0.006457024097442627, 0.006492479801177979, 0.006561791896820069, 0.006467743873596192, 0.006456960201263428, 0.006453216075897217, 0.0064739837646484374, 0.00655344009399414, 0.006477983951568603, 0.0065064959526062015, 0.006490111827850342, 0.006455103874206543, 0.006709375858306885, 0.006467167854309082, 0.006490560054779053, 0.006469664096832276, 0.0064719681739807125, 0.006522592067718506, 0.0064880638122558594, 0.0065593280792236325, 0.006533472061157226, 0.0065270400047302245, 0.006545407772064209, 0.006551551818847656, 0.006492224216461182, 0.006518720149993896, 0.006503744125366211, 0.006520959854125976, 0.006514976024627686, 0.006508831977844238, 0.006502016067504883, 0.006487711906433106, 0.0064960322380065915, 0.006511551856994629, 0.0065146880149841305, 0.006508543968200684, 0.006526976108551025, 0.006539552211761474, 0.006532832145690918, 0.006516992092132568, 0.00650764799118042, 0.006531424045562744, 0.006501760005950928, 0.006507423877716065, 0.006557695865631104, 0.006508543968200684, 0.006494207859039307, 0.006506303787231446, 0.006488383769989014, 0.006534815788269043, 0.006500576019287109, 0.006492159843444824, 0.0065372161865234375, 0.006533504009246826, 0.0065320320129394534, 0.00652672004699707, 0.006503359794616699, 0.006516736030578613, 0.006410496234893799, 0.006513088226318359, 0.0065426878929138186, 0.0065129919052124025, 0.006489568233489991, 0.006554687976837158, 0.006514815807342529, 0.006500383853912353, 0.006524991989135742, 0.006506207942962646, 0.006543360233306885, 0.0065491838455200195, 0.006543680191040039, 0.0065372161865234375, 0.006512639999389648, 0.006510591983795166, 0.006561600208282471, 0.0065512962341308595, 0.006521279811859131, 0.006547488212585449, 0.006540383815765381, 0.0065831680297851565, 0.006518335819244385, 0.006508992195129394, 0.006619455814361572, 0.006518464088439941, 0.00649129581451416, 0.006521344184875488, 0.006510335922241211, 0.006488671779632568, 0.0065186238288879395, 0.00648198413848877, 0.006542719841003418, 0.006506591796875, 0.006484608173370362, 0.0065064959526062015, 0.006602176189422607, 0.006502975940704345, 0.006499648094177246, 0.006458176136016846, 0.006530943870544433, 0.006520832061767578, 0.006471680164337158, 0.006533120155334473, 0.006526400089263916, 0.0064988799095153805, 0.0065146880149841305, 0.006453248023986816, 0.00650761604309082, 0.00648857593536377, 0.006464128017425537, 0.006518688201904297, 0.006539135932922363, 0.006498303890228272, 0.006507808208465576, 0.006494944095611572, 0.00652288007736206, 0.006492159843444824, 0.00646892786026001, 0.006502079963684082, 0.006501376152038574, 0.006479551792144775, 0.0064618239402770995, 0.006415296077728272, 0.006481887817382813, 0.006477791786193848, 0.006472703933715821, 0.006472320079803467, 0.0064899840354919434, 0.0066052799224853515, 0.00662937593460083, 0.006730016231536865, 0.006542111873626709, 0.006521312236785889, 0.006484447956085205, 0.0065484800338745115, 0.0064702401161193844, 0.006455552101135254, 0.006486112117767334, 0.0064570560455322265, 0.006463935852050781, 0.006459296226501465, 0.006468992233276367, 0.006484608173370362, 0.006508543968200684, 0.006481728076934814, 0.006521344184875488, 0.006471007823944092, 0.006451680183410645, 0.006502272129058838, 0.006475776195526123, 0.006518400192260742, 0.006531455993652344, 0.006483808040618897, 0.006525087833404541, 0.006526976108551025, 0.006782368183135986, 0.00666860818862915, 0.0065716800689697265, 0.0066341118812561035, 0.006496255874633789, 0.006497663974761963, 0.006548096179962158, 0.006524159908294678, 0.006521599769592285, 0.006653952121734619, 0.006530784130096436, 0.006512864112854004, 0.00660316801071167, 0.006510240077972412, 0.006497695922851563, 0.0065296320915222165, 0.0065372161865234375, 0.00649232006072998, 0.00651584005355835, 0.006505055904388428, 0.006559872150421143, 0.006499328136444092, 0.006506815910339355, 0.006471519947052002, 0.006490880012512207, 0.006542592048645019, 0.006486879825592041, 0.006485824108123779, 0.006516736030578613, 0.006477952003479004, 0.0064011201858520505, 0.00648419189453125, 0.006496607780456543, 0.0065630397796630855, 0.006503071784973145, 0.006477952003479004, 0.006502528190612793, 0.006481855869293213, 0.006494143962860108, 0.0065413122177124024, 0.006504447937011719, 0.006489727973937988, 0.006482367992401123, 0.006464511871337891, 0.0065155520439147945, 0.006495456218719483, 0.00645414400100708, 0.006473855972290039, 0.00650870418548584, 0.006481696128845215, 0.006493951797485352, 0.006457824230194092, 0.006526944160461426, 0.006500351905822754, 0.006489855766296387, 0.006497983932495117, 0.006461215972900391, 0.006470176219940185, 0.006502495765686035, 0.006479775905609131, 0.006591551780700684, 0.006499263763427734, 0.006531136035919189, 0.006500480175018311, 0.006487872123718262, 0.006494431972503662, 0.006516511917114258, 0.006479872226715088, 0.006529024124145508, 0.006492159843444824, 0.006480063915252686, 0.006543168067932129, 0.006500576019287109, 0.006474688053131104, 0.006484511852264404, 0.006465951919555664, 0.006500256061553955, 0.006479616165161133, 0.006668320178985596, 0.006512576103210449, 0.006555935859680176, 0.006445248126983643, 0.006459231853485107, 0.006432735919952392, 0.006483839988708496, 0.0064759039878845214, 0.006458816051483155, 0.0065214080810546875, 0.006511616230010986, 0.006501599788665772, 0.006516448020935059, 0.006502624034881592, 0.006528031826019287, 0.006418752193450928, 0.006481599807739258, 0.006567935943603515, 0.006515903949737549, 0.006474048137664795, 0.006600863933563232, 0.006490464210510254, 0.006463488101959228, 0.006508096218109131, 0.006488096237182617, 0.0064966721534729, 0.006500351905822754, 0.006445216178894043, 0.0065103678703308104, 0.006530144214630127, 0.00649724817276001, 0.006579520225524903, 0.006587071895599365, 0.006534207820892334, 0.006521632194519043, 0.006529024124145508, 0.006531231880187989, 0.006516736030578613, 0.006500351905822754, 0.006504191875457764, 0.006500095844268799, 0.006485631942749023, 0.006525824069976807, 0.0064839677810668945, 0.006498400211334228, 0.006549407958984375, 0.006538911819458008, 0.0065474557876586915, 0.006510655879974365, 0.006498591899871826, 0.006533023834228516, 0.006502624034881592, 0.006555520057678223, 0.006513792037963867, 0.006506944179534912, 0.006553055763244629, 0.006527967929840088, 0.0066293120384216305, 0.0065229439735412596, 0.0065045437812805175, 0.007114848136901855, 0.007272255897521973, 0.006692863941192627, 0.006533120155334473, 0.006524576187133789, 0.006531424045562744, 0.006500351905822754, 0.006543360233306885, 0.006563839912414551, 0.006502399921417237, 0.00652288007736206, 0.006653215885162354, 0.006554624080657959, 0.0065263681411743164, 0.006512800216674805, 0.006625696182250977, 0.006510335922241211, 0.006564000129699707, 0.0064876160621643066, 0.006506847858428955, 0.006535583972930908, 0.006803135871887207, 0.0065651841163635255, 0.006643871784210205, 0.006520639896392822, 0.0065277118682861325, 0.0064973759651184085, 0.006498847961425781, 0.0064802560806274416, 0.00651200008392334, 0.006486112117767334, 0.006488224029541015, 0.006559296131134033, 0.0066176319122314455, 0.00646563196182251, 0.0064898238182067875, 0.00652345609664917, 0.0064982080459594724, 0.006485151767730713, 0.006500671863555909, 0.0065030078887939454, 0.006676576137542725, 0.006507936000823974, 0.006696608066558838, 0.006544256210327148, 0.006535168170928955, 0.006597631931304931, 0.006500671863555909, 0.006519392013549805, 0.006505472183227539, 0.00714035177230835, 0.0065474557876586915, 0.006548575878143311, 0.006468512058258056, 0.0065147199630737306, 0.0064774398803710935, 0.006567520141601562, 0.006513216018676758, 0.006496448040008545, 0.00653926420211792, 0.0065231361389160155, 0.006445151805877685, 0.0065326719284057614, 0.0065129599571228025, 0.0065491518974304195, 0.006505983829498291, 0.006468095779418945, 0.006460608005523682, 0.0064951682090759275, 0.006449120044708252, 0.006506527900695801, 0.006497632026672364, 0.006542272090911865, 0.006540128231048584, 0.006556447982788086, 0.006459487915039063, 0.006475776195526123, 0.006470816135406494, 0.006505311965942383, 0.006490111827850342, 0.006446815967559814, 0.006398272037506103, 0.006524640083312988, 0.006508543968200684, 0.006465536117553711, 0.006479648113250733, 0.006580448150634766, 0.0065577921867370605, 0.006536672115325928, 0.006601151943206787, 0.006565887928009034, 0.006658048152923584, 0.0065576000213623045, 0.006535615921020508, 0.006498976230621338, 0.006533696174621582, 0.006487648010253907, 0.0064990720748901365, 0.006547552108764648, 0.006481503963470459, 0.0065335359573364254, 0.006582335948944092, 0.006518720149993896, 0.006573056221008301, 0.006534143924713135, 0.0065064959526062015, 0.006490272045135498, 0.006487296104431152, 0.0065296320915222165, 0.006516287803649903, 0.006510079860687256, 0.00655625581741333, 0.006594528198242187, 0.0064843521118164065, 0.006471007823944092, 0.006564544200897217, 0.006570176124572754, 0.0067612800598144535, 0.006540575981140137, 0.006501344203948975, 0.0065071358680725095, 0.006584415912628174, 0.006576128005981445, 0.006494207859039307, 0.006489247798919678, 0.006518879890441895, 0.006488831996917724, 0.006509568214416504, 0.006517471790313721, 0.00652239990234375, 0.006480127811431885, 0.006523392200469971, 0.00653056001663208, 0.006519296169281006, 0.006511871814727783, 0.006477888107299805, 0.00672435188293457, 0.0065322241783142086, 0.006816351890563965, 0.0065963840484619144, 0.006532608032226562, 0.006577087879180908, 0.006512639999389648, 0.0065064959526062015, 0.006429440021514893, 0.006573056221008301, 0.006622208118438721, 0.006577631950378418, 0.006547135829925537, 0.006497407913208008, 0.006586080074310303, 0.00655344009399414, 0.006535071849822998, 0.0064906878471374515, 0.006501183986663818, 0.006617983818054199, 0.006527167797088623, 0.006512383937835694, 0.00652623987197876, 0.00652345609664917, 0.0065099201202392576, 0.006556640148162842, 0.006553599834442139, 0.006541215896606446, 0.0065476479530334475, 0.0065370240211486816, 0.006620319843292236, 0.006499328136444092, 0.006487328052520752, 0.006511168003082275, 0.006516736030578613, 0.006541567802429199, 0.006678175926208496, 0.006650303840637207, 0.006611648082733154, 0.006526944160461426, 0.006503424167633057, 0.006542367935180664, 0.006556863784790039, 0.00652675199508667, 0.006913951873779297, 0.006737599849700928, 0.0064691839218139644, 0.006513792037963867, 0.006493663787841797, 0.006580480098724365, 0.006516064167022705, 0.0064784960746765135, 0.006485055923461914, 0.006480735778808594, 0.006530655860900879, 0.006506912231445313, 0.006616608142852783, 0.0065214080810546875, 0.006528160095214844, 0.006703968048095703, 0.006561408042907715, 0.006543392181396484, 0.006506912231445313, 0.00649619197845459, 0.006555776119232178, 0.006487936019897461, 0.006575647830963135, 0.0064834880828857425, 0.0064778242111206055, 0.006496799945831299, 0.006515103816986084, 0.006404096126556396, 0.006512639999389648, 0.006701344013214111, 0.006514400005340576, 0.006511839866638183, 0.006482656002044678, 0.0065229439735412596, 0.006502528190612793, 0.006469727993011475, 0.006509471893310547, 0.006491007804870605, 0.006475872039794922, 0.006475679874420166, 0.006465792179107666, 0.006553215980529785, 0.006609055995941162, 0.006461408138275147, 0.006668320178985596, 0.006487167835235595, 0.006441823959350586, 0.006500351905822754, 0.0064596481323242185, 0.006477375984191894, 0.006449183940887451, 0.006471839904785157, 0.006498303890228272, 0.006491807937622071, 0.00650275182723999, 0.0064800319671630855, 0.006467423915863037, 0.006567327976226806, 0.006501279830932617, 0.006481599807739258, 0.006557695865631104, 0.006625279903411865, 0.006768320083618164, 0.007166272163391114, 0.0072765440940856935, 0.007260159969329834, 0.006524928092956543, 0.006563488006591797, 0.006498655796051025, 0.006526976108551025, 0.006550975799560547, 0.006586944103240967, 0.006581952095031738, 0.0065030078887939454, 0.006492959976196289, 0.006582911968231201, 0.00651251220703125, 0.006486112117767334, 0.006518208026885987, 0.006517663955688476, 0.006498303890228272, 0.006520832061767578, 0.00648960018157959, 0.006479455947875977, 0.006501279830932617, 0.006455584049224854, 0.006534880161285401, 0.006550528049468994, 0.006458367824554443, 0.006504320144653321, 0.006478015899658203, 0.0064646401405334475, 0.006574783802032471, 0.006489855766296387, 0.006498688220977783, 0.006492032051086426, 0.006494239807128906, 0.0064759039878845214, 0.006622399806976318, 0.006474624156951905, 0.006502304077148437, 0.006588287830352783, 0.0064832639694213864, 0.006480832099914551, 0.006665984153747559, 0.006513984203338623, 0.006583199977874756, 0.0064633598327636715, 0.006499328136444092, 0.006476704120635986, 0.006479263782501221, 0.006485663890838623, 0.006474688053131104, 0.006498335838317871, 0.006485919952392578, 0.006524831771850586, 0.006601984024047851, 0.0065279040336608885, 0.006494048118591308, 0.006488096237182617, 0.006529151916503906, 0.006549503803253174, 0.00653107213973999, 0.006476831912994385, 0.0065335679054260255, 0.0065049281120300294, 0.006497983932495117, 0.006527359962463379, 0.0065413122177124024, 0.006519999980926514, 0.006558527946472168, 0.006502399921417237, 0.006536896228790283, 0.006514592170715332, 0.0065328960418701175, 0.006541759967803955, 0.006506688117980957, 0.006594111919403076, 0.006523327827453613, 0.0065170879364013674, 0.006546688079833984, 0.006498720169067383, 0.006535168170928955, 0.006540863990783691, 0.00662883186340332, 0.0064921917915344235, 0.006517600059509277, 0.006549600124359131, 0.0065372161865234375, 0.006512447834014893, 0.006465727806091308, 0.006525023937225342, 0.0066128640174865725, 0.006381375789642334, 0.006533664226531982, 0.00648089599609375, 0.0064642238616943355, 0.0064822077751159664, 0.006481919765472412, 0.006454912185668946, 0.00646998405456543, 0.00644649600982666, 0.006515327930450439, 0.00653107213973999, 0.006485568046569824, 0.0065642881393432615, 0.006560031890869141, 0.006473663806915283, 0.006487040042877197, 0.006476736068725586, 0.0065203838348388676, 0.0065456957817077634, 0.006520991802215576, 0.006526815891265869, 0.00654099178314209, 0.006459712028503418, 0.006469632148742676, 0.006504447937011719, 0.006489439964294434, 0.006490719795227051, 0.006719552040100098, 0.006719456195831299, 0.006551583766937256, 0.006504672050476074, 0.006539040088653565, 0.006482016086578369, 0.0065352959632873535, 0.006475584030151367, 0.00648524808883667, 0.006514400005340576, 0.006493184089660644, 0.006502399921417237, 0.0064737281799316405, 0.006508480072021485, 0.006555136203765869, 0.00647430419921875, 0.0064654722213745115, 0.006508287906646729, 0.006504096031188965, 0.006467807769775391, 0.0065251197814941405, 0.006505824089050293, 0.0064869441986083985, 0.006506048202514649, 0.00648473596572876, 0.006495872020721436, 0.006508736133575439, 0.0064992961883544925, 0.006529695987701416, 0.006533376216888428, 0.006495552062988281, 0.006539968013763428, 0.006608895778656006, 0.006533120155334473, 0.0065168957710266115, 0.006539328098297119, 0.006474080085754394, 0.0065411520004272465, 0.006487520217895508, 0.006522496223449707, 0.006509600162506103, 0.006493311882019043, 0.006552320003509521, 0.006516160011291504, 0.00650707197189331, 0.006516736030578613, 0.0064769601821899414, 0.007024735927581787, 0.006976096153259277, 0.006523039817810059, 0.006496255874633789, 0.006518176078796387, 0.006496960163116455, 0.006514592170715332, 0.006507967948913574, 0.006496831893920898, 0.006553887844085693, 0.006544191837310791, 0.006508639812469482, 0.006490431785583496, 0.006466400146484375, 0.006600351810455323, 0.006479296207427978, 0.006470208168029785, 0.006598368167877198, 0.006471776008605957, 0.006456831932067871, 0.006496960163116455, 0.0064793601036071775, 0.006486527919769287, 0.006487648010253907, 0.006490528106689453, 0.00653107213973999, 0.006496128082275391, 0.006499839782714844, 0.006496895790100097, 0.006622879981994629, 0.006603104114532471, 0.0066706881523132325, 0.006448480129241943, 0.006512256145477295, 0.006541664123535156, 0.00651094388961792, 0.0065259838104248045, 0.00648905611038208, 0.00649126386642456, 0.006615359783172608, 0.006506080150604248, 0.006492991924285888, 0.006498528003692627, 0.006489888191223144, 0.006534687995910645, 0.006482560157775879, 0.006450816154479981, 0.0064757118225097655, 0.006463903903961182, 0.006557695865631104, 0.006482304096221924, 0.006489120006561279, 0.006395008087158203, 0.0065170879364013674, 0.006468128204345703, 0.006479743957519531, 0.006462751865386963, 0.006503232002258301, 0.00647603178024292, 0.006547232151031494, 0.006508063793182373, 0.00647324800491333, 0.006448416233062744, 0.0064498882293701175, 0.006470592021942139, 0.006528927803039551, 0.006456607818603515, 0.006478367805480957, 0.006560031890869141, 0.0065372161865234375, 0.006454432010650634, 0.006451072216033936, 0.006443999767303467, 0.006493216037750244, 0.006488639831542969, 0.006459807872772217, 0.0064880638122558594, 0.006461440086364746, 0.00646943998336792, 0.006525216102600098, 0.006485919952392578, 0.006492159843444824, 0.006485119819641113, 0.00643775987625122, 0.0065372161865234375, 0.006453248023986816, 0.006492159843444824, 0.006469151973724365, 0.0064642238616943355, 0.006563583850860596, 0.006467840194702148, 0.006463232040405273, 0.00649232006072998, 0.006473567962646485, 0.006510591983795166, 0.00648364782333374, 0.006455615997314453, 0.006474815845489502, 0.006468255996704101, 0.006469791889190674, 0.00651043176651001, 0.0065632319450378415, 0.006480671882629394, 0.006504511833190918, 0.006642047882080078, 0.00652672004699707, 0.006484960079193115, 0.006470592021942139, 0.006495584011077881, 0.006455967903137207, 0.006492159843444824, 0.006479199886322022, 0.006474400043487549, 0.006467872142791748, 0.006483168125152588, 0.006377344131469727, 0.006494495868682861, 0.0064718079566955565, 0.006563168048858642, 0.006488448143005371, 0.00648633623123169, 0.006483007907867431, 0.006468192100524902, 0.0064584641456604, 0.006487167835235595, 0.006593696117401123, 0.006563551902770996, 0.00662390422821045, 0.006465407848358154, 0.006527391910552979, 0.0065146880149841305, 0.00648140811920166, 0.006550015926361084, 0.006520639896392822, 0.006455552101135254, 0.006563776016235352, 0.006472703933715821, 0.006502528190612793, 0.0064828162193298336, 0.006475552082061768, 0.006498528003692627, 0.006482175827026367, 0.006487648010253907, 0.006486464023590088, 0.006466944217681885, 0.006548096179962158, 0.006511583805084229, 0.006742784023284912, 0.00692633581161499, 0.006557727813720703, 0.006520800113677979, 0.006590464115142822, 0.0065326719284057614, 0.006510687828063965, 0.006565983772277832, 0.0065640959739685055, 0.0065493440628051755, 0.006578335762023926, 0.006516928195953369, 0.006566783905029297, 0.006574624061584473, 0.006537568092346191, 0.006584383964538574, 0.0065474557876586915, 0.006496255874633789, 0.006647808074951172, 0.006583327770233154, 0.0065361599922180175, 0.006588128089904785, 0.006517280101776123, 0.006597631931304931, 0.006524799823760986, 0.00649510383605957, 0.0065413122177124024, 0.006518367767333984, 0.0065131840705871585, 0.006536704063415527, 0.006550943851470948, 0.006432384014129639, 0.006534111976623535, 0.00657203197479248, 0.006488287925720215, 0.006635295867919922, 0.006509568214416504, 0.006486976146697998, 0.006524064064025879, 0.006503168106079101, 0.006530816078186035, 0.006517151832580567, 0.006615039825439453, 0.006497824192047119, 0.006474112033843994, 0.006475359916687012, 0.006487743854522705, 0.006470047950744629, 0.006507199764251709, 0.006499616146087647, 0.006476223945617676, 0.006487552165985107, 0.006503136157989502, 0.006483551979064942, 0.006498176097869873, 0.006455615997314453, 0.006498303890228272, 0.006490111827850342, 0.006516736030578613, 0.006526976108551025, 0.006474944114685058, 0.006478655815124512, 0.006684031963348389, 0.006851327896118164, 0.0064795842170715335, 0.006494368076324463, 0.006469632148742676, 0.006498079776763916, 0.006485631942749023, 0.0067142400741577145, 0.006530784130096436, 0.006540448188781738, 0.006471903800964356, 0.006618048191070557, 0.006499519824981689, 0.006494080066680908, 0.006482687950134278, 0.006460927963256836, 0.00655295991897583, 0.006497280120849609, 0.006471680164337158, 0.0064898238182067875, 0.006467872142791748, 0.007306464195251465, 0.007582752227783203, 0.006999807834625244, 0.006547103881835938, 0.00648367977142334, 0.007534719944000244, 0.006704639911651611, 0.006495232105255127, 0.006524479866027832, 0.00651478385925293, 0.00662278413772583, 0.00644265604019165, 0.0064685440063476565, 0.006463808059692383, 0.006510272026062012, 0.006436960220336914, 0.006488096237182617, 0.0064832639694213864, 0.006480095863342285, 0.006484511852264404, 0.006576191902160644, 0.006514431953430176, 0.0067010560035705566, 0.0067338237762451176, 0.006529024124145508, 0.006536928176879883, 0.0065047359466552735, 0.006583456039428711, 0.006539296150207519, 0.006503039836883545, 0.006527359962463379, 0.006509984016418457, 0.006486400127410889, 0.006530079841613769, 0.0065136637687683106, 0.006512224197387695, 0.006544064044952393, 0.006501535892486573, 0.006511168003082275, 0.006504703998565673, 0.006561312198638916, 0.006552095890045166, 0.006503392219543457, 0.006459680080413819, 0.0064926080703735355, 0.0065272641181945805, 0.006493120193481445, 0.0065359678268432616, 0.0065064959526062015, 0.006526976108551025, 0.006518784046173095, 0.006516799926757812, 0.006507487773895264, 0.006491104125976563, 0.006548736095428467, 0.006598527908325195, 0.0064858880043029785, 0.00651958417892456, 0.006496479988098144, 0.006518655776977539, 0.006541440010070801, 0.006498144149780273, 0.006496416091918945, 0.006477024078369141, 0.006484767913818359, 0.006553599834442139, 0.006864895820617676, 0.008388607978820802, 0.006631423950195312, 0.007474624156951905, 0.006629568099975586, 0.006662528038024903, 0.006506303787231446, 0.00653331184387207, 0.006412288188934326, 0.006497600078582764, 0.006519040107727051, 0.006521279811859131, 0.0064899840354919434, 0.006520959854125976, 0.006508543968200684, 0.006510591983795166, 0.006529376029968261, 0.006596255779266358, 0.006536799907684326, 0.006541952133178711, 0.006520607948303223, 0.006558815956115723, 0.006517663955688476, 0.00657155179977417, 0.006519264221191407, 0.0064737281799316405, 0.006508543968200684, 0.006508800029754638, 0.00651043176651001, 0.006534080028533936, 0.006495200157165528, 0.006494207859039307, 0.006764800071716309, 0.006493951797485352, 0.0071283202171325685, 0.006556000232696533, 0.006523295879364014, 0.0064839677810668945, 0.006493984222412109, 0.006467807769775391, 0.006500351905822754, 0.006493887901306152, 0.006471551895141601, 0.006636288166046143, 0.006514368057250977, 0.006498303890228272, 0.006508543968200684, 0.006461440086364746, 0.006501408100128174, 0.006517983913421631, 0.0064832639694213864, 0.0065006399154663086, 0.006541215896606446, 0.006572127819061279, 0.0064963197708129885, 0.006504447937011719, 0.00650822401046753, 0.006498720169067383, 0.00652288007736206, 0.00649616003036499, 0.0065079998970031736, 0.006484640121459961, 0.006459360122680664, 0.006490176200866699, 0.006601984024047851, 0.0065279040336608885, 0.0064899840354919434, 0.006516704082489014, 0.006494143962860108, 0.006518208026885987, 0.006795839786529541, 0.00643775987625122, 0.00651148796081543, 0.006683648109436035, 0.006484096050262451, 0.0065205440521240235, 0.006572319984436035, 0.007193664073944092, 0.006500383853912353, 0.006564479827880859, 0.006551871776580811, 0.006502079963684082, 0.006493760108947754, 0.0065008001327514645, 0.006526976108551025, 0.006552768230438232, 0.0065482878684997555, 0.006520415782928467, 0.0064966721534729, 0.006504191875457764, 0.006506752014160156, 0.006616799831390381, 0.006625664234161377, 0.006516223907470703, 0.0065335359573364254, 0.006518784046173095, 0.006492159843444824, 0.006486015796661377, 0.006555295944213867, 0.006530528068542481, 0.006518847942352295, 0.006484799861907959, 0.0065268797874450685, 0.00652297592163086, 0.006512639999389648, 0.0065127677917480465, 0.006524799823760986, 0.0066859197616577146, 0.006470431804656982, 0.0065289921760559085, 0.00655679988861084, 0.006588640213012695, 0.006509247779846191, 0.006497759819030762, 0.0064949760437011715, 0.006460319995880127, 0.006508543968200684, 0.006484864234924316, 0.0064880638122558594, 0.006481503963470459, 0.006496416091918945, 0.006452608108520508, 0.006531424045562744, 0.006701600074768066, 0.00647983980178833, 0.006574111938476563, 0.006488255977630615, 0.006486144065856934, 0.006521759986877441, 0.006478464126586914, 0.006512800216674805, 0.006475071907043457, 0.006468287944793701, 0.006506239891052246]",tokens/s,152.83656030206106,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2172.416,2198.798336,0.0,1820.327936,1730.89792,s,1,8.8145322265625,8.8145322265625,0.0,8.8145322265625,8.8145322265625,8.8145322265625,8.8145322265625,[8.8145322265625],,kWh,5.4517574154160305e-05,6.00652971744652e-06,1.8165570088005123e-05,7.868967395961195e-05,,MB,2246.254592,2343.501824,0.0,1935.671296,1893.834752,s,10,1.6642487487792967,0.16642487487792965,0.0006596255032235849,0.16620549011230468,0.1671137954711914,0.16756105575561525,0.1679188639831543,"[0.16800831604003907, 0.16573887634277343, 0.16613337707519532, 0.16627760314941406, 0.16562451171875, 0.167014404296875, 0.16610520935058593, 0.1661298522949219, 0.166712890625, 0.16650370788574217]",tokens/s,1538.2315906067076,kWh,4.944369268124862e-06,5.452589416985818e-07,3.2784424375666374e-06,8.768070647390081e-06,tokens/kWh,29196845.040955663,MB,2252.017664,2490.302464,0.0,2082.471936,1953.4336,s,10,15.85852490234375,1.585852490234375,0.014475317429753993,1.5874735717773438,1.6008872436523438,1.601325762939453,1.6016765783691407,"[1.6017642822265625, 1.5851982421875, 1.600789794921875, 1.5940562744140625, 1.58705517578125, 1.5878919677734376, 1.5962750244140624, 1.57608447265625, 1.5796583251953125, 1.5497513427734375]",tokens/s,39.72626734702744,kWh,4.576581528312513e-05,5.047719653414422e-06,2.790886029003318e-05,7.872239522657272e-05,tokens/kWh,800280.5277796523,,s,630,15.855024709701548,0.025166705888415138,0.0006164887478600973,0.025137167930603027,0.025476288795471192,0.025661361122131347,0.02766415050506593,"[0.025282880783081055, 0.02501523208618164, 0.0251013126373291, 0.025157632827758788, 0.025169919967651368, 0.02531942367553711, 0.027202655792236328, 0.032416671752929685, 0.025204736709594725, 0.02534809684753418, 0.0252620792388916, 0.025031967163085936, 0.025336544036865236, 0.025215200424194336, 0.02540028762817383, 0.025381984710693358, 0.025124576568603514, 0.025089408874511717, 0.025366943359375, 0.025346271514892577, 0.0259420166015625, 0.02517168045043945, 0.025024799346923827, 0.025298944473266603, 0.025257984161376954, 0.024887296676635744, 0.02493382453918457, 0.02548419189453125, 0.025253536224365235, 0.02515782356262207, 0.025106239318847656, 0.025001983642578125, 0.025075647354125978, 0.025120832443237304, 0.02519363212585449, 0.0250849609375, 0.025177888870239258, 0.025137184143066406, 0.025325439453125, 0.025603904724121093, 0.025297216415405274, 0.025186304092407227, 0.025095359802246094, 0.025078432083129883, 0.025792671203613282, 0.025382911682128906, 0.025382400512695313, 0.025339967727661134, 0.025913791656494142, 0.025481088638305664, 0.02530112075805664, 0.025445695877075195, 0.02531193542480469, 0.025251840591430662, 0.025341951370239257, 0.025184255599975586, 0.025653247833251954, 0.025274368286132814, 0.025276416778564452, 0.02656051254272461, 0.024985151290893556, 0.02497987174987793, 0.024931903839111327, 0.02748236846923828, 0.025526304244995118, 0.027293376922607422, 0.02734876823425293, 0.02484480094909668, 0.02480227279663086, 0.024827968597412108, 0.02485321617126465, 0.025049087524414062, 0.025184255599975586, 0.025241695404052734, 0.02501523208618164, 0.024949728012084962, 0.025001535415649413, 0.02511097526550293, 0.024938079833984376, 0.025055648803710938, 0.025067520141601563, 0.02513315200805664, 0.025143199920654297, 0.025058847427368164, 0.025028928756713868, 0.024919904708862305, 0.025123136520385742, 0.025329376220703127, 0.025128639221191407, 0.02511929512023926, 0.025126943588256834, 0.024895488739013674, 0.024866783142089843, 0.024999744415283204, 0.024813791275024415, 0.025025760650634766, 0.025106847763061522, 0.025014463424682616, 0.02506582450866699, 0.02498944091796875, 0.02514246368408203, 0.025431968688964843, 0.025017343521118163, 0.02511257553100586, 0.024997888565063478, 0.025068704605102538, 0.024923328399658204, 0.02498684883117676, 0.024953279495239258, 0.025013792037963868, 0.024924127578735352, 0.024961055755615233, 0.02509257507324219, 0.025319072723388673, 0.025124767303466796, 0.024914400100708008, 0.02531849670410156, 0.025819263458251952, 0.02512563133239746, 0.025177696228027343, 0.025131328582763672, 0.025075551986694335, 0.024833759307861327, 0.024703519821166992, 0.024635391235351564, 0.024619007110595705, 0.02555721664428711, 0.025305984497070312, 0.02543609619140625, 0.02524460792541504, 0.02567350387573242, 0.025628480911254883, 0.02556153678894043, 0.025190399169921874, 0.025476863861083984, 0.02528281593322754, 0.0252105598449707, 0.024985919952392577, 0.025100160598754882, 0.025005823135375978, 0.02521945571899414, 0.025047040939331053, 0.025122528076171876, 0.025084192276000977, 0.025054784774780275, 0.025182655334472656, 0.025161727905273438, 0.02513865661621094, 0.025037343978881837, 0.025008352279663085, 0.024942047119140626, 0.024854496002197267, 0.025038656234741212, 0.02493615913391113, 0.024777023315429688, 0.024924671173095703, 0.024919679641723633, 0.025128896713256837, 0.025135360717773437, 0.025548992156982423, 0.025320575714111327, 0.02523971176147461, 0.025069664001464844, 0.028514944076538085, 0.028008447647094727, 0.02507776069641113, 0.025203807830810547, 0.02510531234741211, 0.02514739227294922, 0.02550169563293457, 0.02527824020385742, 0.02532745552062988, 0.025348480224609376, 0.02525388717651367, 0.025856000900268555, 0.02539926338195801, 0.025372703552246093, 0.025314367294311524, 0.02527519989013672, 0.025110687255859375, 0.025724000930786133, 0.029483903884887697, 0.02532352066040039, 0.025354143142700195, 0.025460832595825194, 0.025622528076171876, 0.025268224716186522, 0.025282560348510744, 0.02524313545227051, 0.025537919998168946, 0.025334144592285158, 0.02544233512878418, 0.025211103439331056, 0.025165855407714845, 0.025168991088867186, 0.025169855117797853, 0.0251975040435791, 0.025309183120727538, 0.025211999893188477, 0.02512361526489258, 0.025100128173828125, 0.025096063613891603, 0.025557024002075195, 0.025354623794555664, 0.025235456466674806, 0.025298944473266603, 0.025548736572265626, 0.025765600204467772, 0.025552383422851564, 0.025232223510742186, 0.025214975357055663, 0.025113887786865234, 0.025213663101196288, 0.025366527557373047, 0.02513715171813965, 0.02512406349182129, 0.0251362247467041, 0.02526585578918457, 0.02534169578552246, 0.025207040786743164, 0.025161407470703126, 0.025196863174438477, 0.025233407974243165, 0.0251343994140625, 0.02503152084350586, 0.025329504013061523, 0.025241600036621094, 0.02548940849304199, 0.02509119987487793, 0.02515657615661621, 0.025414976119995117, 0.025334367752075194, 0.025207071304321288, 0.025192287445068358, 0.025266048431396484, 0.025165504455566406, 0.025181631088256835, 0.02550668716430664, 0.025377920150756836, 0.025262975692749025, 0.02521660804748535, 0.025330400466918944, 0.02554969596862793, 0.025289535522460938, 0.025264127731323242, 0.025350143432617187, 0.025298944473266603, 0.02562825584411621, 0.025667999267578127, 0.025448448181152345, 0.025649152755737304, 0.025300991058349608, 0.02551558494567871, 0.025245920181274414, 0.02535001564025879, 0.02581760025024414, 0.025167680740356444, 0.025196512222290038, 0.02520684814453125, 0.025185407638549803, 0.02509913635253906, 0.025163455963134764, 0.02515817642211914, 0.025059104919433594, 0.025243583679199218, 0.025122400283813476, 0.02506985664367676, 0.024960800170898436, 0.024977567672729493, 0.02502681541442871, 0.025018367767333984, 0.02488934326171875, 0.02505107116699219, 0.025094207763671876, 0.025100288391113282, 0.024860416412353516, 0.024793407440185548, 0.024753408432006838, 0.024978111267089844, 0.02507699203491211, 0.025071552276611328, 0.025709375381469727, 0.025208831787109375, 0.02520028877258301, 0.02514963150024414, 0.025104543685913087, 0.025223392486572266, 0.025155168533325195, 0.025045120239257812, 0.025104448318481444, 0.025165279388427733, 0.025325727462768555, 0.025029247283935546, 0.025125984191894532, 0.025100095748901367, 0.025237823486328127, 0.025053728103637696, 0.025198591232299804, 0.02524310493469238, 0.02510620880126953, 0.02539132881164551, 0.025106975555419922, 0.025173120498657227, 0.0251646728515625, 0.025198591232299804, 0.025345056533813477, 0.025312000274658204, 0.025243711471557618, 0.02727952003479004, 0.025312768936157228, 0.0250863037109375, 0.02523673629760742, 0.025066144943237306, 0.02505244827270508, 0.02500912094116211, 0.02553638458251953, 0.02530646324157715, 0.02535503959655762, 0.025149440765380858, 0.025044992446899415, 0.024975360870361327, 0.025491455078125, 0.025476224899291994, 0.025039487838745118, 0.02510873603820801, 0.02509993553161621, 0.02512345504760742, 0.025593568801879883, 0.025769952774047852, 0.02521049690246582, 0.025129472732543946, 0.025180192947387697, 0.025138208389282227, 0.02539401626586914, 0.025048416137695314, 0.025055904388427735, 0.025036800384521486, 0.025110111236572266, 0.025051551818847655, 0.024930303573608398, 0.024968351364135742, 0.02516399955749512, 0.025137792587280272, 0.025157632827758788, 0.025221120834350585, 0.025103967666625978, 0.025153951644897463, 0.025190048217773438, 0.02514364814758301, 0.025080928802490233, 0.0251125431060791, 0.025035072326660156, 0.025092735290527343, 0.02513248062133789, 0.02528927993774414, 0.02553670310974121, 0.025296703338623047, 0.025337600708007814, 0.025184576034545898, 0.02522515106201172, 0.02520412826538086, 0.025183935165405274, 0.025153696060180666, 0.02507804870605469, 0.02503932762145996, 0.02518016052246094, 0.025105567932128908, 0.025214111328125, 0.025253568649291992, 0.02513715171813965, 0.02529689598083496, 0.02507334327697754, 0.025210432052612305, 0.025266944885253908, 0.025288415908813477, 0.025343904495239256, 0.025250175476074218, 0.025316736221313477, 0.025461984634399415, 0.025258975982666014, 0.025334848403930663, 0.025379583358764647, 0.025235456466674806, 0.025117759704589845, 0.025217887878417968, 0.02500364875793457, 0.024977184295654296, 0.025023168563842773, 0.025398431777954103, 0.02517897605895996, 0.026425344467163086, 0.025206432342529297, 0.025395103454589844, 0.02516537666320801, 0.025205120086669922, 0.02521958351135254, 0.02511257553100586, 0.02501955223083496, 0.02520355224609375, 0.02503670310974121, 0.025143392562866212, 0.025609695434570312, 0.02508025550842285, 0.02515772819519043, 0.02508185577392578, 0.025071807861328125, 0.02514134407043457, 0.025134111404418947, 0.02522700881958008, 0.025098751068115235, 0.025182111740112305, 0.025141504287719725, 0.025700639724731446, 0.02501158332824707, 0.02496780776977539, 0.025022464752197264, 0.02591107177734375, 0.02494220733642578, 0.024701536178588866, 0.024805376052856445, 0.0249487361907959, 0.02492838478088379, 0.02510630416870117, 0.025231424331665038, 0.025169984817504883, 0.025087839126586915, 0.025133087158203126, 0.026482688903808595, 0.029913087844848633, 0.027738399505615234, 0.025261791229248046, 0.025392480850219726, 0.025154367446899414, 0.02521683120727539, 0.02526166343688965, 0.025190656661987304, 0.025135295867919922, 0.025075712203979493, 0.025020416259765626, 0.025515647888183595, 0.025232927322387695, 0.025396671295166016, 0.02518572807312012, 0.025437183380126953, 0.02529280090332031, 0.02500559997558594, 0.024928735733032226, 0.02496512031555176, 0.025135103225708007, 0.02533705520629883, 0.025125024795532226, 0.02519718360900879, 0.024964799880981447, 0.02491628837585449, 0.025004032135009766, 0.025036800384521486, 0.025145471572875975, 0.025138336181640623, 0.02509903907775879, 0.024936384201049804, 0.024714975357055663, 0.02480531120300293, 0.02482809638977051, 0.024735103607177733, 0.024806175231933594, 0.024725055694580077, 0.02458892822265625, 0.024631103515625, 0.024475648880004884, 0.024647743225097656, 0.024709056854248047, 0.024530048370361327, 0.024501247406005858, 0.02477462387084961, 0.02466396713256836, 0.024517663955688475, 0.024463615417480468, 0.024629535675048827, 0.02460905647277832, 0.02550559997558594, 0.02499350357055664, 0.025049407958984374, 0.025020736694335938, 0.02508297538757324, 0.025118751525878905, 0.025070144653320314, 0.025262432098388674, 0.02501180839538574, 0.025166208267211915, 0.02508799934387207, 0.02517356872558594, 0.025217472076416017, 0.025208768844604493, 0.025550912857055665, 0.025299104690551757, 0.02534281539916992, 0.02526518440246582, 0.025288671493530274, 0.025257984161376954, 0.02511894416809082, 0.025259807586669923, 0.025233407974243165, 0.025233407974243165, 0.025296255111694334, 0.025465791702270507, 0.025202463150024414, 0.025182271957397463, 0.025093696594238282, 0.02511680030822754, 0.02505958366394043, 0.025077695846557616, 0.02523993682861328, 0.02505638313293457, 0.0250067195892334, 0.02516547203063965, 0.02489993667602539, 0.02510767936706543, 0.025145536422729493, 0.025010112762451173, 0.02533033561706543, 0.026093568801879883, 0.029319168090820313, 0.02516691207885742, 0.02520364761352539, 0.02515711975097656, 0.024951295852661134, 0.024870912551879884, 0.0246965446472168, 0.02466003227233887, 0.02454924774169922, 0.02453334426879883, 0.02432204818725586, 0.024156160354614258, 0.024413440704345705, 0.024505088806152344, 0.02434611129760742, 0.024433151245117187, 0.02470911979675293, 0.026217599868774415, 0.024747135162353516, 0.02479692840576172, 0.02492416000366211, 0.0249487361907959, 0.024747392654418946, 0.024699552536010742, 0.024712543487548828, 0.02505513572692871, 0.02504572868347168, 0.024827775955200197, 0.025365951538085938, 0.027249343872070314, 0.024896703720092773, 0.02485331153869629, 0.024473600387573242, 0.024606719970703125, 0.024772064208984375, 0.025100000381469728, 0.02524652862548828, 0.02507980728149414, 0.025049087524414062, 0.025069055557250978, 0.025132640838623047, 0.02508896064758301, 0.02513711929321289, 0.025061376571655275, 0.02506883239746094, 0.025086368560791016, 0.025300416946411133, 0.025221696853637697, 0.024961376190185548, 0.02487612724304199, 0.0247445125579834, 0.024440160751342775, 0.024236799240112305, 0.024745887756347656, 0.02453708839416504, 0.024279327392578126, 0.024166112899780275, 0.023985151290893555, 0.024092927932739257, 0.02411392021179199, 0.02402521514892578, 0.024016767501831054, 0.02405171203613281, 0.02402659225463867, 0.02419152069091797, 0.024276992797851563, 0.02408963203430176, 0.024182975769042967, 0.024179487228393554, 0.024137727737426756, 0.024191999435424806, 0.024050304412841797, 0.02405414390563965, 0.023988224029541014, 0.024287008285522462, 0.024226015090942382, 0.024341791152954102, 0.024380128860473634, 0.02431795120239258, 0.024334144592285157, 0.024277151107788084, 0.02430953598022461, 0.024330495834350586, 0.02432137680053711, 0.02436777687072754, 0.02445516777038574, 0.02458163261413574, 0.024381952285766603, 0.024369152069091796, 0.02447670364379883, 0.0246363525390625, 0.024960159301757812, 0.025069759368896483, 0.025055936813354492, 0.025146944046020508, 0.025020864486694334, 0.025178016662597655, 0.02517411231994629, 0.025185823440551758, 0.025405536651611327, 0.02541401672363281, 0.025100160598754882, 0.02526630401611328, 0.025140575408935547, 0.025305952072143555, 0.025333568572998046, 0.025616384506225585, 0.025214975357055663, 0.02522217559814453]",tokens/s,39.73503741148439,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4356.624384,4562.28864,0.0,4183.81824,4182.069248,s,1,9.8477255859375,9.8477255859375,0.0,9.8477255859375,9.8477255859375,9.8477255859375,9.8477255859375,[9.8477255859375],,kWh,8.873898320000061e-05,9.781352760492478e-06,2.9000856533999923e-05,0.00012752119249449302,,MB,1599.602688,4715.380736,0.0,4307.550208,4281.174016,s,10,3.7143634033203123,0.3714363403320312,0.0021803293657190388,0.37187149047851564,0.373250830078125,0.37330784912109377,0.37335346435546873,"[0.3652881774902344, 0.37182183837890626, 0.371921142578125, 0.3733648681640625, 0.3726637268066406, 0.3722630615234375, 0.3711059875488281, 0.3711553039550781, 0.3732381591796875, 0.3715411376953125]",tokens/s,689.2163533895435,kWh,1.0779766479166616e-05,1.1880993018705978e-06,7.134093008857145e-06,1.910195878989436e-05,tokens/kWh,13401766.950488525,MB,1605.07904,4730.0608,0.0,4322.230272,4281.176576,s,10,18.6219794921875,1.86219794921875,0.009161592099191089,1.8625853881835939,1.871479797363281,1.873463653564453,1.8750507385253905,"[1.8694849853515625, 1.85217333984375, 1.8689837646484375, 1.8710389404296874, 1.870054443359375, 1.8488328857421874, 1.8560394287109374, 1.85618701171875, 1.8537371826171876, 1.875447509765625]",tokens/s,33.83098989365253,kWh,5.4357870552083466e-05,5.9959499420913065e-06,3.603005263354278e-05,9.638387312771752e-05,tokens/kWh,653636.3185625378,,s,630,18.61888622093199,0.029553787652273025,0.0006580003981161237,0.029398816108703613,0.029977926063537597,0.030449244976043698,0.03256973907470704,"[0.030482912063598634, 0.029822080612182618, 0.02954444885253906, 0.02929676818847656, 0.02929305648803711, 0.02929414367675781, 0.029298879623413085, 0.029274368286132814, 0.029339296340942383, 0.029404895782470703, 0.029364864349365236, 0.029224960327148438, 0.029251359939575197, 0.02938697624206543, 0.029298688888549803, 0.02939289665222168, 0.029327360153198243, 0.029404256820678713, 0.029612960815429686, 0.029558656692504882, 0.029438079833984374, 0.029304832458496095, 0.029226272583007813, 0.029287071228027345, 0.02920800018310547, 0.029280576705932617, 0.029329727172851563, 0.029425664901733397, 0.034744129180908204, 0.03146563148498535, 0.029431840896606446, 0.029241247177124022, 0.02941756820678711, 0.029624319076538085, 0.02933363151550293, 0.029220735549926758, 0.029247264862060546, 0.029114591598510743, 0.029079551696777343, 0.02934169578552246, 0.02898476791381836, 0.02940985679626465, 0.02918604850769043, 0.032671745300292966, 0.02939084815979004, 0.029706239700317383, 0.02942972755432129, 0.029808671951293945, 0.029558784484863283, 0.02977382469177246, 0.030174943923950197, 0.029894176483154296, 0.029827840805053712, 0.0333616943359375, 0.029632736206054687, 0.029714719772338867, 0.029402816772460937, 0.029367904663085937, 0.029554752349853514, 0.029526496887207033, 0.029449535369873048, 0.029434463500976563, 0.029308479309082033, 0.031189664840698243, 0.030099071502685548, 0.029792608261108397, 0.02973676872253418, 0.02947711944580078, 0.029155328750610353, 0.029117919921875, 0.029744800567626954, 0.029596223831176757, 0.02944236755371094, 0.029327360153198243, 0.029237247467041014, 0.02923654365539551, 0.02950009536743164, 0.02940928077697754, 0.029128704071044922, 0.02912665557861328, 0.029473983764648437, 0.029635040283203126, 0.029486528396606447, 0.029324031829833983, 0.029132959365844726, 0.02897648048400879, 0.028914335250854493, 0.029145023345947266, 0.029339712142944337, 0.029236448287963866, 0.029108287811279297, 0.02910416030883789, 0.029239967346191408, 0.029679647445678713, 0.029318143844604492, 0.02916454315185547, 0.029414495468139647, 0.029267072677612305, 0.029241119384765625, 0.029317312240600586, 0.029511295318603515, 0.029216320037841796, 0.02935251235961914, 0.029397056579589843, 0.029404287338256837, 0.03014950370788574, 0.02940342330932617, 0.0295830078125, 0.02920044708251953, 0.029449663162231444, 0.029331872940063477, 0.02970060729980469, 0.02935772705078125, 0.029190143585205077, 0.02920857620239258, 0.02912665557861328, 0.02928553581237793, 0.029288320541381835, 0.029336736679077147, 0.029085504531860353, 0.029447872161865233, 0.02923347282409668, 0.029433504104614257, 0.029501792907714843, 0.02940108871459961, 0.029453344345092773, 0.030838464736938475, 0.02966953659057617, 0.029659263610839842, 0.029404895782470703, 0.02925391960144043, 0.02943180847167969, 0.02933558464050293, 0.02935625648498535, 0.02912214469909668, 0.029366399765014647, 0.029460479736328125, 0.029773664474487305, 0.029768991470336913, 0.029858848571777345, 0.029502368927001952, 0.02967977523803711, 0.029542367935180665, 0.029950815200805662, 0.030129568099975586, 0.02993212890625, 0.02962816047668457, 0.029544704437255858, 0.02955228805541992, 0.02958950424194336, 0.029671039581298828, 0.030005983352661133, 0.02990287971496582, 0.02981270408630371, 0.029863391876220703, 0.029786624908447266, 0.030042144775390626, 0.029818143844604492, 0.029898624420166015, 0.029967199325561522, 0.03089971160888672, 0.029875968933105467, 0.030124799728393555, 0.029784063339233398, 0.030670944213867186, 0.029798080444335937, 0.029724895477294924, 0.029487104415893556, 0.02937651252746582, 0.029310592651367186, 0.02929088020324707, 0.029476896286010742, 0.029275968551635743, 0.02969820785522461, 0.029896703720092774, 0.029838783264160156, 0.029401344299316408, 0.029534528732299805, 0.02917580795288086, 0.029442047119140623, 0.02920243263244629, 0.029294591903686523, 0.0292044792175293, 0.0290631046295166, 0.030123136520385743, 0.029388736724853516, 0.029397472381591797, 0.029454879760742188, 0.029360128402709962, 0.030927135467529298, 0.0300516471862793, 0.029977279663085936, 0.030060543060302734, 0.029882368087768556, 0.029742464065551758, 0.029702783584594727, 0.03033087921142578, 0.029612031936645508, 0.029920799255371094, 0.029626848220825196, 0.029646848678588866, 0.029710336685180663, 0.02960588836669922, 0.029380607604980468, 0.029322975158691405, 0.02905116844177246, 0.02940675163269043, 0.02930940818786621, 0.029313024520874024, 0.029280256271362305, 0.02936947250366211, 0.029607967376708986, 0.029684576034545898, 0.029456384658813478, 0.029456192016601563, 0.03030035209655762, 0.03010259246826172, 0.03063596725463867, 0.029228031158447267, 0.029222911834716796, 0.029305919647216797, 0.029508544921875, 0.029247488021850586, 0.029517568588256837, 0.02932262420654297, 0.029190080642700195, 0.029260255813598632, 0.029295072555541993, 0.02928223991394043, 0.029411392211914064, 0.02944758415222168, 0.029235231399536134, 0.0295850887298584, 0.029579967498779298, 0.029421695709228514, 0.02922604751586914, 0.029809663772583008, 0.02953753662109375, 0.029340415954589843, 0.03168047904968262, 0.030681119918823243, 0.029556991577148438, 0.029794048309326172, 0.029875776290893555, 0.029831743240356444, 0.029855743408203125, 0.029958015441894532, 0.03005580711364746, 0.030210687637329103, 0.03012784004211426, 0.029782079696655275, 0.029851615905761717, 0.030861312866210938, 0.029816768646240235, 0.02987740707397461, 0.03382160186767578, 0.033925121307373046, 0.029743104934692382, 0.029755392074584962, 0.029746240615844726, 0.029781951904296874, 0.029831167221069335, 0.02998374366760254, 0.02967065620422363, 0.029706720352172852, 0.029290752410888674, 0.029259807586669923, 0.02922710418701172, 0.029548095703125, 0.029360479354858398, 0.029542272567749023, 0.029122655868530273, 0.029177888870239258, 0.029162687301635744, 0.029078336715698243, 0.029263744354248045, 0.029292671203613282, 0.029226112365722656, 0.02975993537902832, 0.029712831497192383, 0.029665279388427734, 0.029812736511230467, 0.02964873504638672, 0.029595808029174806, 0.029644351959228515, 0.029726272583007814, 0.02995859146118164, 0.02978665542602539, 0.030074783325195312, 0.029724672317504884, 0.029831167221069335, 0.029642751693725586, 0.029386751174926756, 0.029454336166381836, 0.02957049560546875, 0.029456384658813478, 0.029440479278564455, 0.02940889549255371, 0.029466400146484373, 0.02923766326904297, 0.029240800857543946, 0.029278207778930664, 0.029205312728881837, 0.029437952041625977, 0.029310976028442383, 0.02959564781188965, 0.029752607345581054, 0.029512416839599608, 0.02914739227294922, 0.02927292823791504, 0.029311679840087892, 0.029155551910400392, 0.02927552032470703, 0.0299400634765625, 0.029259136199951172, 0.0313474235534668, 0.029718624114990235, 0.030408096313476563, 0.03231999969482422, 0.029546560287475584, 0.029366207122802735, 0.029491199493408202, 0.029290496826171877, 0.029314304351806642, 0.02959791946411133, 0.029464544296264647, 0.029254016876220704, 0.029106367111206056, 0.029157152175903322, 0.029077728271484374, 0.029230943679809572, 0.029140640258789062, 0.02937468719482422, 0.0292825927734375, 0.029272064208984375, 0.02907151985168457, 0.029235040664672852, 0.029120512008666992, 0.02914303970336914, 0.029101728439331054, 0.029091968536376953, 0.029325536727905274, 0.02916147232055664, 0.029181024551391602, 0.029111200332641602, 0.029061119079589845, 0.029083648681640626, 0.029079551696777343, 0.029233152389526368, 0.029296640396118165, 0.029120512008666992, 0.029149183273315428, 0.02992742347717285, 0.02924937629699707, 0.029204639434814453, 0.029124607086181642, 0.029081600189208984, 0.029159648895263672, 0.029228832244873045, 0.02938470458984375, 0.029145280838012697, 0.029417280197143555, 0.029058528900146485, 0.029055519104003905, 0.029040767669677735, 0.029191808700561525, 0.029296287536621092, 0.029418079376220704, 0.02942300796508789, 0.029239904403686522, 0.02923654365539551, 0.029213375091552734, 0.02913484764099121, 0.029239295959472656, 0.02910428810119629, 0.029220703125, 0.02917580795288086, 0.029256799697875976, 0.030508991241455077, 0.029890560150146486, 0.03144710350036621, 0.029822912216186524, 0.03004323196411133, 0.029637535095214843, 0.029461824417114257, 0.02947961616516113, 0.02975948715209961, 0.029648319244384765, 0.029802400588989256, 0.029738880157470702, 0.0294420166015625, 0.029440832138061524, 0.029321216583251954, 0.02920038414001465, 0.02934307289123535, 0.029227680206298828, 0.02910207939147949, 0.02920857620239258, 0.029257055282592773, 0.029462879180908202, 0.029767679214477538, 0.029349536895751954, 0.02918467140197754, 0.029206079483032228, 0.029251935958862305, 0.029085792541503907, 0.029376447677612303, 0.02918111991882324, 0.02897599983215332, 0.02921228790283203, 0.02909555244445801, 0.029125375747680662, 0.029335744857788087, 0.029679487228393555, 0.029890495300292967, 0.029439807891845703, 0.029317312240600586, 0.02906675148010254, 0.029107776641845703, 0.02909417533874512, 0.029211296081542968, 0.02945155143737793, 0.029543136596679686, 0.029425664901733397, 0.029519872665405275, 0.029386751174926756, 0.029467647552490234, 0.029258752822875978, 0.02911846351623535, 0.02921062469482422, 0.029050527572631837, 0.02937059211730957, 0.029331167221069335, 0.029296863555908204, 0.029095455169677733, 0.029260448455810547, 0.029219039916992186, 0.029441343307495118, 0.029493728637695314, 0.030680192947387695, 0.02986422348022461, 0.03139433670043945, 0.03000115203857422, 0.029487104415893556, 0.02940108871459961, 0.029429759979248047, 0.029180927276611326, 0.029434879302978514, 0.029631744384765624, 0.029811296463012695, 0.029596832275390624, 0.0291727352142334, 0.02931616020202637, 0.02926643180847168, 0.029243072509765624, 0.028989408493041994, 0.029124448776245117, 0.029125408172607423, 0.0292325439453125, 0.02928102493286133, 0.02931711959838867, 0.029190143585205077, 0.02923423957824707, 0.029182912826538086, 0.029258975982666014, 0.029413375854492187, 0.029338367462158205, 0.0295199031829834, 0.029302047729492187, 0.029332191467285156, 0.029617792129516603, 0.029155168533325195, 0.030354080200195314, 0.030721920013427734, 0.02936422348022461, 0.02943935966491699, 0.029634719848632814, 0.029512159347534178, 0.029734912872314452, 0.029550592422485353, 0.030027776718139648, 0.029367807388305665, 0.02947532844543457, 0.029379903793334963, 0.029911167144775392, 0.029476896286010742, 0.029106719970703125, 0.029527616500854493, 0.029325056076049804, 0.029465280532836913, 0.029214719772338867, 0.029472768783569334, 0.029382623672485352, 0.029462560653686524, 0.029265504837036133, 0.02923356819152832, 0.029261823654174804, 0.02917523193359375, 0.02937913513183594, 0.029312511444091797, 0.02945884895324707, 0.029208671569824218, 0.02937446403503418, 0.029345279693603517, 0.030140735626220702, 0.029562944412231444, 0.02918400001525879, 0.02921062469482422, 0.02937651252746582, 0.02917580795288086, 0.02924310493469238, 0.029196575164794923, 0.02924367904663086, 0.02931273651123047, 0.02926972770690918, 0.03281094360351562, 0.029489343643188476, 0.02947088050842285, 0.029130752563476563, 0.029169599533081056, 0.029124671936035157, 0.029360128402709962, 0.029138944625854493, 0.02921062469482422, 0.029073087692260743, 0.02959312057495117, 0.0290548152923584, 0.029123199462890624, 0.029110784530639647, 0.029187551498413088, 0.029196640014648438, 0.029157215118408204, 0.02928451156616211, 0.02924060821533203, 0.029567712783813475, 0.02953625679016113, 0.029474815368652343, 0.029303968429565428, 0.02926380729675293, 0.029600383758544922, 0.029255519866943358, 0.029292991638183594, 0.02934169578552246, 0.029306047439575194, 0.02927084732055664, 0.030701759338378907, 0.02956211280822754, 0.02931273651123047, 0.02908678436279297, 0.02934752082824707, 0.029497215270996094, 0.029335775375366212, 0.02939846420288086, 0.029215072631835936, 0.029343135833740236, 0.029209407806396484, 0.029393024444580078, 0.029175680160522462, 0.029081600189208984, 0.02959676742553711, 0.02916035270690918, 0.029263872146606446, 0.029212671279907225, 0.029357952117919923, 0.02952739143371582, 0.030978719711303712, 0.029614112854003907, 0.030810111999511718, 0.029646879196166993, 0.029475936889648436, 0.029499488830566405, 0.029602783203125, 0.029327167510986327, 0.029335552215576172, 0.02934163284301758, 0.029548608779907226, 0.029647008895874023, 0.029529951095581056, 0.02958131217956543, 0.029472768783569334, 0.03018057632446289, 0.029399168014526366, 0.029589471817016603, 0.02950147247314453, 0.029395263671875, 0.029284255981445313, 0.02960223960876465, 0.029488895416259767, 0.029419776916503906, 0.02928179168701172, 0.02962892723083496, 0.030008960723876953, 0.03782486343383789, 0.02988035202026367, 0.029473056793212892, 0.02937740707397461, 0.029492063522338866, 0.030115840911865234, 0.02937446403503418, 0.029233152389526368, 0.029384607315063475, 0.029388160705566407, 0.029301471710205078, 0.02936832046508789, 0.02936729621887207, 0.029432512283325194, 0.030513343811035157, 0.030652799606323243, 0.030134016036987305, 0.030004480361938476, 0.030009376525878907, 0.029935808181762696, 0.029714975357055664, 0.03034636878967285, 0.030593536376953126, 0.02963007926940918, 0.029301504135131835, 0.029403200149536134, 0.029374176025390625, 0.029468896865844727, 0.029474815368652343, 0.02931711959838867, 0.02931292724609375, 0.029266016006469726, 0.029366111755371092, 0.029606048583984374, 0.02931711959838867, 0.029130752563476563, 0.030665023803710938, 0.029875904083251952]",tokens/s,33.8366104462109,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1501.093888,1324.285952,0.0,945.815552,943.480832,s,1,8.4035087890625,8.4035087890625,0.0,8.4035087890625,8.4035087890625,8.4035087890625,8.4035087890625,[8.4035087890625],,kWh,4.1327311754164e-05,4.550933413764803e-06,1.3277510622005523e-05,5.9155755789934325e-05,,MB,1589.542912,1487.863808,0.0,1080.03328,1046.51776,s,10,0.8268167724609374,0.08268167724609374,0.0004368021488074932,0.08262483215332031,0.08331148834228516,0.08346441802978516,0.08358676177978516,"[0.08361734771728516, 0.08257561492919922, 0.0824674530029297, 0.08216486358642579, 0.08275836944580078, 0.08218051147460938, 0.08327750396728516, 0.08237484741210938, 0.08272621154785156, 0.08267404937744141]",tokens/s,3096.211984646146,kWh,2.5114694163461434e-06,2.768441387566843e-07,1.6708014316068285e-06,4.459114986709656e-06,tokens/kWh,57410495.30299291,MB,1593.438208,1573.84704,0.0,1166.016512,1082.822656,s,10,11.529435424804689,1.1529435424804686,0.002247484354686342,1.1528699340820312,1.155279248046875,1.1558662475585937,1.1563358471679688,"[1.152810791015625, 1.15289111328125, 1.1527913818359374, 1.1551488037109374, 1.1535712890625, 1.1528487548828126, 1.1502054443359375, 1.1481671142578125, 1.1545474853515625, 1.1564532470703126]",tokens/s,54.64274500767001,kWh,3.383504718865373e-05,3.731713839729946e-06,1.7744325211593323e-05,5.531108623997701e-05,tokens/kWh,1139012.16343254,,s,630,11.527169504165656,0.018297094451056584,0.0003323471075669016,0.018204559326171876,0.01853084735870361,0.018927657604217527,0.01959991451263428,"[0.018458911895751953, 0.018184671401977538, 0.01805721664428711, 0.018102272033691406, 0.01833776092529297, 0.01850774383544922, 0.01816172790527344, 0.018067455291748045, 0.01827014350891113, 0.01813100814819336, 0.018089984893798827, 0.01811359977722168, 0.018141984939575195, 0.018163007736206056, 0.01816643142700195, 0.018217151641845702, 0.01816582489013672, 0.01832748794555664, 0.018210111618041994, 0.018090368270874024, 0.018047103881835936, 0.01814134407043457, 0.018185375213623046, 0.018227615356445313, 0.018135520935058595, 0.018077695846557617, 0.018153663635253905, 0.0181463680267334, 0.01811327934265137, 0.018124479293823242, 0.018143552780151367, 0.018131231307983397, 0.018112224578857423, 0.019215391159057616, 0.018207712173461912, 0.018198528289794923, 0.01816783905029297, 0.01830499267578125, 0.018237760543823242, 0.018340736389160155, 0.01835647964477539, 0.018370784759521485, 0.018574975967407228, 0.018284351348876952, 0.018283424377441407, 0.018289920806884765, 0.018270879745483398, 0.018410751342773438, 0.018221023559570313, 0.01821526336669922, 0.02178102493286133, 0.01844428825378418, 0.01826188850402832, 0.018288768768310548, 0.018212480545043944, 0.018347391128540037, 0.018267135620117187, 0.01818601608276367, 0.01817545509338379, 0.018150144577026368, 0.01830694389343262, 0.018187936782836915, 0.018325023651123047, 0.018329919815063475, 0.018243263244628907, 0.018178400039672853, 0.018134016036987305, 0.018115360260009764, 0.018322784423828124, 0.01823593521118164, 0.018228832244873046, 0.01825846481323242, 0.018108415603637695, 0.01819196891784668, 0.018594207763671874, 0.018593791961669923, 0.018307071685791015, 0.01830473518371582, 0.01831760025024414, 0.01831248092651367, 0.018244352340698242, 0.018257919311523436, 0.01826812744140625, 0.019224576950073242, 0.018817024230957033, 0.018292383193969728, 0.01817795181274414, 0.01817612838745117, 0.01817964744567871, 0.018199296951293947, 0.01824563217163086, 0.018126848220825196, 0.02006380844116211, 0.018229183197021485, 0.018243392944335936, 0.018264768600463867, 0.01822105598449707, 0.01829033660888672, 0.018224672317504884, 0.018184959411621095, 0.01829449653625488, 0.018214975357055664, 0.018216928482055663, 0.018229440689086916, 0.01820419120788574, 0.018227807998657225, 0.01825984001159668, 0.018193727493286134, 0.018231840133666993, 0.018173280715942382, 0.018167871475219727, 0.018250207901000976, 0.01849734306335449, 0.018170175552368165, 0.018157855987548828, 0.018214176177978516, 0.01827840042114258, 0.01819926452636719, 0.01826201629638672, 0.018142784118652344, 0.018322879791259766, 0.018147520065307617, 0.01825027275085449, 0.01816316795349121, 0.018301759719848633, 0.018184255599975586, 0.018536447525024414, 0.020339744567871094, 0.018977760314941406, 0.018364639282226564, 0.01815456008911133, 0.018121152877807616, 0.018124191284179688, 0.019116928100585937, 0.018251775741577148, 0.018329599380493163, 0.01815283203125, 0.01810291290283203, 0.01816329574584961, 0.0181231689453125, 0.01819366455078125, 0.018489599227905273, 0.018395647048950196, 0.018157888412475585, 0.018266048431396485, 0.018083583831787108, 0.018453727722167967, 0.01830108833312988, 0.018354816436767576, 0.018195711135864257, 0.01816204833984375, 0.01823196792602539, 0.018122463226318358, 0.018199552536010744, 0.01808896064758301, 0.018122047424316407, 0.01813164710998535, 0.018241439819335938, 0.01849558448791504, 0.01815894317626953, 0.018057504653930665, 0.018123136520385743, 0.018269664764404298, 0.018210336685180663, 0.01810527992248535, 0.018109760284423827, 0.01804569625854492, 0.018170080184936523, 0.018112287521362305, 0.018077695846557617, 0.018218751907348632, 0.018163839340209962, 0.018945600509643553, 0.01824211120605469, 0.018207935333251952, 0.01816156768798828, 0.018216096878051757, 0.01815936088562012, 0.01819443130493164, 0.01834320068359375, 0.018526815414428712, 0.018234752655029298, 0.018184831619262695, 0.01828611183166504, 0.018092639923095705, 0.018224544525146484, 0.018080352783203125, 0.018097280502319336, 0.01900364875793457, 0.019517183303833008, 0.01814121627807617, 0.01832364845275879, 0.018444608688354493, 0.018110944747924806, 0.018131967544555663, 0.01816268730163574, 0.018298879623413086, 0.018296735763549805, 0.019089120864868164, 0.018239519119262696, 0.01820707130432129, 0.018239488601684572, 0.01840127944946289, 0.018167327880859375, 0.018141664505004883, 0.01804035186767578, 0.018601791381835937, 0.018159456253051758, 0.018228031158447264, 0.018255872726440428, 0.018140832901000978, 0.01810047912597656, 0.01815510368347168, 0.018727424621582032, 0.019129440307617186, 0.01823187255859375, 0.01822550392150879, 0.01834614372253418, 0.018201696395874024, 0.01804569625854492, 0.018069503784179687, 0.018096128463745118, 0.018128896713256838, 0.018136735916137695, 0.01819683265686035, 0.018108415603637695, 0.018143232345581056, 0.018130367279052734, 0.018119232177734375, 0.018134016036987305, 0.018107391357421874, 0.018136415481567383, 0.018121376037597656, 0.018134271621704102, 0.018194944381713866, 0.01815577507019043, 0.018099903106689453, 0.0180513916015625, 0.018144895553588867, 0.019227008819580078, 0.018339935302734374, 0.018433055877685546, 0.018467008590698244, 0.01865388870239258, 0.019785375595092775, 0.018686304092407225, 0.018758655548095703, 0.018766847610473633, 0.018288639068603514, 0.01821286392211914, 0.01822550392150879, 0.018158464431762694, 0.018840576171875, 0.01864838409423828, 0.01853068733215332, 0.018528575897216796, 0.01886115264892578, 0.018275232315063478, 0.018259967803955078, 0.01819443130493164, 0.01827833557128906, 0.018384960174560545, 0.01844380760192871, 0.018581535339355467, 0.018335391998291015, 0.018287391662597657, 0.018315168380737306, 0.01824127960205078, 0.018141408920288087, 0.01818796730041504, 0.01819411277770996, 0.018368799209594725, 0.018131423950195312, 0.01817190361022949, 0.018080896377563476, 0.018100576400756838, 0.01813763236999512, 0.018239488601684572, 0.018214912414550782, 0.018884607315063476, 0.019231935501098633, 0.01859231948852539, 0.01823468780517578, 0.018214879989624025, 0.01808403205871582, 0.018115360260009764, 0.018100223541259765, 0.01804697608947754, 0.018155519485473632, 0.019152576446533204, 0.018188831329345703, 0.018183967590332032, 0.01823289680480957, 0.018118751525878905, 0.018186368942260743, 0.01818047904968262, 0.01818828773498535, 0.018452320098876953, 0.01818137550354004, 0.018196672439575196, 0.01815545654296875, 0.018137727737426758, 0.01817964744567871, 0.01826246452331543, 0.018233343124389647, 0.018192384719848635, 0.018446176528930665, 0.01822934341430664, 0.018151487350463867, 0.018255807876586913, 0.018133056640625, 0.018435903549194336, 0.018211008071899414, 0.01813484764099121, 0.018280832290649415, 0.018132640838623048, 0.01813360023498535, 0.018048799514770508, 0.018155744552612305, 0.018067455291748045, 0.018102144241333006, 0.018178176879882813, 0.01812623977661133, 0.018176607131958008, 0.018135072708129883, 0.018182111740112306, 0.018146623611450197, 0.01815622329711914, 0.018094079971313477, 0.018259904861450196, 0.018199712753295898, 0.01820150375366211, 0.018167808532714845, 0.018089984893798827, 0.01809436798095703, 0.018036447525024413, 0.018067455291748045, 0.018010271072387694, 0.018120672225952147, 0.019124095916748046, 0.01838447952270508, 0.018578880310058592, 0.018086431503295898, 0.01808633613586426, 0.01815465545654297, 0.018120735168457032, 0.0181092472076416, 0.018100223541259765, 0.019085119247436524, 0.018468416213989258, 0.01957142448425293, 0.018171808242797852, 0.01821273612976074, 0.018339712142944335, 0.018204927444458008, 0.018092031478881835, 0.018184192657470705, 0.01828156852722168, 0.018115583419799804, 0.018116512298583985, 0.018198751449584962, 0.01919977569580078, 0.01832364845275879, 0.01815737533569336, 0.018137088775634767, 0.01808950424194336, 0.01817558479309082, 0.01807244873046875, 0.018186176300048828, 0.018296895980834962, 0.01815670394897461, 0.018111328125, 0.01943142318725586, 0.020368671417236327, 0.01828118324279785, 0.018219104766845705, 0.01828223991394043, 0.018253984451293944, 0.018263551712036134, 0.01809049606323242, 0.018333696365356447, 0.01807910346984863, 0.01809062385559082, 0.01818623924255371, 0.019267839431762697, 0.018130687713623046, 0.018045087814331055, 0.018081632614135743, 0.018044927597045898, 0.01912556838989258, 0.01822358322143555, 0.018090208053588866, 0.01803500747680664, 0.01812233543395996, 0.018038240432739258, 0.018216991424560548, 0.01812441635131836, 0.018094303131103516, 0.0180948486328125, 0.01819443130493164, 0.018335744857788085, 0.018184192657470705, 0.018101343154907225, 0.018055999755859375, 0.018136159896850586, 0.018117631912231445, 0.01807360076904297, 0.018276639938354492, 0.0182554874420166, 0.019285472869873047, 0.01890572738647461, 0.01878201675415039, 0.018249120712280274, 0.018245439529418945, 0.01823632049560547, 0.01809324836730957, 0.01811974334716797, 0.01807459259033203, 0.01840640068054199, 0.018245695114135742, 0.018132511138916015, 0.018178207397460937, 0.018110559463500975, 0.018280448913574218, 0.018081792831420897, 0.018083423614501954, 0.018102399826049803, 0.018208608627319336, 0.01841606330871582, 0.018617631912231446, 0.018381471633911132, 0.018239551544189454, 0.018135135650634765, 0.01818943977355957, 0.018121503829956056, 0.018358591079711915, 0.018089567184448242, 0.01803856086730957, 0.018123071670532228, 0.018040416717529296, 0.018899360656738282, 0.01822163200378418, 0.018304832458496095, 0.018153919219970702, 0.018112159729003905, 0.018280799865722657, 0.018151552200317382, 0.018143104553222655, 0.01839308738708496, 0.01856060791015625, 0.018280864715576172, 0.018225151062011717, 0.018229248046875, 0.01821900749206543, 0.0181628475189209, 0.018346111297607423, 0.0181624641418457, 0.01832748794555664, 0.018137216567993164, 0.018130815505981446, 0.018118656158447266, 0.01817804718017578, 0.018167808532714845, 0.01818556785583496, 0.018120672225952147, 0.01816441535949707, 0.018224672317504884, 0.018209247589111327, 0.01828160095214844, 0.018090879440307617, 0.018147327423095702, 0.01822719955444336, 0.01818780708312988, 0.0181560001373291, 0.018135232925415037, 0.018153280258178712, 0.018140607833862305, 0.01806015968322754, 0.018144927978515624, 0.018175296783447266, 0.018117055892944336, 0.018047264099121094, 0.018100223541259765, 0.01809110450744629, 0.018254751205444335, 0.01873459243774414, 0.018355743408203125, 0.018221183776855467, 0.018170175552368165, 0.01812944030761719, 0.018137088775634767, 0.018170015335083008, 0.018192256927490234, 0.01827734375, 0.018215072631835937, 0.01917411231994629, 0.01824985694885254, 0.018208671569824218, 0.018268255233764647, 0.018097824096679687, 0.01819100761413574, 0.018132671356201172, 0.018138816833496094, 0.018460800170898437, 0.018995071411132814, 0.018319168090820313, 0.01818864059448242, 0.018505727767944336, 0.018464767456054687, 0.01840447998046875, 0.018367359161376953, 0.018149215698242186, 0.018110496520996094, 0.018118783950805663, 0.01814691162109375, 0.018225568771362305, 0.018134559631347656, 0.01815395164489746, 0.01825382423400879, 0.018159616470336915, 0.018128896713256838, 0.01814240074157715, 0.01810256004333496, 0.018141727447509765, 0.01818979263305664, 0.018147296905517578, 0.018289087295532226, 0.01817344093322754, 0.018137727737426758, 0.01816985511779785, 0.0184965763092041, 0.018289600372314453, 0.01828985595703125, 0.01826464080810547, 0.018267488479614256, 0.018408191680908202, 0.01818003273010254, 0.018272192001342773, 0.018140512466430662, 0.01811520004272461, 0.018172224044799803, 0.01818320083618164, 0.01808483123779297, 0.01821696090698242, 0.018051071166992186, 0.019091455459594727, 0.018181535720825197, 0.018107135772705077, 0.018171743392944337, 0.018146400451660157, 0.01813596725463867, 0.01840742492675781, 0.01837494468688965, 0.018650848388671874, 0.018485023498535157, 0.018392416000366212, 0.018357152938842772, 0.018194400787353515, 0.018093984603881837, 0.018399391174316406, 0.018079776763916016, 0.018304607391357423, 0.020652351379394532, 0.01961155128479004, 0.018309215545654296, 0.018264064788818358, 0.01815331268310547, 0.018258848190307618, 0.01819443130493164, 0.01843177604675293, 0.018120927810668944, 0.018182144165039063, 0.018152448654174806, 0.01815449523925781, 0.01806780815124512, 0.01811814308166504, 0.01858531188964844, 0.01825584030151367, 0.018252191543579103, 0.018177536010742186, 0.018208480834960936, 0.018352991104125978, 0.01829478454589844, 0.01830297660827637, 0.018431999206542968, 0.01918156814575195, 0.01852544021606445, 0.018491264343261718, 0.018478111267089845, 0.01845631980895996, 0.018456640243530272, 0.018409503936767577, 0.018526208877563476, 0.01858905601501465, 0.01851158332824707, 0.019485631942749024, 0.01853228759765625, 0.01837264060974121, 0.018249408721923828, 0.018217632293701172, 0.018124448776245118, 0.01826959991455078, 0.01834377670288086, 0.018706687927246092, 0.01859382438659668, 0.018280576705932618, 0.01829859161376953, 0.018192127227783204, 0.01827724838256836, 0.018171743392944337, 0.018148704528808592, 0.018279232025146485, 0.018366111755371093, 0.018315616607666015, 0.018210464477539063, 0.018288991928100587, 0.018193599700927734, 0.018225120544433593, 0.018207584381103516, 0.01837648010253906, 0.018501792907714844, 0.018472991943359374, 0.018415647506713866, 0.018571264266967775, 0.018308576583862306, 0.018246431350708008, 0.01821980857849121, 0.01815648078918457, 0.018156608581542968, 0.018278656005859376]",tokens/s,54.653486250231055,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 205, in run_text_generation_memory_tracking _ = backend.generate(self.inputs, self.config.generate_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 430, in generate return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 280, in forward attn_weights = nn.functional.softmax(attn_weights, dim=-1, dtype=torch.float32).to(query_states.dtype) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/functional.py"", line 1887, in softmax ret = input.softmax(dim, dtype=dtype) RuntimeError: CUDA error: an illegal memory access was encountered CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. For debugging consider passing CUDA_LAUNCH_BLOCKING=1. Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1041.743872,867.106816,0.0,488.636416,482.553856,s,1,7.8583193359375,7.8583193359375,0.0,7.8583193359375,7.8583193359375,7.8583193359375,7.8583193359375,[7.8583193359375],,kWh,2.8190545075002166e-05,3.1021099707537287e-06,9.49389648402521e-06,4.078655152978111e-05,,MB,1243.959296,1024.393216,0.0,616.562688,581.925888,s,10,0.33909443283081053,0.033909443283081056,0.00020581722807567733,0.033835599899291996,0.03415680732727051,0.03424669094085693,0.034318597831726076,"[0.03413683319091797, 0.03406169509887695, 0.03382361602783203, 0.0337279052734375, 0.034006336212158206, 0.033847583770751956, 0.033692222595214844, 0.03433657455444336, 0.03370121765136719, 0.03376044845581055]",tokens/s,7549.519402688923,kWh,1.0397370668743825e-06,1.1466459175257054e-07,6.882419503558293e-07,1.8426436089827824e-06,tokens/kWh,138930826.74914163,MB,1278.291968,1039.07328,0.0,631.242752,597.192192,s,10,14.378984497070313,1.4378984497070313,0.010146728865944542,1.4366253051757814,1.4495190185546873,1.4552125732421874,1.4597674169921875,"[1.4361473388671875, 1.4241981201171876, 1.4270159912109375, 1.43397021484375, 1.4609061279296875, 1.4414090576171874, 1.4482537841796874, 1.437103271484375, 1.4393765869140625, 1.43060400390625]",tokens/s,43.81394250257111,kWh,4.168290228937644e-05,4.597195103610237e-06,1.646704955164374e-05,6.27471469446304e-05,tokens/kWh,1004029.7139819396,,s,630,14.373176967620855,0.02281456661527119,0.00047591966537088664,0.02269748783111572,0.023278319358825687,0.023458812618255612,0.024824110603332533,"[0.02253209686279297, 0.022571136474609375, 0.022504703521728515, 0.022395135879516602, 0.022505216598510742, 0.022462560653686525, 0.022640960693359375, 0.022789791107177736, 0.02298633575439453, 0.022700992584228516, 0.022674688339233397, 0.02308790397644043, 0.023357440948486328, 0.02303385543823242, 0.023007232666015624, 0.02279648017883301, 0.022486751556396484, 0.02263868713378906, 0.022525888442993164, 0.022524192810058595, 0.022453023910522462, 0.02255558395385742, 0.022415008544921875, 0.02253660774230957, 0.022458368301391602, 0.022411136627197265, 0.02248716735839844, 0.022421503067016603, 0.022444032669067384, 0.02247865676879883, 0.02266111946105957, 0.02279020881652832, 0.02285126495361328, 0.02298854446411133, 0.023042144775390624, 0.023126016616821288, 0.023278175354003908, 0.022971935272216797, 0.023128543853759766, 0.023219392776489257, 0.023202367782592773, 0.023128383636474608, 0.023088703155517577, 0.022982463836669922, 0.02269254493713379, 0.022658111572265625, 0.02259836769104004, 0.02254364776611328, 0.02278892707824707, 0.02273904037475586, 0.022889663696289062, 0.022905088424682616, 0.023042272567749024, 0.023061887741088867, 0.022924160003662108, 0.023072063446044924, 0.022760128021240233, 0.022607744216918944, 0.022668800354003905, 0.022481695175170898, 0.022887775421142578, 0.023280384063720704, 0.023572223663330078, 0.02255900764465332, 0.022659679412841797, 0.022520959854125975, 0.02251046371459961, 0.022785728454589843, 0.02328607940673828, 0.022579200744628908, 0.022503776550292967, 0.02247977638244629, 0.022552799224853516, 0.02245644760131836, 0.02234204864501953, 0.022548479080200197, 0.022411264419555665, 0.02253209686279297, 0.022506624221801757, 0.022459264755249023, 0.02253968048095703, 0.022581855773925782, 0.022728544235229492, 0.022872032165527342, 0.02268899154663086, 0.022406112670898436, 0.022495231628417968, 0.022370271682739258, 0.022411296844482422, 0.022312959671020507, 0.022366207122802736, 0.022466976165771483, 0.02246348762512207, 0.022503103256225586, 0.022456735610961915, 0.02253059196472168, 0.022494367599487305, 0.022452224731445314, 0.022745920181274415, 0.022886688232421876, 0.023256799697875977, 0.0229171199798584, 0.022747135162353514, 0.0226296329498291, 0.02270899200439453, 0.02254643249511719, 0.02250268745422363, 0.022528064727783202, 0.022614688873291017, 0.022552576065063477, 0.022410816192626953, 0.022423999786376953, 0.022461631774902343, 0.022417312622070314, 0.022436800003051757, 0.02232111930847168, 0.022386688232421875, 0.022853631973266602, 0.022730688095092773, 0.02296022415161133, 0.02296214485168457, 0.023000736236572266, 0.022878559112548828, 0.022832223892211914, 0.022645599365234376, 0.022503711700439452, 0.022335487365722655, 0.022606912612915038, 0.02259552001953125, 0.02268083190917969, 0.022586111068725587, 0.022429792404174805, 0.022559648513793946, 0.022479040145874023, 0.02260767936706543, 0.02240716743469238, 0.02253343963623047, 0.022441791534423827, 0.022532991409301758, 0.022576160430908203, 0.022704799652099608, 0.022663679122924805, 0.022820671081542968, 0.022769792556762695, 0.022677312850952147, 0.022723648071289064, 0.022590688705444336, 0.024086591720581054, 0.022619359970092772, 0.022620672225952147, 0.022511199951171876, 0.02318377685546875, 0.022728672027587892, 0.022558656692504883, 0.022487136840820314, 0.022525951385498046, 0.022747360229492187, 0.022470048904418945, 0.022394559860229493, 0.022520095825195312, 0.022554975509643554, 0.02256697654724121, 0.022466560363769532, 0.02249113655090332, 0.022571008682250978, 0.022634496688842775, 0.02251692771911621, 0.022463296890258787, 0.02256220817565918, 0.023134687423706054, 0.022909055709838866, 0.022804479598999023, 0.022593856811523438, 0.02264031982421875, 0.022521184921264648, 0.02250819206237793, 0.022814495086669922, 0.022807872772216797, 0.02303580856323242, 0.022659328460693358, 0.02256358337402344, 0.022508768081665038, 0.022494144439697265, 0.022667007446289064, 0.022421600341796875, 0.022708223342895507, 0.022616064071655274, 0.022849248886108398, 0.022632736206054688, 0.022496992111206055, 0.022515552520751953, 0.02247657585144043, 0.022542303085327148, 0.0225020809173584, 0.02267068862915039, 0.022643360137939453, 0.02263654327392578, 0.02271379280090332, 0.0227128963470459, 0.022867231369018554, 0.02264860725402832, 0.022884416580200195, 0.02268864059448242, 0.022787872314453124, 0.022675872802734375, 0.022648639678955078, 0.022543392181396484, 0.02258367919921875, 0.02258515167236328, 0.022735616683959962, 0.022603519439697267, 0.023009056091308593, 0.02338991928100586, 0.02309622383117676, 0.02267862319946289, 0.022614112854003908, 0.022547136306762694, 0.022470016479492188, 0.02261030387878418, 0.022982912063598634, 0.022542335510253905, 0.022623584747314452, 0.022589567184448243, 0.022545087814331056, 0.02253603172302246, 0.022687871932983397, 0.022665088653564452, 0.02287820816040039, 0.02266111946105957, 0.022668800354003905, 0.02304025650024414, 0.022907232284545897, 0.02307708740234375, 0.023418624877929686, 0.02373801612854004, 0.022910335540771484, 0.02279091262817383, 0.02261417579650879, 0.022849536895751952, 0.022675455093383787, 0.022697984695434572, 0.022589439392089843, 0.02255801582336426, 0.022560543060302734, 0.022424480438232423, 0.022425792694091798, 0.02343916893005371, 0.022912384033203125, 0.022742752075195313, 0.02295814323425293, 0.023075679779052734, 0.022988800048828126, 0.022623359680175783, 0.02258390426635742, 0.02265033531188965, 0.022457151412963866, 0.02245827293395996, 0.022437343597412108, 0.022360191345214844, 0.02253379249572754, 0.02262915229797363, 0.02294790458679199, 0.027543552398681642, 0.027435327529907228, 0.02443846321105957, 0.023088672637939452, 0.022986591339111326, 0.02313075256347656, 0.023051488876342775, 0.02341894340515137, 0.023089887619018555, 0.023178272247314453, 0.023107967376708984, 0.022948448181152343, 0.022916255950927736, 0.022879072189331055, 0.02300022315979004, 0.023018560409545898, 0.023120800018310548, 0.02314249610900879, 0.02289039993286133, 0.022956928253173827, 0.02284339141845703, 0.02299014472961426, 0.022833280563354492, 0.022747200012207033, 0.022764127731323244, 0.02280790328979492, 0.022901311874389648, 0.022923263549804687, 0.02339798355102539, 0.02314179229736328, 0.02330316734313965, 0.023355392456054686, 0.023783647537231445, 0.023586591720581054, 0.023373823165893554, 0.02375657653808594, 0.0235317440032959, 0.023475648880004883, 0.024932928085327148, 0.02328166389465332, 0.023473312377929687, 0.023229087829589844, 0.023314624786376952, 0.02296188735961914, 0.0229399356842041, 0.022746976852416993, 0.022724927902221678, 0.022839136123657226, 0.02284707260131836, 0.02270047950744629, 0.022810592651367187, 0.022480480194091795, 0.022573471069335938, 0.022275007247924805, 0.022550207138061523, 0.023072223663330078, 0.022670175552368162, 0.02262118339538574, 0.022590463638305663, 0.022828159332275392, 0.022921472549438476, 0.022975072860717774, 0.022733055114746093, 0.02256780815124512, 0.0224736328125, 0.022700159072875977, 0.02238217544555664, 0.022325023651123047, 0.022366079330444336, 0.02232588768005371, 0.02250752067565918, 0.02231283187866211, 0.022482208251953125, 0.022401887893676756, 0.022553695678710937, 0.022485023498535157, 0.022449024200439455, 0.022762847900390626, 0.022978656768798827, 0.02325356864929199, 0.023191871643066405, 0.023211103439331054, 0.023171680450439453, 0.02324684715270996, 0.023142240524291993, 0.023004703521728516, 0.023118463516235352, 0.022982175827026368, 0.02270252799987793, 0.022683679580688478, 0.022787328720092773, 0.02303164863586426, 0.02325187110900879, 0.02328153610229492, 0.023283615112304687, 0.02316444778442383, 0.02311974334716797, 0.023156991958618166, 0.02307164764404297, 0.023154144287109376, 0.023031999588012695, 0.02313612747192383, 0.023058528900146483, 0.022868000030517577, 0.022728607177734374, 0.02264192008972168, 0.02271126365661621, 0.022875776290893556, 0.022951295852661133, 0.022604671478271485, 0.02263859176635742, 0.022697023391723633, 0.022683744430541993, 0.02311404800415039, 0.02517251205444336, 0.023605247497558594, 0.022695903778076173, 0.022631839752197267, 0.022593151092529296, 0.024963455200195314, 0.024557695388793946, 0.023130815505981447, 0.02283091163635254, 0.022644704818725585, 0.022396799087524413, 0.02256912040710449, 0.02250060844421387, 0.022430463790893553, 0.022656063079833984, 0.02263520050048828, 0.022655231475830078, 0.022896223068237305, 0.025541023254394533, 0.023058496475219726, 0.022882144927978517, 0.023152767181396486, 0.02269795227050781, 0.022664512634277344, 0.02299942398071289, 0.022602048873901368, 0.022550527572631835, 0.02353971290588379, 0.022547967910766603, 0.02260633659362793, 0.022477951049804688, 0.022475648880004882, 0.022347776412963868, 0.022452224731445314, 0.022355968475341798, 0.022343679428100584, 0.022386304855346678, 0.02242188835144043, 0.022546016693115234, 0.02253455924987793, 0.02263859176635742, 0.022800384521484376, 0.022819936752319334, 0.02270262336730957, 0.022761856079101563, 0.022674911499023436, 0.022979103088378906, 0.022734848022460938, 0.02271433639526367, 0.023029792785644532, 0.023012928009033203, 0.023357759475708006, 0.02343270492553711, 0.023718528747558594, 0.023398399353027344, 0.023396352767944335, 0.023318111419677736, 0.02335545539855957, 0.023492223739624025, 0.023737056732177735, 0.023434528350830076, 0.023372512817382812, 0.023592575073242188, 0.023507232666015624, 0.02350908851623535, 0.02331148719787598, 0.023626623153686525, 0.02336288070678711, 0.023481216430664063, 0.023465120315551757, 0.023558752059936523, 0.02342470359802246, 0.023451103210449218, 0.02366304016113281, 0.022987071990966796, 0.022794208526611327, 0.02259884834289551, 0.022854656219482423, 0.022953216552734374, 0.022876895904541016, 0.022968351364135744, 0.022791488647460938, 0.02270057678222656, 0.02254579162597656, 0.022659328460693358, 0.02253878402709961, 0.022589439392089843, 0.022490655899047852, 0.02284579277038574, 0.022740800857543944, 0.02275974464416504, 0.02272051239013672, 0.02307811164855957, 0.023063199996948242, 0.02305241584777832, 0.02265648078918457, 0.022612512588500975, 0.022622207641601562, 0.02252716827392578, 0.022515520095825196, 0.022395519256591796, 0.022452863693237304, 0.022428800582885742, 0.022450815200805663, 0.02239811134338379, 0.022465471267700196, 0.022514591217041014, 0.023055423736572267, 0.022764575958251952, 0.023169952392578123, 0.02262361526489258, 0.022487680435180665, 0.022583232879638673, 0.023009344100952147, 0.022540288925170897, 0.02260518455505371, 0.022556800842285157, 0.022884864807128907, 0.022584800720214845, 0.022546367645263674, 0.022684255599975587, 0.022601760864257813, 0.02268742370605469, 0.022495519638061522, 0.022413055419921876, 0.022508064270019532, 0.022920032501220704, 0.022791072845458983, 0.0223600959777832, 0.022822240829467773, 0.023083648681640624, 0.02327961540222168, 0.02332057571411133, 0.023275615692138672, 0.023216032028198243, 0.02305023956298828, 0.02309529685974121, 0.023136415481567384, 0.023130239486694334, 0.023096960067749025, 0.022997087478637695, 0.023050207138061524, 0.022556703567504884, 0.022537696838378907, 0.022464607238769533, 0.022485631942749024, 0.02264473533630371, 0.02281657600402832, 0.022736160278320313, 0.022897247314453126, 0.02279987144470215, 0.022714879989624022, 0.022788543701171875, 0.022972095489501954, 0.022732608795166014, 0.022902271270751954, 0.022819520950317383, 0.023005504608154297, 0.023239360809326173, 0.02314886474609375, 0.02313279914855957, 0.023248672485351562, 0.023398271560668947, 0.023087520599365235, 0.022845439910888672, 0.022899999618530273, 0.022694623947143555, 0.02279327964782715, 0.02295395278930664, 0.022744159698486328, 0.022841472625732422, 0.023104511260986327, 0.02281484794616699, 0.022760063171386718, 0.022873344421386718, 0.02270694351196289, 0.022724607467651366, 0.022705215454101563, 0.022786592483520506, 0.022674976348876955, 0.022604415893554688, 0.02278153610229492, 0.02263311958312988, 0.02249456024169922, 0.022629024505615235, 0.022482048034667967, 0.022395679473876953, 0.022394432067871093, 0.02253059196472168, 0.022468191146850586, 0.022440351486206055, 0.022271999359130858, 0.02294902420043945, 0.022653024673461915, 0.022651647567749022, 0.02256812858581543, 0.022698816299438478, 0.022757375717163086, 0.023590080261230467, 0.022707008361816407, 0.022749183654785156, 0.022571008682250978, 0.02252390480041504, 0.022525440216064452, 0.022467071533203126, 0.02299212837219238, 0.022504064559936525, 0.02247897529602051, 0.02245193672180176, 0.022513952255249024, 0.022435840606689454, 0.022550527572631835, 0.022468608856201173, 0.022576704025268554, 0.022503040313720704, 0.022424383163452147, 0.022435136795043945, 0.022532800674438476, 0.02249942398071289, 0.022589311599731446, 0.022517791748046877, 0.022435392379760742, 0.022630367279052734, 0.022534368515014648, 0.022542591094970702, 0.022498432159423827, 0.022468799591064452, 0.022489215850830076, 0.022528575897216796, 0.02272972869873047, 0.022797056198120117, 0.022689088821411133, 0.02247881507873535, 0.022490079879760743, 0.022587392807006838, 0.02243174362182617, 0.022784000396728517, 0.0226627197265625, 0.02252025604248047, 0.022603776931762694, 0.022591487884521484, 0.02271027183532715, 0.02302070426940918, 0.02343612861633301, 0.02594108772277832, 0.02303011131286621, 0.022900447845458985, 0.02263859176635742, 0.022807392120361328, 0.022773759841918945, 0.02286809539794922, 0.02288150405883789, 0.02274086380004883, 0.02267580795288086]",tokens/s,43.83164567021136,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1550.217216,1546.584064,0.0,1168.113664,1154.613248,s,1,8.64842578125,8.64842578125,0.0,8.64842578125,8.64842578125,8.64842578125,8.64842578125,[8.64842578125],,kWh,4.103143184585557e-05,4.51875467082263e-06,1.3285010628000249e-05,5.8835197144678446e-05,,MB,1614.45888,1796.145152,0.0,1388.314624,1334.065152,s,10,0.8219226303100586,0.08219226303100585,0.0004280390602370377,0.08214347076416015,0.08276129989624023,0.08278773918151856,0.08280889060974121,"[0.08281417846679688, 0.0819373779296875, 0.08164854431152344, 0.08275542449951172, 0.08214393615722657, 0.082610595703125, 0.08180518341064454, 0.08247196960449218, 0.08159241485595703, 0.08214300537109374]",tokens/s,3114.6483934069033,kWh,2.5020421523860822e-06,2.759293593684368e-07,1.6584438718633522e-06,4.4364153836178716e-06,tokens/kWh,57704244.950848915,MB,1623.748608,1796.145152,0.0,1388.314624,1372.847616,s,10,14.96638134765625,1.496638134765625,0.009406909194957883,1.4962703857421875,1.5054152587890626,1.5100327026367186,1.5137266577148436,"[1.49269677734375, 1.50438916015625, 1.514650146484375, 1.4816754150390625, 1.48654833984375, 1.499749755859375, 1.5007628173828125, 1.50422900390625, 1.492791015625, 1.488888916015625]",tokens/s,42.09434367370698,kWh,4.24842924109445e-05,4.684985023969847e-06,2.0886351941735733e-05,6.805562937665008e-05,tokens/kWh,925713.2815763119,,s,630,14.96271859169006,0.023750346970936607,0.0003661580478043759,0.023750319480895996,0.024078694343566895,0.02421709747314453,0.024989873924255383,"[0.02380633544921875, 0.023762943267822266, 0.023607391357421875, 0.02354380798339844, 0.023275264739990233, 0.023283008575439454, 0.023229375839233398, 0.023183359146118163, 0.023592960357666014, 0.023416831970214845, 0.024385536193847656, 0.024706239700317382, 0.023783615112304687, 0.023743104934692384, 0.024041471481323243, 0.02397132873535156, 0.023676639556884767, 0.023415584564208985, 0.023458911895751954, 0.023972768783569336, 0.02384486389160156, 0.023412736892700195, 0.023745920181274412, 0.02337980842590332, 0.023310527801513672, 0.023598751068115233, 0.023786239624023438, 0.023750783920288086, 0.02358278465270996, 0.02364112091064453, 0.02396463966369629, 0.023725568771362306, 0.023619903564453124, 0.023675071716308595, 0.023586559295654296, 0.023671039581298826, 0.023658496856689453, 0.023844415664672852, 0.023769535064697266, 0.0236727352142334, 0.023636064529418944, 0.02361510467529297, 0.023546207427978517, 0.023480127334594727, 0.023560415267944335, 0.023631872177124022, 0.023518335342407225, 0.023503488540649414, 0.023717567443847655, 0.02366111946105957, 0.02370355224609375, 0.023769088745117187, 0.023799808502197265, 0.023985504150390625, 0.023896095275878906, 0.02388140869140625, 0.02381100845336914, 0.02373017692565918, 0.023791616439819335, 0.02364998435974121, 0.02373868751525879, 0.023752704620361328, 0.023810047149658203, 0.023736255645751953, 0.023730367660522462, 0.023744319915771483, 0.02378384017944336, 0.02387763214111328, 0.02389811134338379, 0.023823711395263673, 0.023720895767211914, 0.02379952049255371, 0.02370332717895508, 0.023720159530639648, 0.02366979217529297, 0.02381923294067383, 0.02386124801635742, 0.023922016143798828, 0.02406671905517578, 0.024233503341674803, 0.02417856025695801, 0.023800159454345705, 0.024000448226928713, 0.024113183975219728, 0.024078559875488282, 0.02417056083679199, 0.024141727447509767, 0.02396940803527832, 0.023980512619018554, 0.02400592041015625, 0.024105152130126952, 0.024222240447998047, 0.0241213436126709, 0.02386105537414551, 0.023830720901489258, 0.023958528518676758, 0.02396236801147461, 0.02398147201538086, 0.023927648544311522, 0.02386332893371582, 0.024174591064453126, 0.023848928451538087, 0.023760896682739258, 0.023787519454956055, 0.023777280807495117, 0.023807071685791017, 0.023802719116210937, 0.0242729606628418, 0.023896032333374024, 0.02367647933959961, 0.023789215087890624, 0.023743135452270508, 0.02366192054748535, 0.02374291229248047, 0.02383091163635254, 0.023754079818725585, 0.023728479385375978, 0.02387990379333496, 0.023610944747924804, 0.0235762882232666, 0.023685951232910157, 0.02365235137939453, 0.023635967254638672, 0.02370560073852539, 0.023918079376220702, 0.023871999740600586, 0.023883775711059572, 0.023834272384643553, 0.024004831314086914, 0.02396563148498535, 0.02396793556213379, 0.023924736022949217, 0.023982080459594726, 0.023940576553344726, 0.024035871505737303, 0.024410400390625, 0.024126304626464843, 0.02394179153442383, 0.02369990348815918, 0.023778783798217774, 0.02382265663146973, 0.02380985641479492, 0.02381756782531738, 0.023816864013671876, 0.023861663818359375, 0.023836448669433595, 0.023840160369873048, 0.023806560516357423, 0.023803903579711915, 0.023821887969970704, 0.023792064666748047, 0.023911903381347657, 0.024098623275756837, 0.02398896026611328, 0.023910400390625, 0.023988224029541014, 0.0241582088470459, 0.023795711517333985, 0.02389017677307129, 0.023876480102539063, 0.023923583984375, 0.024029184341430664, 0.024104223251342774, 0.024167135238647462, 0.023954591751098632, 0.023878496170043947, 0.023975936889648438, 0.024077375411987303, 0.024036287307739258, 0.024095903396606444, 0.023894847869873045, 0.023740447998046876, 0.023742464065551756, 0.023820255279541017, 0.024213119506835936, 0.02391859245300293, 0.024068511962890626, 0.025285888671875, 0.027642431259155272, 0.023999935150146486, 0.024421119689941408, 0.023995807647705078, 0.023908895492553713, 0.024059968948364256, 0.024195072174072265, 0.023985599517822264, 0.02398908805847168, 0.024044895172119142, 0.023935359954833986, 0.02414240074157715, 0.024014848709106446, 0.023961599349975587, 0.024110271453857423, 0.02407302474975586, 0.024231903076171873, 0.023986207962036134, 0.02397536087036133, 0.02382480049133301, 0.023787935256958007, 0.023869184494018553, 0.023887744903564455, 0.02401638412475586, 0.0240230712890625, 0.023997024536132814, 0.023961599349975587, 0.024037376403808593, 0.02372403144836426, 0.023621631622314454, 0.023901695251464843, 0.024130048751831053, 0.023625728607177734, 0.023442527770996095, 0.023337440490722658, 0.02329644775390625, 0.023389728546142578, 0.023187744140625, 0.023165119171142577, 0.02304204750061035, 0.023128095626831054, 0.023194751739501952, 0.02319856071472168, 0.023181312561035155, 0.023144447326660156, 0.023308000564575194, 0.023361824035644532, 0.023322784423828125, 0.02353545570373535, 0.02351103973388672, 0.02352249526977539, 0.02346668815612793, 0.023432735443115235, 0.023351903915405273, 0.02332057571411133, 0.02345084762573242, 0.023435295104980467, 0.023268096923828124, 0.02326323127746582, 0.0232857608795166, 0.023318527221679687, 0.023307327270507813, 0.023241664886474608, 0.023283231735229493, 0.02329030418395996, 0.02319977569580078, 0.023179264068603517, 0.023080608367919923, 0.023216480255126952, 0.023215232849121095, 0.02312895965576172, 0.023166976928710937, 0.023171072006225587, 0.023169023513793945, 0.02349679946899414, 0.023485599517822267, 0.023452512741088866, 0.023613439559936524, 0.02354380798339844, 0.0234967041015625, 0.024188127517700195, 0.023703935623168946, 0.02372150421142578, 0.0236712646484375, 0.023754175186157227, 0.023684064865112306, 0.023672256469726562, 0.023607872009277345, 0.023785375595092775, 0.023607391357421875, 0.023754751205444336, 0.023651647567749023, 0.023724159240722655, 0.023788095474243164, 0.023684736251831054, 0.023550336837768554, 0.023697216033935545, 0.023654592514038085, 0.02369740867614746, 0.023481632232666017, 0.023756607055664063, 0.02366556739807129, 0.023613311767578124, 0.0235807991027832, 0.02368409538269043, 0.02379395294189453, 0.023722719192504883, 0.023650304794311523, 0.023556095123291015, 0.023433311462402344, 0.024747520446777343, 0.024240383148193358, 0.023642175674438475, 0.023421024322509764, 0.02331430435180664, 0.023183488845825197, 0.02322617530822754, 0.02335353660583496, 0.023288864135742188, 0.023264223098754883, 0.023234016418457033, 0.02329631996154785, 0.023570655822753906, 0.023371776580810546, 0.023267616271972658, 0.024314815521240235, 0.023368480682373047, 0.023322463989257813, 0.02323468780517578, 0.023203872680664064, 0.023279424667358398, 0.023338720321655272, 0.023559839248657226, 0.023535423278808594, 0.023660831451416016, 0.023567071914672853, 0.02373436737060547, 0.02372822380065918, 0.023723360061645507, 0.023646783828735352, 0.023539039611816408, 0.023618175506591798, 0.02343129539489746, 0.02372198486328125, 0.023570432662963867, 0.02367647933959961, 0.023873600006103515, 0.023658912658691408, 0.023586175918579103, 0.023499359130859376, 0.02361100769042969, 0.023609695434570314, 0.023631904602050783, 0.023734272003173826, 0.023580671310424805, 0.023528511047363282, 0.023518144607543947, 0.023740415573120118, 0.02370355224609375, 0.02369270324707031, 0.02389462471008301, 0.024393728256225586, 0.023952959060668945, 0.023883232116699217, 0.023796703338623045, 0.023836383819580077, 0.02393110466003418, 0.023908287048339843, 0.023781503677368164, 0.02381814384460449, 0.023865440368652343, 0.023836544036865234, 0.023779199600219725, 0.023758207321166992, 0.023693695068359374, 0.023654495239257813, 0.023775487899780273, 0.023580831527709963, 0.023658496856689453, 0.02366873550415039, 0.02368307113647461, 0.023826431274414063, 0.02384486389160156, 0.023758848190307616, 0.023838720321655273, 0.023917631149291994, 0.023966655731201172, 0.02394316864013672, 0.024379392623901368, 0.02391561508178711, 0.023954336166381835, 0.02388582420349121, 0.023918079376220702, 0.024045984268188478, 0.024041151046752928, 0.023981855392456054, 0.02397657585144043, 0.0239649600982666, 0.024201183319091796, 0.024169216156005858, 0.023879871368408204, 0.023948831558227537, 0.023875776290893554, 0.0238144645690918, 0.023680864334106447, 0.02360704040527344, 0.02357904052734375, 0.023479711532592772, 0.023871936798095704, 0.023535776138305663, 0.023825504302978515, 0.023738847732543946, 0.023808576583862304, 0.02430544090270996, 0.024410015106201173, 0.024332639694213867, 0.0238057918548584, 0.02375017547607422, 0.02368355178833008, 0.02352889633178711, 0.023537599563598632, 0.0235546875, 0.023547103881835937, 0.023415584564208985, 0.023390207290649414, 0.02355583953857422, 0.023529727935791014, 0.023467199325561523, 0.023388704299926757, 0.023362943649291992, 0.023393184661865234, 0.023750463485717774, 0.02434681510925293, 0.0253798713684082, 0.02373731231689453, 0.023505088806152343, 0.023314016342163086, 0.023388320922851563, 0.02332415962219238, 0.02338604736328125, 0.023805856704711914, 0.024320735931396484, 0.024049663543701173, 0.023772159576416017, 0.023532543182373047, 0.025851903915405275, 0.02457804870605469, 0.02367283248901367, 0.023801023483276368, 0.023797855377197266, 0.02368953514099121, 0.023625120162963868, 0.023758975982666016, 0.023931743621826172, 0.02377110481262207, 0.02384492874145508, 0.023842815399169923, 0.024127071380615234, 0.023869535446166993, 0.023980096817016603, 0.023963615417480467, 0.024010751724243166, 0.023977472305297853, 0.02386387252807617, 0.02387353515625, 0.02371513557434082, 0.023756864547729493, 0.02378816032409668, 0.023883295059204102, 0.02382896041870117, 0.024020639419555664, 0.023609695434570314, 0.02364825630187988, 0.023536895751953123, 0.02363849639892578, 0.023666976928710937, 0.023756799697875978, 0.023948831558227537, 0.023799936294555665, 0.02363369560241699, 0.023728063583374023, 0.023757408142089844, 0.023689247131347655, 0.02391244888305664, 0.023740415573120118, 0.023842815399169923, 0.023830528259277343, 0.02382956886291504, 0.023901119232177734, 0.02407539176940918, 0.02398201560974121, 0.0239520320892334, 0.02390768051147461, 0.024017375946044923, 0.023982559204101563, 0.023887136459350585, 0.023816352844238282, 0.023881536483764648, 0.024017663955688478, 0.023812095642089845, 0.02390969657897949, 0.024220352172851563, 0.024079904556274415, 0.023921119689941407, 0.024071231842041015, 0.023952320098876954, 0.02401590347290039, 0.023935968399047852, 0.023992319107055664, 0.02388172721862793, 0.024201215744018553, 0.02388991928100586, 0.0240676155090332, 0.023807615280151368, 0.023935903549194337, 0.023981023788452148, 0.023804479598999024, 0.02379155158996582, 0.023772863388061522, 0.023798559188842775, 0.023611391067504883, 0.023610528945922853, 0.025088863372802736, 0.023813600540161132, 0.023751136779785156, 0.023517248153686523, 0.023779327392578126, 0.023602336883544923, 0.023452512741088866, 0.023428735733032228, 0.023343488693237303, 0.02348195266723633, 0.02354854393005371, 0.02337887954711914, 0.02349888038635254, 0.02341142463684082, 0.023400447845458985, 0.02371788787841797, 0.02406399917602539, 0.023939008712768554, 0.023869472503662108, 0.024172224044799805, 0.023875072479248048, 0.023824832916259767, 0.024041439056396486, 0.024080928802490235, 0.02406108856201172, 0.02413849639892578, 0.024035327911376952, 0.024495744705200197, 0.024295808792114258, 0.024159231185913087, 0.024738815307617186, 0.0242457275390625, 0.024043968200683594, 0.02411734390258789, 0.024096031188964844, 0.023892704010009765, 0.023932928085327147, 0.02389561653137207, 0.023828927993774413, 0.02373222351074219, 0.023793024063110353, 0.023661184310913085, 0.023434911727905273, 0.023347551345825196, 0.023389888763427735, 0.023303647994995118, 0.023432031631469726, 0.023459840774536132, 0.02354368019104004, 0.0234149112701416, 0.023345151901245118, 0.023390207290649414, 0.023360511779785157, 0.023419904708862304, 0.023230464935302734, 0.023053632736206055, 0.023214784622192383, 0.02334851264953613, 0.023274208068847658, 0.023371776580810546, 0.02328985595703125, 0.023617536544799804, 0.02353705596923828, 0.02362019157409668, 0.02373161506652832, 0.02366320037841797, 0.023580671310424805, 0.023758495330810547, 0.023785408020019532, 0.023663232803344727, 0.02367081642150879, 0.02370345687866211, 0.023670879364013672, 0.02366054344177246, 0.023463615417480467, 0.02348988723754883, 0.02595948791503906, 0.025237375259399415, 0.023840576171875, 0.023843040466308595, 0.023504896163940428, 0.023326719284057617, 0.023255136489868163, 0.02325289535522461, 0.023154239654541015, 0.023198144912719727, 0.023203840255737306, 0.023257087707519532, 0.02306172752380371, 0.02313091278076172, 0.023111679077148437, 0.023408639907836915, 0.02411075210571289, 0.023654048919677734, 0.023642240524291994, 0.023669599533081054, 0.02380975914001465, 0.023785472869873047, 0.023991296768188477, 0.023809024810791016, 0.023653375625610353, 0.023579647064208984, 0.023441247940063477, 0.023471935272216797, 0.023531871795654295, 0.02371295928955078, 0.02347292709350586, 0.02338991928100586, 0.023412256240844725, 0.02339859199523926, 0.023742111206054687, 0.023665599822998047, 0.023830528259277343, 0.02357040023803711, 0.02346518325805664, 0.02381702423095703, 0.023586816787719726, 0.023599103927612306, 0.023371103286743165, 0.02350966453552246, 0.023410688400268553, 0.023365631103515624, 0.02360540771484375, 0.02351702308654785, 0.023782848358154297, 0.024240703582763673, 0.02364112091064453, 0.02354275131225586, 0.02353971290588379, 0.023638015747070314]",tokens/s,42.10464803835093,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,6729.650176,7525.564416,0.0,7147.094016,7138.9184,s,1,11.4090703125,11.4090703125,0.0,11.4090703125,11.4090703125,11.4090703125,11.4090703125,[11.4090703125],,kWh,0.00012896135883750805,1.4217553013790054e-05,4.2785312005994e-05,0.0001859642238572921,,MB,1605.595136,8226.013184,0.0,7818.182656,7724.300288,s,10,6.541907592773438,0.6541907592773437,0.0010968707736659236,0.6545454406738281,0.6554325134277343,0.6555990386962891,0.6557322589111328,"[0.6523662719726563, 0.6524636840820313, 0.6538348388671875, 0.6533842163085938, 0.6545287475585938, 0.6553955078125, 0.6548684692382812, 0.6547381591796875, 0.6545621337890625, 0.6557655639648438]",tokens/s,391.32316739354764,kWh,1.9139248714321864e-05,2.110738477930151e-06,1.2704662941499106e-05,3.395465013375112e-05,tokens/kWh,7539468.054937621,MB,1611.640832,8372.813824,0.0,7964.983296,7904.605696,s,10,30.872391845703124,3.087239184570312,0.009571757553960468,3.0865322265625,3.094320703125,3.1031818359375,3.1102707421875,"[3.077819091796875, 3.076754638671875, 3.08461279296875, 3.07926171875, 3.087107666015625, 3.087256103515625, 3.089228515625, 3.085956787109375, 3.0923515625, 3.11204296875]",tokens/s,20.40658213813403,kWh,8.99548742615112e-05,9.922422753871875e-06,5.9754061692101484e-05,0.00015963135870748454,tokens/kWh,394659.29821122397,,s,630,30.868402111053467,0.04899746366833883,0.000977710370331209,0.048934560775756836,0.04963838310241699,0.04993952903747558,0.05121825778961182,"[0.0511861457824707, 0.04846723175048828, 0.04803587341308594, 0.04802835083007812, 0.048527359008789066, 0.048216064453125, 0.04802969741821289, 0.047993854522705076, 0.047874656677246094, 0.04905615997314453, 0.04876697540283203, 0.048246784210205076, 0.04849200057983399, 0.04856681442260742, 0.04820102310180664, 0.04856902313232422, 0.048574462890625, 0.04836761474609375, 0.049616897583007816, 0.04921548843383789, 0.04869036865234375, 0.0485153923034668, 0.04882688140869141, 0.04887065505981445, 0.04970985412597656, 0.048368927001953124, 0.04843727874755859, 0.0487922248840332, 0.04877721786499024, 0.04836966323852539, 0.04904755020141602, 0.04895334243774414, 0.04922067260742188, 0.04842793655395508, 0.04895452880859375, 0.048632190704345706, 0.04861731338500976, 0.04861609649658203, 0.04899020767211914, 0.048936958312988284, 0.04855801773071289, 0.04964966583251953, 0.049079456329345704, 0.04900751876831055, 0.04865024185180664, 0.0493383674621582, 0.04933817672729492, 0.04878281784057617, 0.04896380615234375, 0.04905625534057617, 0.04910079956054687, 0.049235233306884764, 0.04918534469604492, 0.04887670516967774, 0.04871433639526367, 0.04894927978515625, 0.04920563125610351, 0.04914585494995117, 0.04892470550537109, 0.049538463592529294, 0.04931142425537109, 0.049592288970947265, 0.049414623260498045, 0.05143142318725586, 0.04864614486694336, 0.04810137557983398, 0.047768638610839846, 0.047700927734375, 0.047508575439453124, 0.04803062438964844, 0.048639999389648435, 0.0488812141418457, 0.04853219223022461, 0.0483691520690918, 0.048172286987304684, 0.04848534393310547, 0.04876435089111328, 0.04869174575805664, 0.04877132797241211, 0.04833052825927735, 0.04869094467163086, 0.04916044616699219, 0.04934860610961914, 0.04927078247070313, 0.04863155364990234, 0.04868531036376953, 0.04806198501586914, 0.04839267349243164, 0.04816656112670899, 0.04837615966796875, 0.04834678268432617, 0.048882015228271486, 0.04875836944580078, 0.04839670562744141, 0.04872185516357422, 0.04912271881103516, 0.048615806579589846, 0.04923183822631836, 0.04884921646118164, 0.04869055938720703, 0.04894518280029297, 0.049310302734375, 0.04903500747680664, 0.04870374298095703, 0.0489697265625, 0.049006591796875, 0.048811840057373046, 0.04921772766113281, 0.04910079956054687, 0.04883660888671875, 0.04878934478759766, 0.04894121551513672, 0.04865433502197265, 0.04894841766357422, 0.04905657577514649, 0.049414142608642575, 0.049119232177734375, 0.04881139373779297, 0.04918540954589844, 0.04933631896972656, 0.04925395202636719, 0.050059486389160156, 0.04977062225341797, 0.0490599365234375, 0.049209342956542966, 0.04957388687133789, 0.05131222534179687, 0.04826124954223633, 0.04803968048095703, 0.04798252868652344, 0.04846448135375977, 0.04786380767822265, 0.04809318542480469, 0.04869852828979492, 0.04877091217041016, 0.048288352966308595, 0.04828966522216797, 0.04813059234619141, 0.048807937622070315, 0.04840038299560547, 0.04871145629882812, 0.04855356979370117, 0.048415359497070314, 0.049408000946044923, 0.04879897689819336, 0.04953523254394531, 0.04994508743286133, 0.049167903900146484, 0.04876502227783203, 0.04884310531616211, 0.04845340728759766, 0.04849395370483398, 0.04879756927490234, 0.04872499084472656, 0.04825804901123047, 0.04849356842041016, 0.048840320587158204, 0.049053825378417966, 0.0487056655883789, 0.048711521148681644, 0.04950374221801758, 0.048509185791015624, 0.04899657440185547, 0.04925881576538086, 0.04916428756713867, 0.04893286514282227, 0.0501104621887207, 0.04923593521118164, 0.04928460693359375, 0.049207839965820316, 0.0486495361328125, 0.049056449890136716, 0.04882022476196289, 0.0488199348449707, 0.04887567901611328, 0.049078399658203126, 0.04951039886474609, 0.049065662384033204, 0.04898233413696289, 0.0491640625, 0.04934230422973633, 0.04891686248779297, 0.04921343994140625, 0.049470783233642575, 0.04921798324584961, 0.04960691070556641, 0.04971321487426758, 0.05123065567016601, 0.049169857025146486, 0.05111603164672852, 0.048981822967529294, 0.048506175994873044, 0.04861017608642578, 0.048903358459472655, 0.04848630523681641, 0.04821228790283203, 0.04822390365600586, 0.0485549430847168, 0.048164161682128906, 0.048470497131347656, 0.0485781135559082, 0.0484334716796875, 0.048255329132080076, 0.04819353485107422, 0.048963584899902345, 0.04873625564575195, 0.04829289627075195, 0.049212318420410156, 0.04910662460327148, 0.04905612945556641, 0.048707584381103515, 0.04871782302856445, 0.04911260986328125, 0.04900230407714844, 0.04855875015258789, 0.04885689544677734, 0.048596446990966796, 0.04895388793945313, 0.04880099105834961, 0.048567264556884766, 0.04838399887084961, 0.049030399322509764, 0.04850777435302735, 0.04925558471679688, 0.049080543518066407, 0.04886783981323242, 0.04876809692382812, 0.048630657196044924, 0.04999900817871094, 0.04873846435546875, 0.049291999816894534, 0.04926252746582031, 0.04882751846313477, 0.04877580642700195, 0.04850755310058594, 0.04886428833007812, 0.048845439910888674, 0.04898611068725586, 0.048963584899902345, 0.04871782302856445, 0.04867891311645508, 0.04907827377319336, 0.04914169692993164, 0.04917049789428711, 0.049251487731933594, 0.048747360229492186, 0.04904531097412109, 0.04924403381347656, 0.04900614547729492, 0.049422592163085935, 0.049388031005859374, 0.04950812911987305, 0.052891456604003906, 0.049322177886962894, 0.04810137557983398, 0.04811126327514648, 0.048656734466552734, 0.048159744262695314, 0.0477416000366211, 0.04853987121582031, 0.04840163040161133, 0.04803267288208008, 0.04841062545776367, 0.04832361602783203, 0.04863616180419922, 0.04891302490234375, 0.048783454895019535, 0.04860313415527344, 0.0484598388671875, 0.04877260971069336, 0.048925121307373046, 0.05010419082641602, 0.049852542877197266, 0.04863180923461914, 0.04856217575073242, 0.04828364944458008, 0.0487927360534668, 0.049496318817138674, 0.04850339126586914, 0.04856787109375, 0.04844287872314453, 0.04887033462524414, 0.049012607574462894, 0.048457183837890626, 0.04921001434326172, 0.05062646484375, 0.048293087005615236, 0.04891532897949219, 0.04912300872802734, 0.04872351837158203, 0.04964745712280273, 0.04958505630493164, 0.04912646484375, 0.04903769683837891, 0.04907267379760742, 0.04899023818969726, 0.048773342132568356, 0.04880534362792969, 0.04886707305908203, 0.04899283218383789, 0.049235969543457034, 0.04890828704833984, 0.04890419387817383, 0.04903923034667969, 0.04901811218261719, 0.04900259017944336, 0.04909648132324219, 0.049510944366455076, 0.04944883346557617, 0.049310302734375, 0.04950425720214844, 0.05018624114990235, 0.049768001556396484, 0.04933004760742187, 0.04927936172485352, 0.05104703903198242, 0.048719070434570313, 0.04796495819091797, 0.04776335906982422, 0.04804412841796875, 0.047808158874511716, 0.04817750549316406, 0.04833280181884766, 0.04878950500488281, 0.0486657600402832, 0.04810838317871094, 0.04840604782104492, 0.048329185485839844, 0.04839648056030273, 0.048406272888183596, 0.048191551208496095, 0.04813603210449219, 0.04840367889404297, 0.04976236724853516, 0.04993273544311523, 0.04977094268798828, 0.04938751983642578, 0.04915126419067383, 0.04871440124511719, 0.048398399353027345, 0.048361473083496094, 0.04888115310668945, 0.04869171142578125, 0.049018878936767575, 0.049168384552001954, 0.049005599975585935, 0.04871478271484375, 0.04920035171508789, 0.04913398361206055, 0.048908607482910156, 0.048742401123046876, 0.049176513671875, 0.05066950225830078, 0.049285247802734376, 0.04984009552001953, 0.04973366546630859, 0.049248062133789065, 0.04934265518188476, 0.048844223022460935, 0.04896416091918945, 0.0493568000793457, 0.049342369079589846, 0.04939785766601563, 0.04928489685058594, 0.04860540771484375, 0.05005231857299805, 0.04913011169433594, 0.04909891128540039, 0.048814079284667966, 0.04909260940551758, 0.04934009552001953, 0.049600833892822264, 0.04904550552368164, 0.04884604644775391, 0.04943731307983398, 0.04972150421142578, 0.049637374877929685, 0.04930748748779297, 0.05118790435791016, 0.048519168853759766, 0.04838396835327148, 0.048095264434814454, 0.04855807876586914, 0.048123649597167965, 0.0484477424621582, 0.04867839813232422, 0.04843366241455078, 0.04828160095214844, 0.048936256408691405, 0.04914019012451172, 0.0491517105102539, 0.048978431701660156, 0.04936281585693359, 0.04815990447998047, 0.04883350372314453, 0.04862905502319336, 0.048855743408203124, 0.04910038375854492, 0.049094047546386715, 0.04894412612915039, 0.04866361618041992, 0.04857952117919922, 0.04869062423706055, 0.04855174255371094, 0.04889267349243164, 0.04840857696533203, 0.048451583862304685, 0.04866457748413086, 0.04885488128662109, 0.04917059326171875, 0.049426433563232425, 0.04930559921264648, 0.04916630554199219, 0.04914790344238281, 0.04915203094482422, 0.04899391937255859, 0.049092510223388675, 0.048964065551757814, 0.04936284637451172, 0.04968790435791016, 0.04949478530883789, 0.04914329528808594, 0.0492509765625, 0.04882979202270508, 0.04915484619140625, 0.04869692611694336, 0.04895756912231446, 0.0491069450378418, 0.0490250244140625, 0.04947087860107422, 0.04923657608032227, 0.049075519561767575, 0.0494024658203125, 0.049425792694091794, 0.04912815856933594, 0.049825790405273435, 0.049601665496826174, 0.049593215942382814, 0.04960255813598633, 0.050003360748291016, 0.04965193557739258, 0.051533824920654295, 0.04860518264770508, 0.04825222396850586, 0.04814713668823242, 0.04884659194946289, 0.0482918701171875, 0.04817737579345703, 0.04801257705688477, 0.04853833770751953, 0.04871987152099609, 0.04885504150390625, 0.04856137466430664, 0.0486379508972168, 0.04865718460083008, 0.048535552978515625, 0.048551937103271485, 0.048527359008789066, 0.049000415802001956, 0.049261985778808595, 0.04945980834960938, 0.04997328186035156, 0.04969267272949219, 0.0490332145690918, 0.04899020767211914, 0.048898048400878906, 0.04841830444335937, 0.04846745681762695, 0.04866128158569336, 0.0488870735168457, 0.04851504135131836, 0.04893795013427735, 0.04899225616455078, 0.04874176025390625, 0.04910704040527344, 0.04927926254272461, 0.049333568572998046, 0.04859795379638672, 0.04883660888671875, 0.04918272018432617, 0.04894105529785156, 0.0495816650390625, 0.04944284820556641, 0.04896710586547852, 0.04921235275268555, 0.04878908920288086, 0.048720287322998046, 0.04848230361938476, 0.049209342956542966, 0.04861737442016602, 0.048750686645507815, 0.049235969543457034, 0.049069854736328126, 0.04893718338012695, 0.04892038345336914, 0.04893920135498047, 0.04843724822998047, 0.04907379150390625, 0.04921177673339844, 0.049733345031738284, 0.0496888656616211, 0.050151134490966795, 0.050014495849609375, 0.04971686553955078, 0.05080665588378906, 0.048567935943603514, 0.048161598205566404, 0.048510974884033206, 0.048492542266845705, 0.048366592407226565, 0.04816793441772461, 0.048244384765625, 0.04878780746459961, 0.048280960083007814, 0.048366207122802735, 0.04940390396118164, 0.049582080841064455, 0.048691200256347655, 0.048123905181884766, 0.04846131134033203, 0.04896521759033203, 0.048487327575683595, 0.048844799041748044, 0.049127422332763675, 0.04952463912963867, 0.049031265258789064, 0.049069950103759766, 0.048752574920654296, 0.048764575958251954, 0.04925494384765625, 0.0495797119140625, 0.048582912445068356, 0.0487916145324707, 0.049152000427246094, 0.04909875106811523, 0.049772544860839846, 0.049342369079589846, 0.04884489440917969, 0.04896710586547852, 0.049138240814208985, 0.04909875106811523, 0.04876406478881836, 0.04973791885375976, 0.04949795150756836, 0.0493408317565918, 0.04932854461669922, 0.0498581771850586, 0.049398143768310546, 0.04897587203979492, 0.049143585205078125, 0.048844799041748044, 0.04898633575439453, 0.049205249786376956, 0.04951788711547851, 0.049464065551757815, 0.048850879669189454, 0.0491879997253418, 0.04884348678588867, 0.04933849716186523, 0.04928102493286133, 0.04943667221069336, 0.04958380889892578, 0.04959059143066406, 0.04967628860473633, 0.049377281188964846, 0.04979075241088867, 0.04970124816894531, 0.05191968154907227, 0.06838175964355468, 0.04760870361328125, 0.04880393600463867, 0.04853142547607422, 0.048424415588378907, 0.04827795028686523, 0.047958110809326174, 0.048469631195068356, 0.047973793029785154, 0.04838889694213867, 0.04820601654052734, 0.04816886520385742, 0.04845577621459961, 0.048465633392333986, 0.04883280181884766, 0.04856217575073242, 0.04833420944213867, 0.049429119110107424, 0.050740673065185544, 0.050119232177734375, 0.0493568000793457, 0.049211006164550784, 0.04855152130126953, 0.04902377700805664, 0.049084415435791014, 0.048889854431152346, 0.048732158660888675, 0.04864745712280273, 0.048523998260498045, 0.04969062423706055, 0.04896460723876953, 0.048950271606445314, 0.048658432006835936, 0.04900864028930664, 0.0486596794128418, 0.04889782333374024, 0.04958652877807617, 0.05011727905273437, 0.05005855941772461, 0.04968719863891601, 0.04950982284545898, 0.049707294464111325, 0.04937350463867188, 0.04914995193481445, 0.0489238395690918, 0.049003326416015625, 0.048976993560791014, 0.04897270584106445, 0.04929478454589844, 0.04914172744750977, 0.04891459274291992, 0.04880428695678711, 0.048928768157958984, 0.049014785766601565, 0.0486707534790039, 0.0494601936340332, 0.04949913787841797, 0.05009612655639648, 0.04970086288452148, 0.05040332794189453, 0.05008544158935547, 0.049711231231689454]",tokens/s,20.409219684695227,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,11127.222272,12225.282048,0.0,11846.811648,11814.785024,s,1,14.2289521484375,14.2289521484375,0.0,14.2289521484375,14.2289521484375,14.2289521484375,14.2289521484375,[14.2289521484375],,kWh,0.0002074531606000164,2.287619585885001e-05,6.829616574800107e-05,0.0002986255222068675,,MB,2072.178688,13158.514688,0.0,12750.68416,12641.86368,s,10,11.69583349609375,1.1695833496093748,0.0011074682986908374,1.1700194702148439,1.1705676147460937,1.1707565246582032,1.1709076525878908,"[1.1671827392578125, 1.1681357421875, 1.169578125, 1.170525634765625, 1.17003515625, 1.1700037841796875, 1.1703682861328124, 1.1689547119140624, 1.1709454345703125, 1.1701038818359375]",tokens/s,218.88136496257457,kWh,3.419648953291509e-05,3.7690069015955685e-06,2.2736184855599863e-05,6.070168129011051e-05,tokens/kWh,4217346.118907375,MB,2076.442624,13452.115968,0.0,13044.28544,12933.698048,s,10,50.42909423828126,5.0429094238281245,0.010713131044623299,5.045888671875,5.05390078125,5.0548341796875,5.0555808984374995,"[5.02344580078125, 5.03154052734375, 5.0300498046875, 5.03985888671875, 5.04284716796875, 5.055767578125, 5.04893017578125, 5.051208984375, 5.053693359375, 5.051751953125]",tokens/s,12.492788330149315,kWh,0.00014738280290417,1.6259827501932307e-05,9.79809950513999e-05,0.0002616236254575022,tokens/kWh,240803.94073674222,,s,630,50.42495937347408,0.08003961805313353,0.0012047015714267615,0.08003118515014648,0.08085504379272461,0.08119948577880859,0.08279828002929687,"[0.08202239990234375, 0.07817011260986328, 0.07856742095947265, 0.07859164428710938, 0.07911459350585938, 0.07797328186035156, 0.07938633728027343, 0.07966156768798828, 0.07951071929931641, 0.07934611511230469, 0.07875580596923829, 0.08002710723876953, 0.08064275360107422, 0.0799450912475586, 0.07833299255371094, 0.0786655044555664, 0.07802790069580078, 0.07945104217529297, 0.07972589111328125, 0.07980902099609374, 0.07916969299316406, 0.07876345825195312, 0.07837369537353515, 0.07974671936035156, 0.08063785552978515, 0.08017228698730469, 0.07948745727539062, 0.07952790069580078, 0.07863760375976563, 0.07819459533691406, 0.07928566741943359, 0.079727294921875, 0.08014546966552734, 0.08003065490722656, 0.07963648223876953, 0.07846633911132812, 0.07965360260009766, 0.08066777801513672, 0.08011660766601562, 0.08049606323242188, 0.07969235229492187, 0.07981791687011719, 0.07998323059082031, 0.08002690887451172, 0.07951862335205079, 0.08059913635253907, 0.08027088165283203, 0.08024668884277343, 0.07993936157226562, 0.08069395446777344, 0.08069107055664063, 0.08032585906982422, 0.08090716552734376, 0.07994687652587891, 0.08025138854980468, 0.08074240112304687, 0.07999942779541015, 0.07999686431884766, 0.08049049377441406, 0.08077705383300782, 0.08001363372802735, 0.08046339416503906, 0.08096578979492187, 0.0835343017578125, 0.07810419464111328, 0.07861702728271484, 0.07875993347167969, 0.07910320281982422, 0.07794563293457031, 0.07910720062255859, 0.0795841293334961, 0.07844659423828125, 0.07941075134277344, 0.07931241607666016, 0.08012470245361328, 0.08071308898925782, 0.07985024261474609, 0.07810867309570313, 0.07891295623779297, 0.07934611511230469, 0.07939897918701172, 0.07895174407958984, 0.07971241760253907, 0.07925743865966797, 0.07929933166503907, 0.07827251434326171, 0.07999488067626953, 0.08039833831787109, 0.08075654602050782, 0.07964649963378906, 0.07949763488769532, 0.0801434555053711, 0.07950224304199219, 0.08066819000244141, 0.07974671936035156, 0.07939974212646485, 0.08008704376220703, 0.07939891052246094, 0.07908716583251953, 0.08019503784179688, 0.07970719909667968, 0.08031632232666015, 0.07902758026123047, 0.08018598175048829, 0.07970611572265625, 0.08047135925292968, 0.07975580596923829, 0.07994793701171875, 0.08039218902587891, 0.08016671752929687, 0.08014867401123046, 0.08084870147705078, 0.08029811096191407, 0.08087353515625, 0.08011097717285157, 0.08075328063964844, 0.08015475463867187, 0.08049043273925781, 0.07981644439697265, 0.08087718200683594, 0.08140652465820312, 0.07990681457519531, 0.08044963073730468, 0.08093686676025391, 0.08105101013183594, 0.08097036743164063, 0.08178422546386718, 0.0778737564086914, 0.07885311889648437, 0.07947161865234376, 0.0794439697265625, 0.07947203063964844, 0.07845500946044921, 0.07841011047363282, 0.07793379211425781, 0.07976834869384766, 0.07905484771728516, 0.07920130920410157, 0.08019023895263672, 0.07980665588378906, 0.07923302459716797, 0.08001126098632813, 0.07951503753662109, 0.07991356658935547, 0.07924230194091797, 0.07892626953125, 0.0789184341430664, 0.0800191650390625, 0.07949465942382812, 0.07904086303710937, 0.08089395141601563, 0.08018748474121094, 0.08011170959472656, 0.08039421081542969, 0.07999839782714843, 0.07934595489501953, 0.07966544342041015, 0.0795990753173828, 0.07917158508300781, 0.07998928070068359, 0.07979622650146484, 0.08026521301269532, 0.07939891052246094, 0.07992652893066406, 0.08001187133789063, 0.08014454650878906, 0.08001126098632813, 0.0796033935546875, 0.08004640197753907, 0.07982284545898438, 0.0788359375, 0.07995321655273438, 0.08008096313476562, 0.0805823974609375, 0.0799566421508789, 0.07983513641357422, 0.08144204711914063, 0.08071033477783203, 0.08030332946777344, 0.0803674545288086, 0.0819947509765625, 0.0795832290649414, 0.08040447998046875, 0.08035123443603516, 0.0800747528076172, 0.08039833831787109, 0.08110899353027344, 0.08029798126220702, 0.08095334625244141, 0.0828392333984375, 0.07806198120117187, 0.07916134643554687, 0.07862214660644531, 0.07906339263916015, 0.07891990661621094, 0.07980441284179687, 0.07903132629394531, 0.07842095947265625, 0.07847917175292969, 0.0784051513671875, 0.07905709075927735, 0.08156195068359375, 0.0805827865600586, 0.07967948913574219, 0.07941840362548828, 0.07950640106201172, 0.07891792297363281, 0.07928988647460937, 0.08475667572021485, 0.08009728240966797, 0.07883958435058594, 0.07980802917480469, 0.0797085418701172, 0.08099667358398438, 0.08063516998291016, 0.08024070739746093, 0.08033110046386718, 0.0795726089477539, 0.0801596450805664, 0.08076390075683594, 0.07972329711914063, 0.07857913970947265, 0.07829503631591797, 0.0798326416015625, 0.08000128173828125, 0.08059161376953125, 0.08047545623779297, 0.08052336120605469, 0.07986646270751953, 0.07891725158691407, 0.08057231903076172, 0.07993532562255859, 0.07992784118652344, 0.07938278198242188, 0.07981858825683594, 0.0806379165649414, 0.08003734588623047, 0.08007942199707031, 0.08065634918212891, 0.08014771270751953, 0.08071183776855469, 0.08021651458740234, 0.08061971282958984, 0.08085298919677734, 0.08048639678955079, 0.07980643463134765, 0.07986179351806641, 0.0805514907836914, 0.08051248168945313, 0.08056288146972657, 0.08138780975341797, 0.08117657470703125, 0.08308306884765625, 0.078017822265625, 0.0793375015258789, 0.07864611053466797, 0.07843014526367187, 0.07839110565185547, 0.07856768035888671, 0.07947596740722657, 0.07975299072265625, 0.0790599365234375, 0.07823564910888672, 0.08095318603515625, 0.08110899353027344, 0.07955996704101563, 0.07958617401123047, 0.07955865478515625, 0.07912416076660156, 0.07958076477050781, 0.07961177825927734, 0.0796592025756836, 0.08012457275390625, 0.07991670227050782, 0.07975279998779297, 0.08049689483642578, 0.08065484619140625, 0.080500732421875, 0.07977983856201172, 0.07979779052734375, 0.07927375793457031, 0.08083936309814453, 0.08047561645507813, 0.08018179321289062, 0.08013321685791015, 0.08013673400878907, 0.07979865264892579, 0.08054579162597657, 0.0803616943359375, 0.08037763214111328, 0.08008620452880859, 0.07965078735351562, 0.07967852783203125, 0.08035206604003907, 0.07899849700927734, 0.07948697662353515, 0.08041661071777344, 0.08098627471923828, 0.08059001922607421, 0.08056915283203125, 0.08063542175292969, 0.08012643432617188, 0.08127446746826172, 0.08047388458251953, 0.08060582733154296, 0.07966925048828125, 0.08074211120605469, 0.0802523193359375, 0.08067881774902344, 0.0807147216796875, 0.08044748687744141, 0.08080537414550781, 0.08026950073242188, 0.08128569793701172, 0.08078025817871094, 0.08260041809082032, 0.07780124664306641, 0.078974365234375, 0.07772457885742187, 0.07980841827392578, 0.07981228637695313, 0.07924972534179688, 0.07912844848632812, 0.079665283203125, 0.07954150390625, 0.07949298858642578, 0.07997325134277344, 0.08013385772705078, 0.07919235229492187, 0.07950950622558593, 0.07904374694824219, 0.07904547119140624, 0.08030617523193359, 0.07933132934570312, 0.0795948486328125, 0.07898563385009766, 0.08004633331298829, 0.07961395263671875, 0.08034480285644531, 0.08109468841552735, 0.08066687774658203, 0.07996211242675781, 0.07929241943359375, 0.07982284545898438, 0.07984754943847656, 0.08003961944580078, 0.08027369689941406, 0.07988419342041016, 0.07969586944580079, 0.07988224029541016, 0.08014768218994141, 0.08029878234863282, 0.0798883819580078, 0.08016690826416016, 0.07977558135986328, 0.10099523162841798, 0.0751094741821289, 0.07972054290771484, 0.07915174102783203, 0.07932947540283203, 0.08023654174804687, 0.07990886688232422, 0.08022425842285157, 0.08081785583496094, 0.08087789154052734, 0.08048844909667968, 0.08153702545166015, 0.08276742553710938, 0.08016326141357422, 0.07964672088623047, 0.08062361907958984, 0.08073011016845703, 0.08084889221191406, 0.08042716979980469, 0.07946943664550782, 0.08045868682861328, 0.08133350372314453, 0.08083535766601563, 0.08281088256835938, 0.07819468688964844, 0.0787837142944336, 0.07909638214111328, 0.0784095687866211, 0.07843059539794922, 0.07951181030273438, 0.07959117126464844, 0.07859398651123047, 0.07862892913818359, 0.07986380767822265, 0.08073216247558594, 0.07984703826904296, 0.07969769287109375, 0.08000163269042969, 0.07982694244384765, 0.07962828826904297, 0.07913200378417969, 0.08030646514892578, 0.07916582489013672, 0.07974515533447266, 0.07991801452636718, 0.08004422760009766, 0.08020995330810547, 0.08005705261230468, 0.08031215667724609, 0.07999644470214844, 0.08046656036376953, 0.07968153381347656, 0.07918182373046875, 0.08052735900878906, 0.07966310119628907, 0.08011145782470704, 0.08012748718261718, 0.08035574340820313, 0.08065849304199219, 0.08064742279052735, 0.08028575897216797, 0.08058354949951171, 0.08019558715820313, 0.0801607666015625, 0.08009932708740235, 0.0806645736694336, 0.07974297332763672, 0.08039218902587891, 0.08077926635742187, 0.08033235168457031, 0.08140230560302734, 0.08115945434570312, 0.08065020751953125, 0.08051360321044922, 0.08078099060058594, 0.08051900482177735, 0.08082669067382812, 0.08073177337646484, 0.08044617462158203, 0.08070105743408203, 0.08081369781494141, 0.08031308746337891, 0.08110284423828125, 0.08118681335449218, 0.08131378936767578, 0.08087776184082031, 0.08203343963623047, 0.0788705291748047, 0.07800211334228516, 0.0788212127685547, 0.07983235168457031, 0.07948544311523438, 0.07987654113769531, 0.07957504272460937, 0.07935737609863282, 0.07880556488037109, 0.07954431915283203, 0.08023177337646484, 0.079917724609375, 0.08033916473388672, 0.07889282989501953, 0.07928006744384766, 0.07998265838623046, 0.07980960083007813, 0.07996102142333984, 0.07972892761230468, 0.07998191833496093, 0.07984371185302734, 0.08030207824707031, 0.08003379058837891, 0.08021526336669922, 0.08027011108398438, 0.07981465911865235, 0.0796874237060547, 0.07962239837646484, 0.08001760101318359, 0.08017708587646484, 0.08005353546142578, 0.08002953338623046, 0.07935574340820313, 0.07980124664306641, 0.08065843200683594, 0.0809184341430664, 0.08074588775634765, 0.08022905731201171, 0.08016281890869141, 0.08080786895751953, 0.08060912322998047, 0.08020764923095704, 0.08062406158447266, 0.08066643524169922, 0.08069929504394531, 0.08029827117919922, 0.08076902770996094, 0.08056124877929688, 0.08175504302978516, 0.0803430404663086, 0.08018915557861328, 0.08071395111083984, 0.08028575897216797, 0.08060041809082032, 0.080623779296875, 0.08120985412597656, 0.08060256195068359, 0.08084524536132813, 0.0807508773803711, 0.08079872131347657, 0.08210313415527344, 0.0803799057006836, 0.08197731018066406, 0.07891353607177734, 0.07926681518554687, 0.07878265380859376, 0.079927490234375, 0.07981938934326172, 0.08007884979248046, 0.07963986968994141, 0.0802863998413086, 0.07934944152832031, 0.07852003479003906, 0.07901238250732422, 0.08063616180419922, 0.07974278259277344, 0.0799817886352539, 0.07999097442626953, 0.08056278228759765, 0.08002329254150391, 0.07961420440673828, 0.08006246185302734, 0.07993753814697266, 0.08011161804199218, 0.07995552062988281, 0.07961030578613282, 0.08038380432128907, 0.08015248107910156, 0.080193603515625, 0.07987599945068359, 0.08072978973388673, 0.0800895004272461, 0.0796810531616211, 0.07945286560058594, 0.080552001953125, 0.08072185516357422, 0.08068077087402344, 0.07938019561767579, 0.08030252838134766, 0.0801095962524414, 0.08088166046142578, 0.08081584167480468, 0.08126287841796875, 0.08030617523193359, 0.0798023681640625, 0.08090966033935547, 0.08051939392089844, 0.0799543685913086, 0.08028739166259766, 0.08052713775634765, 0.0801698226928711, 0.08027721405029296, 0.07998992156982422, 0.08052374267578125, 0.08137561798095704, 0.08071987152099609, 0.0801812515258789, 0.0808157730102539, 0.08049603271484375, 0.08115039825439453, 0.08061593627929688, 0.08083570861816407, 0.08094156646728516, 0.08095747375488281, 0.08082262420654297, 0.08344767761230469, 0.07886227416992188, 0.07932319641113281, 0.07937359619140626, 0.07923715209960938, 0.07954736328125, 0.07995597076416015, 0.07898524475097657, 0.07883158111572265, 0.07951360321044922, 0.07836662292480469, 0.0800317153930664, 0.08160883331298828, 0.08018534088134766, 0.0795340805053711, 0.07982899475097656, 0.07962009429931641, 0.07976668548583984, 0.079595458984375, 0.07957097625732422, 0.07942438507080078, 0.08001126098632813, 0.07960745239257813, 0.07999727630615235, 0.08083468627929688, 0.08038349151611328, 0.08024921417236328, 0.07963648223876953, 0.08050688171386719, 0.08024269104003906, 0.08002559661865234, 0.08040761566162109, 0.07979084777832031, 0.07981276702880859, 0.07929859161376954, 0.07993068695068359, 0.08107283020019532, 0.08031231689453125, 0.08054579162597657, 0.080574462890625, 0.08040243530273437, 0.08026290893554687, 0.0799889907836914, 0.08022332763671874, 0.07904045104980469, 0.08054064178466797, 0.0801607666015625, 0.08037907409667969, 0.08105197143554688, 0.0807060775756836, 0.08126998138427734, 0.08064911651611328, 0.0808479995727539, 0.08123056030273437, 0.08054988861083984, 0.08071920013427734, 0.08062601470947266, 0.08052767944335938, 0.08094310760498047, 0.08080121612548828, 0.08081660461425781, 0.08070767974853515, 0.08101395416259766]",tokens/s,12.4938127432862,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.747328,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3548.95872,4490.985472,0.0,4112.515072,3976.487424,s,1,10.02813671875,10.02813671875,0.0,10.02813671875,10.02813671875,10.02813671875,10.02813671875,[10.02813671875],,kWh,8.055114220416802e-05,8.878118307045344e-06,2.702307717399998e-05,0.00011645233768521334,,MB,1479.049216,4690.214912,0.0,4282.384384,4102.199808,s,10,3.114191009521484,0.3114191009521484,0.0034445911652642505,0.31207496643066407,0.3150069061279297,0.31512971954345703,0.3152279702758789,"[0.3030219421386719, 0.3110392150878906, 0.30835064697265624, 0.31066754150390624, 0.312292724609375, 0.31525253295898437, 0.3149796142578125, 0.3143934326171875, 0.31233615112304686, 0.3118572082519531]",tokens/s,822.0433467866702,kWh,8.933777769318192e-06,9.85237133470348e-07,5.937311147151524e-06,1.5856326049940063e-05,tokens/kWh,16144975.777725488,MB,1483.063296,4698.60352,0.0,4290.772992,4102.202368,s,10,20.3044150390625,2.0304415039062502,0.010348142657225981,2.028146728515625,2.040732727050781,2.0472989807128905,2.052551983642578,"[2.036552001953125, 2.0176671142578124, 2.0279625244140624, 2.017232421875, 2.0392735595703124, 2.053865234375, 2.0263463134765627, 2.03277294921875, 2.0283309326171874, 2.0244119873046875]",tokens/s,31.027734548765824,kWh,5.962421416026515e-05,6.576492029975314e-06,3.9472558514048464e-05,0.00010567326470428893,tokens/kWh,596177.2845411394,,s,630,20.301882244110097,0.032225209911285885,0.0005863946240674489,0.03213471984863281,0.032721768951416014,0.03309802722930908,0.03427442245483398,"[0.03317695999145508, 0.03250255966186524, 0.032330623626708986, 0.032279518127441405, 0.032067359924316405, 0.03225171279907227, 0.03252633666992188, 0.03202703857421875, 0.03313151931762695, 0.03214179229736328, 0.0319780158996582, 0.03199699211120605, 0.032199615478515624, 0.03259801483154297, 0.03223756790161133, 0.032925537109375, 0.03217158508300781, 0.03228927993774414, 0.033048736572265626, 0.03235763168334961, 0.03208822250366211, 0.03260662460327148, 0.032441726684570314, 0.03621468734741211, 0.032592159271240234, 0.0321921272277832, 0.0324027214050293, 0.03225360107421875, 0.03210019302368164, 0.03216534423828125, 0.03230134582519531, 0.03206329727172851, 0.032424545288085936, 0.03211859130859375, 0.03212716674804687, 0.03237580871582031, 0.03257855987548828, 0.03234815979003906, 0.032323585510253904, 0.03328799819946289, 0.03234016036987305, 0.032159648895263675, 0.03292752075195313, 0.03248569488525391, 0.031916032791137694, 0.032182174682617186, 0.03212271881103516, 0.03231769561767578, 0.0321497917175293, 0.03229872131347656, 0.03192831993103027, 0.031937824249267575, 0.032021217346191407, 0.03198975944519043, 0.031905792236328126, 0.03177267265319824, 0.03179929542541504, 0.03176217651367187, 0.03161523246765137, 0.03161868858337402, 0.031871360778808595, 0.03211264038085938, 0.03184230422973633, 0.033083393096923826, 0.032188255310058596, 0.03227481460571289, 0.03212860870361328, 0.032610462188720706, 0.03301174545288086, 0.032247135162353516, 0.03254732894897461, 0.03219039916992188, 0.03210079956054687, 0.031826719284057614, 0.03183865547180176, 0.031928895950317386, 0.0320184326171875, 0.03180953598022461, 0.031841600418090824, 0.031977888107299804, 0.03180175971984863, 0.0319071044921875, 0.03185103988647461, 0.0318253116607666, 0.0316463680267334, 0.03184025573730469, 0.031667232513427734, 0.031836544036865234, 0.03175075149536133, 0.03229254531860352, 0.03186086463928223, 0.03169468879699707, 0.03265980911254883, 0.031947839736938474, 0.03209724807739258, 0.03240745544433594, 0.032767166137695314, 0.03215039825439453, 0.03199590492248535, 0.03222528076171875, 0.031784959793090824, 0.032092159271240234, 0.03171711921691894, 0.03159612846374512, 0.031687519073486325, 0.03190303993225098, 0.03195459175109863, 0.031711679458618164, 0.03190361595153809, 0.03174774360656738, 0.0318874568939209, 0.03191801643371582, 0.03185494422912598, 0.032257728576660157, 0.032414112091064456, 0.03191654396057129, 0.032161537170410155, 0.03175420761108398, 0.03160902404785156, 0.03169020843505859, 0.03210841751098633, 0.03221372985839844, 0.03160671997070313, 0.03258134460449219, 0.03189583969116211, 0.03161497688293457, 0.032796607971191404, 0.032150592803955075, 0.031976512908935543, 0.03169593620300293, 0.03181865692138672, 0.03199935913085938, 0.03214390563964844, 0.03202899169921875, 0.03187494468688965, 0.031856191635131835, 0.03204508972167969, 0.031860960006713866, 0.03180953598022461, 0.03188121604919433, 0.031889408111572266, 0.031542560577392575, 0.031674879074096676, 0.03246716690063477, 0.03419161605834961, 0.03310316848754883, 0.0320250244140625, 0.03207372665405273, 0.0318525447845459, 0.03250697708129883, 0.032138145446777344, 0.031980607986450194, 0.0319081916809082, 0.032291038513183594, 0.03275107192993164, 0.03246992111206055, 0.0325459213256836, 0.03235017776489258, 0.032396129608154293, 0.03224787139892578, 0.03256934356689453, 0.03249382400512695, 0.032204734802246095, 0.03236355209350586, 0.032295902252197264, 0.03251385498046875, 0.03258531188964844, 0.032188961029052734, 0.032122112274169924, 0.03184614372253418, 0.03185670471191406, 0.03185724830627441, 0.031959264755249024, 0.03237811279296875, 0.03203964614868164, 0.032272289276123044, 0.03297654342651367, 0.03232611083984375, 0.03235631942749023, 0.032032798767089844, 0.03255699157714844, 0.03190496063232422, 0.031948959350585934, 0.03180604743957519, 0.0318691520690918, 0.0322784309387207, 0.03182537651062012, 0.032145503997802735, 0.03181817626953125, 0.03330774307250976, 0.03224563217163086, 0.032094112396240236, 0.03207379150390625, 0.032043041229248045, 0.03227238464355469, 0.03204095840454101, 0.03180544090270996, 0.032643070220947264, 0.031897472381591796, 0.03164521598815918, 0.03195555114746094, 0.032333824157714845, 0.03190169525146484, 0.03178291130065918, 0.03196313667297363, 0.031861055374145505, 0.031949888229370116, 0.03242015838623047, 0.03183238410949707, 0.031854591369628905, 0.03181363105773926, 0.03200115203857422, 0.03187187194824219, 0.032048255920410156, 0.032256160736083984, 0.032327392578125, 0.03210079956054687, 0.031742176055908206, 0.03218262481689453, 0.03209609603881836, 0.032133281707763674, 0.03208726501464844, 0.03187382316589355, 0.03224092864990234, 0.032217823028564456, 0.03203481674194336, 0.03296460723876953, 0.03193196868896484, 0.03262675094604492, 0.03158409690856934, 0.0317936954498291, 0.031890623092651366, 0.03195167922973633, 0.03215359878540039, 0.03174611282348633, 0.031593439102172854, 0.031976192474365235, 0.03170844841003418, 0.03211974334716797, 0.031705408096313475, 0.031758016586303714, 0.031731712341308595, 0.03189145660400391, 0.03175779151916504, 0.031838336944580076, 0.03175827217102051, 0.03170556831359863, 0.03201228713989258, 0.03190508842468262, 0.03187750434875488, 0.032073886871337894, 0.03197148895263672, 0.034414974212646485, 0.03209961700439453, 0.03227068710327148, 0.03181401634216308, 0.0318338565826416, 0.03185411262512207, 0.03180150413513184, 0.03227036666870117, 0.031742496490478514, 0.031747903823852536, 0.03166819190979004, 0.03238739013671875, 0.03210614395141602, 0.03187110328674316, 0.03195712089538574, 0.03176243209838867, 0.031854591369628905, 0.031768575668334964, 0.0318351993560791, 0.032134078979492185, 0.0321923828125, 0.03233599853515625, 0.03225600051879883, 0.03209830474853516, 0.03216793441772461, 0.03232915115356445, 0.03281379318237305, 0.032597854614257814, 0.03335567855834961, 0.03323839950561523, 0.032166622161865235, 0.032026622772216795, 0.03196723175048828, 0.03250128173828125, 0.032129566192626954, 0.0325645751953125, 0.03210419082641602, 0.03224838256835937, 0.03264131164550781, 0.0324587516784668, 0.03215510559082031, 0.032584449768066404, 0.033021728515625, 0.03223324966430664, 0.03540124893188477, 0.03249737548828125, 0.032144191741943356, 0.032362655639648436, 0.032309249877929686, 0.03208758544921875, 0.03214134216308594, 0.03236294555664063, 0.03218022537231445, 0.03259756851196289, 0.032364990234375, 0.03269836807250977, 0.032745471954345705, 0.03293183898925781, 0.03223961639404297, 0.032059391021728514, 0.032960479736328124, 0.03245043182373047, 0.03309174346923828, 0.03318374252319336, 0.032395263671875, 0.03222492980957031, 0.032296768188476564, 0.032182304382324216, 0.03233804702758789, 0.03209616088867188, 0.03226451110839844, 0.032239070892333986, 0.03210710525512695, 0.032091232299804685, 0.03220377731323242, 0.03222700881958008, 0.032393535614013674, 0.03259580612182617, 0.0326262092590332, 0.03230502319335937, 0.03286851119995117, 0.0331060791015625, 0.03197996711730957, 0.03200204849243164, 0.03187507247924805, 0.032552383422851563, 0.031911808013916014, 0.04034431838989258, 0.0338328971862793, 0.0324048957824707, 0.03284643173217774, 0.03221913528442383, 0.03217203140258789, 0.03241510391235351, 0.03258000183105469, 0.03238729476928711, 0.03256707382202149, 0.03251228713989258, 0.03221907043457031, 0.03234729766845703, 0.032365409851074216, 0.0323743667602539, 0.032362911224365236, 0.03229695892333984, 0.032401504516601565, 0.03368540954589844, 0.03235747146606445, 0.03236547088623047, 0.03213459014892578, 0.03344870376586914, 0.03210425567626953, 0.03297654342651367, 0.032465248107910155, 0.03231916809082031, 0.03246432113647461, 0.032205696105957034, 0.03266672134399414, 0.03229776000976563, 0.03250790405273438, 0.03237081527709961, 0.03205120086669922, 0.03256524658203125, 0.032194561004638675, 0.03249942398071289, 0.034259201049804684, 0.03293135833740234, 0.03337033462524414, 0.03235430526733398, 0.03192118453979492, 0.03196217536926269, 0.03239926528930664, 0.03174947166442871, 0.033391265869140624, 0.03174399948120117, 0.03178895950317383, 0.031856735229492186, 0.03196284866333008, 0.03271913528442383, 0.032231422424316404, 0.03199795150756836, 0.03195011138916016, 0.03180364799499512, 0.03230767822265625, 0.03194684791564941, 0.03189545631408691, 0.031741151809692385, 0.031912736892700196, 0.03177824020385742, 0.032134849548339846, 0.03190787124633789, 0.032045921325683596, 0.032004096984863284, 0.031997055053710935, 0.03233804702758789, 0.032275264739990234, 0.032008129119873045, 0.031882591247558593, 0.0323809928894043, 0.031871807098388674, 0.0319465274810791, 0.03239833450317383, 0.03280003356933594, 0.03298070526123047, 0.034484222412109376, 0.032355712890625, 0.03177078437805176, 0.03199231910705566, 0.03176806449890137, 0.032160224914550783, 0.031926271438598636, 0.03179110336303711, 0.03200204849243164, 0.03290521621704102, 0.03264851379394531, 0.03182076835632324, 0.031844032287597655, 0.031817760467529294, 0.03185663986206055, 0.03197686386108398, 0.03219055938720703, 0.032233951568603515, 0.032001121520996094, 0.032152511596679687, 0.03177638435363769, 0.03188307189941406, 0.032012863159179686, 0.03195289611816406, 0.03244236755371094, 0.032556671142578125, 0.033536094665527344, 0.03229673767089844, 0.03229708862304687, 0.03216505432128906, 0.03214815902709961, 0.03254185485839844, 0.03211363220214844, 0.03217510223388672, 0.032294975280761716, 0.03239724731445313, 0.032254974365234376, 0.03211027145385742, 0.032260414123535155, 0.032259265899658204, 0.03230998229980469, 0.03237283325195312, 0.03232489776611328, 0.032293472290039066, 0.03215776062011719, 0.03182143974304199, 0.031816127777099606, 0.03165763282775879, 0.03164505577087402, 0.031492576599121094, 0.03264665603637695, 0.031648128509521485, 0.031634464263916015, 0.033498592376708984, 0.031782432556152346, 0.03178303909301758, 0.03247561645507813, 0.032287967681884765, 0.03192697525024414, 0.03233171081542969, 0.03217808151245117, 0.03219276809692383, 0.03222323226928711, 0.031989664077758787, 0.031924320220947267, 0.032296382904052734, 0.03236726379394531, 0.03247504043579102, 0.03244963073730469, 0.03223030471801758, 0.032200702667236326, 0.032086017608642575, 0.03202592086791992, 0.03206009674072265, 0.032199905395507815, 0.03239811325073242, 0.032176128387451174, 0.03227974319458008, 0.03242681503295899, 0.032217086791992186, 0.032036865234375, 0.032573440551757815, 0.03307724761962891, 0.03208396911621094, 0.03301769638061523, 0.03226959991455078, 0.032219711303710936, 0.03266387176513672, 0.03340697479248047, 0.03315657424926758, 0.03227088165283203, 0.03199795150756836, 0.03213721466064453, 0.032268032073974606, 0.03200864028930664, 0.031815103530883786, 0.031842336654663086, 0.03171504020690918, 0.03163030433654785, 0.03177807998657227, 0.03176489639282227, 0.03359052658081055, 0.03338927841186524, 0.0348012809753418, 0.03322675323486328, 0.03290969467163086, 0.03198975944519043, 0.03168460845947266, 0.03172966384887695, 0.032284671783447266, 0.03181270408630371, 0.03163843154907227, 0.03155558395385742, 0.031669727325439455, 0.03161961555480957, 0.03207372665405273, 0.03191967964172363, 0.03214748764038086, 0.032543136596679685, 0.03221452713012695, 0.03208000183105469, 0.03215603256225586, 0.032083358764648434, 0.031897727966308596, 0.031889888763427736, 0.032376895904541014, 0.032122814178466796, 0.032745471954345705, 0.03179929542541504, 0.03171468734741211, 0.031840288162231445, 0.03216239929199219, 0.032527584075927735, 0.03213593673706055, 0.03271273422241211, 0.03215955352783203, 0.03285932922363281, 0.03210982513427734, 0.031981311798095706, 0.03265740966796875, 0.032061279296875, 0.03185193634033203, 0.03182284736633301, 0.03192940711975098, 0.03203750228881836, 0.03211420822143555, 0.03230569458007813, 0.031678464889526366, 0.03188649559020996, 0.03241660690307617, 0.0325153923034668, 0.032244415283203126, 0.033050495147705077, 0.03215577697753906, 0.031876575469970705, 0.03210089492797852, 0.031794240951538086, 0.0319804801940918, 0.033145984649658206, 0.032107391357421876, 0.032543041229248046, 0.03198755264282226, 0.032165569305419923, 0.03186233520507813, 0.03170569610595703, 0.03162931251525879, 0.031568159103393556, 0.03217567825317383, 0.03174991989135742, 0.03175590324401856, 0.03186764717102051, 0.03197689628601074, 0.03190822410583496, 0.03178463935852051, 0.03178771209716797, 0.03176540756225586, 0.0318940486907959, 0.031940895080566405, 0.03187420845031738, 0.031929471969604495, 0.03184406471252441, 0.03225001525878906, 0.03227571105957031, 0.03224825668334961, 0.032258304595947265, 0.032166942596435544, 0.032279518127441405, 0.032143360137939454, 0.0321976318359375, 0.03300396728515625, 0.032195137023925784, 0.032046207427978514, 0.03208262252807617, 0.0325032958984375, 0.03200921630859375, 0.0342806396484375, 0.03351366424560547, 0.032511905670166014, 0.03291587066650391, 0.03236249542236328, 0.03216313552856445, 0.032184894561767576, 0.03207952117919922, 0.031895519256591794, 0.03205916976928711, 0.03208879852294922, 0.03183318328857422, 0.031675296783447264, 0.0318791675567627, 0.03170918464660644, 0.031901792526245115, 0.03171120071411133, 0.03180867195129394, 0.03200284957885742, 0.032000286102294925]",tokens/s,31.031605465191426,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4908.900352,7227.768832,0.0,6849.298432,6445.09696,s,1,11.409763671875,11.409763671875,0.0,11.409763671875,11.409763671875,11.409763671875,11.409763671875,[11.409763671875],,kWh,0.00012727060596250036,1.4031577753581444e-05,4.089781049598784e-05,0.00018219999421206963,,MB,2795.634688,7580.090368,0.0,7172.25984,6823.319552,s,10,3.659079406738282,0.3659079406738281,0.0022586638860903206,0.3660846557617188,0.36860031433105467,0.36880082855224605,0.3689612399291992,"[0.36158599853515627, 0.36374530029296875, 0.36855264282226563, 0.36434072875976564, 0.3690013427734375, 0.36476950073242187, 0.36626763916015626, 0.36590167236328125, 0.36855575561523435, 0.36635882568359374]",tokens/s,699.6295284780373,kWh,1.0731959706250413e-05,1.1835455571471207e-06,7.125333081214087e-06,1.904083834461162e-05,tokens/kWh,13444786.167855136,MB,2799.960064,7582.18752,0.0,7174.356992,6823.322112,s,10,22.737356689453122,2.2737356689453123,0.010605739310478824,2.2769017333984376,2.285530078125,2.286243310546875,2.286813896484375,"[2.2818232421875, 2.2616044921875, 2.28537158203125, 2.2813994140625, 2.2808974609375, 2.272906005859375, 2.25453369140625, 2.266579345703125, 2.265284912109375, 2.28695654296875]",tokens/s,27.70770624767609,kWh,6.630275453333297e-05,7.313228889816367e-06,4.402154116958458e-05,0.00011763752459273392,tokens/kWh,535543.4009522783,,s,630,22.73450096893311,0.03608650947449699,0.0005440227987745136,0.0360599365234375,0.03659966354370117,0.03701184482574463,0.037776075401306154,"[0.03709203338623047, 0.036034656524658204, 0.037209697723388675, 0.035833889007568356, 0.03566934585571289, 0.03578083038330078, 0.035856895446777344, 0.035621086120605466, 0.03602342224121094, 0.036013057708740234, 0.036138561248779295, 0.03601980972290039, 0.03599955368041992, 0.03596515274047852, 0.03611923217773438, 0.03609363174438476, 0.03618643188476563, 0.03596063995361328, 0.03614944076538086, 0.03641683197021484, 0.03616844940185547, 0.03617536163330078, 0.03633379364013672, 0.03605126571655273, 0.03628636932373047, 0.03619811248779297, 0.04121209716796875, 0.03624560165405273, 0.03650358581542969, 0.03624752044677734, 0.036413440704345705, 0.03710534286499023, 0.03643423843383789, 0.036036094665527346, 0.03605465698242188, 0.036045696258544924, 0.03602227020263672, 0.03609190368652344, 0.03579859161376953, 0.03606572723388672, 0.035917823791503906, 0.03592806243896484, 0.03580710220336914, 0.03592998504638672, 0.03586073684692383, 0.03582156753540039, 0.035999744415283204, 0.03575798416137695, 0.03582326507568359, 0.03583430480957031, 0.035860481262207033, 0.03798406219482422, 0.0359730224609375, 0.03593040084838867, 0.03573964691162109, 0.03564252853393555, 0.03588131332397461, 0.036995582580566407, 0.03616153717041016, 0.03616563034057617, 0.035995582580566406, 0.03667542266845703, 0.03614944076538086, 0.036913150787353514, 0.03591535949707031, 0.035668384552001955, 0.0357534065246582, 0.03563167953491211, 0.03547955322265625, 0.03520512008666992, 0.03529840087890625, 0.03516044616699219, 0.03530931091308594, 0.03499702453613281, 0.035198974609375, 0.03678822326660156, 0.03577446365356445, 0.03653836822509766, 0.035429824829101564, 0.03512982559204102, 0.03528230285644531, 0.03536764907836914, 0.035526336669921874, 0.03558636856079102, 0.03582361602783203, 0.03585228729248047, 0.03733299255371094, 0.03596060943603516, 0.03594982528686524, 0.035892192840576174, 0.03590553665161133, 0.0360274543762207, 0.036082271575927735, 0.035768062591552734, 0.03606147384643555, 0.03650387191772461, 0.03580928039550781, 0.035624801635742186, 0.035616928100585935, 0.03537638473510742, 0.03584022521972656, 0.035737438201904295, 0.03588140869140625, 0.036972862243652344, 0.03587990570068359, 0.03556844711303711, 0.035643550872802736, 0.03546931076049804, 0.035471359252929685, 0.03564652633666992, 0.03585094451904297, 0.03597049713134766, 0.0359958724975586, 0.036011871337890626, 0.03686617660522461, 0.036862049102783206, 0.03618828964233398, 0.03604931259155274, 0.0364071044921875, 0.03704393768310547, 0.036676063537597656, 0.03635740661621094, 0.03597110366821289, 0.03577468872070313, 0.036097824096679686, 0.035576351165771486, 0.036792320251464845, 0.0359444465637207, 0.036171009063720706, 0.03598947143554688, 0.03603263854980469, 0.03600352096557617, 0.036010303497314454, 0.036012096405029295, 0.036030719757080075, 0.03603286361694336, 0.0362105598449707, 0.035870849609375, 0.036018142700195314, 0.03621481704711914, 0.03606937789916992, 0.03665827178955078, 0.036152225494384765, 0.035862529754638675, 0.036116481781005856, 0.036060672760009765, 0.03606147384643555, 0.036116481781005856, 0.036065502166748045, 0.03644416046142578, 0.03656614303588867, 0.03629555130004883, 0.03632271957397461, 0.036491073608398435, 0.036391712188720705, 0.03628646469116211, 0.036245502471923825, 0.036211902618408204, 0.03666566467285156, 0.03613478469848633, 0.03616758346557617, 0.03645718383789062, 0.0361267204284668, 0.03630284881591797, 0.036337535858154295, 0.03651164627075195, 0.036230815887451175, 0.03620512008666992, 0.036124671936035156, 0.03657065582275391, 0.036388832092285155, 0.03606988906860351, 0.03642777633666992, 0.03636553573608398, 0.03612137603759766, 0.0362762222290039, 0.03721011352539062, 0.036796001434326174, 0.03633808135986328, 0.035935520172119144, 0.03595747375488281, 0.036359424591064456, 0.036743934631347654, 0.03669343948364258, 0.03620512008666992, 0.03607548904418945, 0.03627775955200195, 0.03663299179077149, 0.03660403060913086, 0.03693008041381836, 0.03660620880126953, 0.03591987228393555, 0.03624345779418945, 0.03624342346191406, 0.037142559051513674, 0.03628188705444336, 0.03593286514282226, 0.03589836883544922, 0.03597974395751953, 0.0359466552734375, 0.03589487838745117, 0.03591635131835937, 0.035923168182373046, 0.03612956619262695, 0.03619839859008789, 0.036206592559814454, 0.03608607864379883, 0.036130207061767575, 0.03625603103637695, 0.03659964752197266, 0.03637673568725586, 0.03623731231689453, 0.03606512069702148, 0.036155200958251955, 0.03620451354980469, 0.03693606567382812, 0.03640524673461914, 0.036286495208740235, 0.036410526275634764, 0.036057918548583985, 0.03616719818115234, 0.037292160034179685, 0.03683347320556641, 0.03691535949707031, 0.036130847930908205, 0.03608329772949219, 0.036120769500732425, 0.03675689697265625, 0.03616556930541992, 0.0361767349243164, 0.036095455169677736, 0.0361313591003418, 0.03625094223022461, 0.03617657470703125, 0.036181182861328126, 0.03589321517944336, 0.03581564712524414, 0.03581811141967774, 0.03583590316772461, 0.03574179077148437, 0.0357243537902832, 0.03568931198120117, 0.03561471939086914, 0.03600998306274414, 0.03595446395874023, 0.03603068923950195, 0.036038078308105466, 0.0359931526184082, 0.03610287857055664, 0.03705670547485351, 0.03623331069946289, 0.03647283172607422, 0.037802272796630856, 0.03569868850708008, 0.035345760345458985, 0.03525823974609375, 0.03548179244995117, 0.03504803085327148, 0.035110912322998046, 0.035381248474121094, 0.035925312042236326, 0.03594720077514649, 0.03581324768066406, 0.03901225662231445, 0.03611388778686524, 0.03648179244995117, 0.03589734268188476, 0.037344417572021484, 0.036337982177734374, 0.035829761505126956, 0.035568126678466795, 0.035649471282958985, 0.03573884963989258, 0.035741985321044924, 0.035855968475341796, 0.03575091171264649, 0.03591167831420899, 0.03602022552490235, 0.03583523178100586, 0.035764190673828126, 0.035675872802734376, 0.03592607879638672, 0.03623209762573242, 0.036034561157226565, 0.0361082878112793, 0.03613491058349609, 0.03617507171630859, 0.03742166519165039, 0.037626049041748044, 0.03642572784423828, 0.03611238479614258, 0.037064704895019535, 0.036268001556396486, 0.0362209587097168, 0.03654764938354492, 0.03633452987670899, 0.0365219841003418, 0.03622092819213867, 0.036157440185546875, 0.036239360809326174, 0.03635200119018555, 0.03618406295776367, 0.0362762222290039, 0.03632691192626953, 0.036146751403808595, 0.03649631881713867, 0.03652188873291016, 0.03625353622436524, 0.03637392044067383, 0.03645731353759766, 0.03659980773925781, 0.0367507209777832, 0.03624931335449219, 0.0363076171875, 0.0361607666015625, 0.037435615539550784, 0.03634995269775391, 0.0363680305480957, 0.03589779281616211, 0.03618201446533203, 0.036138721466064457, 0.03764048004150391, 0.036139007568359374, 0.036050079345703125, 0.03622515106201172, 0.036088542938232424, 0.03631289672851563, 0.03617302322387695, 0.03628950500488281, 0.03634175872802734, 0.03653222274780273, 0.036169887542724606, 0.03592380905151367, 0.03625881576538086, 0.03619900894165039, 0.03607913589477539, 0.035922817230224606, 0.03604003143310547, 0.035568286895751956, 0.03575404739379883, 0.036931617736816406, 0.03568947219848633, 0.03573648071289062, 0.036000961303710936, 0.03589580917358398, 0.03561094284057617, 0.03559999847412109, 0.03551884841918945, 0.035655040740966794, 0.03567270278930664, 0.03828879928588867, 0.03646752166748047, 0.036138687133789066, 0.03618560028076172, 0.03652608108520508, 0.0363587532043457, 0.0362534065246582, 0.036020511627197264, 0.036015296936035154, 0.03621561431884766, 0.03621046447753906, 0.03631875228881836, 0.037183647155761716, 0.0361416015625, 0.036116481781005856, 0.03564492797851562, 0.03540428924560547, 0.035714080810546875, 0.03597001647949219, 0.035198974609375, 0.03509622573852539, 0.03495971298217773, 0.03528470230102539, 0.03542454528808594, 0.03562905502319336, 0.035827583312988284, 0.03577471923828125, 0.03587443161010742, 0.03678345489501953, 0.036249343872070315, 0.036008544921875, 0.03557164764404297, 0.03618422317504883, 0.035918048858642575, 0.03582361602783203, 0.03632537460327148, 0.03538534545898438, 0.035236064910888674, 0.03521923065185547, 0.03552665710449219, 0.03528230285644531, 0.035232383728027346, 0.03554508972167969, 0.035590145111083986, 0.0352911376953125, 0.035073375701904295, 0.035248287200927736, 0.03529292678833008, 0.03513353729248047, 0.0351580810546875, 0.035106975555419924, 0.03498630523681641, 0.03547148895263672, 0.035248126983642575, 0.03517804718017578, 0.03528489685058594, 0.03531011199951172, 0.03631718444824219, 0.035565055847167966, 0.03506227111816406, 0.035915775299072264, 0.03558399963378906, 0.035573760986328126, 0.035676158905029294, 0.03727974319458008, 0.035844097137451174, 0.03581756973266602, 0.0359463996887207, 0.036134048461914064, 0.03593638229370117, 0.03556662368774414, 0.03569836807250976, 0.0361135368347168, 0.035857280731201174, 0.035672065734863284, 0.035483264923095705, 0.035985984802246095, 0.03581113433837891, 0.036152576446533205, 0.03613977432250977, 0.03632032012939453, 0.03621984100341797, 0.03616563034057617, 0.03624755096435547, 0.03629590225219727, 0.03614799880981445, 0.03640060806274414, 0.0363438720703125, 0.036446689605712894, 0.03656499099731445, 0.03628236770629883, 0.03671878433227539, 0.036030529022216796, 0.03604729461669922, 0.03612998580932617, 0.03595481491088867, 0.03582012939453125, 0.03568230438232422, 0.035845375061035155, 0.03566796875, 0.03585065460205078, 0.03599814224243164, 0.03595481491088867, 0.03644607925415039, 0.03610214233398437, 0.03637859344482422, 0.03622006225585937, 0.035971969604492185, 0.036311038970947264, 0.03740467071533203, 0.036275489807128906, 0.03595484924316406, 0.03609657669067383, 0.03618201446533203, 0.03609743881225586, 0.03613763046264649, 0.0363553581237793, 0.03617244720458984, 0.036470081329345705, 0.03628851318359375, 0.036272830963134765, 0.03629260635375976, 0.03617744064331055, 0.036085376739501955, 0.03604268646240234, 0.036141983032226564, 0.03603251266479492, 0.0359279670715332, 0.035848289489746096, 0.03588710403442383, 0.03592335891723633, 0.03614371109008789, 0.035901439666748046, 0.03640524673461914, 0.03611033630371094, 0.03557580947875977, 0.03573331069946289, 0.036869598388671876, 0.035668704986572264, 0.03552665710449219, 0.03526860809326172, 0.03521900939941406, 0.035622432708740236, 0.03545382308959961, 0.035426334381103514, 0.035575614929199216, 0.03590572738647461, 0.035784702301025394, 0.03565468978881836, 0.03597369766235352, 0.03514409637451172, 0.035643390655517575, 0.03524198532104492, 0.03530752182006836, 0.036947967529296875, 0.03725107192993164, 0.03591987228393555, 0.0356228141784668, 0.0352891845703125, 0.035280895233154294, 0.03554099273681641, 0.03546112060546875, 0.035638816833496095, 0.035781089782714846, 0.035683551788330076, 0.035525409698486325, 0.03562496185302735, 0.03559404754638672, 0.035683807373046876, 0.03575267028808594, 0.035465217590332034, 0.035778560638427735, 0.035527870178222655, 0.03540636825561523, 0.03625807952880859, 0.03702515029907227, 0.035709568023681644, 0.03589120101928711, 0.03597075271606445, 0.036299072265625, 0.03598950576782227, 0.03623526382446289, 0.036150398254394533, 0.037157760620117185, 0.03619232177734375, 0.03617990493774414, 0.03615119934082031, 0.03627999877929688, 0.036159328460693356, 0.03622355270385742, 0.03634969711303711, 0.03615356826782227, 0.035899425506591795, 0.0359683837890625, 0.03596700668334961, 0.03584880065917969, 0.03602841567993164, 0.03584012985229492, 0.036114303588867185, 0.036058433532714845, 0.03581411361694336, 0.03579391860961914, 0.03573398590087891, 0.03594211196899414, 0.03590572738647461, 0.03623596954345703, 0.036016033172607424, 0.03602227020263672, 0.03557539367675781, 0.0357699203491211, 0.03674508666992188, 0.03595280075073242, 0.0356583366394043, 0.035752159118652344, 0.03550163269042969, 0.03583635330200195, 0.03589120101928711, 0.037085182189941404, 0.03614720153808594, 0.03605920028686523, 0.03570393753051758, 0.03571203231811523, 0.035858207702636716, 0.03611443328857422, 0.03596409606933594, 0.03633644866943359, 0.03580464172363281, 0.03590607833862305, 0.03593625640869141, 0.03611852645874023, 0.03612876892089844, 0.03623715209960938, 0.03638211059570313, 0.03624342346191406, 0.03606585693359375, 0.036454624176025394, 0.03620272064208984, 0.03789596939086914, 0.03758451080322266, 0.03643840026855469, 0.036240447998046876, 0.03660076904296875, 0.03630825424194336, 0.036100833892822266, 0.037154815673828126, 0.03607500839233398, 0.03621120071411133, 0.03605299377441406, 0.035845951080322264, 0.03640729522705078, 0.037711936950683596, 0.03617103958129883, 0.03623107147216797, 0.035902271270751955, 0.03606902313232422, 0.036305374145507815, 0.0363765754699707, 0.03975971221923828, 0.0364156494140625, 0.036018177032470705, 0.03605641555786133, 0.03594307327270508, 0.03616563034057617, 0.03593830490112305, 0.03585228729248047, 0.03583724975585938, 0.03596563339233398, 0.036083713531494144, 0.03598649597167969, 0.036442527770996096, 0.036065055847167966, 0.03621555328369141, 0.036364288330078126, 0.03607878494262695, 0.03616032028198242, 0.036245502471923825, 0.036106239318847655, 0.03618611145019531, 0.036308799743652344, 0.03630688095092773]",tokens/s,27.71118666122914,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,889.155584,638.517248,0.0,260.046848,253.883392,s,1,7.8444189453125,7.8444189453125,0.0,7.8444189453125,7.8444189453125,7.8444189453125,7.8444189453125,[7.8444189453125],,kWh,2.212889393751235e-05,2.432461837708945e-06,7.532228248008366e-06,3.209358402322966e-05,,MB,1186.328576,743.374848,0.0,335.54432,313.01632,s,13,0.19102784061431888,0.014694449278024527,7.09708829283819e-05,0.014706656455993653,0.014779667282104492,0.014788204765319824,0.01479198314666748,"[0.014605024337768554, 0.014754176139831543, 0.014706656455993653, 0.014738335609436035, 0.014622112274169922, 0.014633824348449707, 0.014693984031677246, 0.014758111953735352, 0.014735136032104492, 0.014792927742004395, 0.014785056114196777, 0.014570079803466796, 0.014632415771484376]",tokens/s,17421.544363887573,kWh,4.353322681321532e-07,4.800946855380307e-08,2.878500398510174e-07,7.711917765369738e-07,tokens/kWh,331953747.15944785,MB,1220.120576,768.540672,0.0,360.710144,313.805312,s,13,10.196140502929687,0.7843185002253606,0.005833425561564566,0.7838369140625,0.7906074829101563,0.7941986938476563,0.7972828051757812,"[0.7829224853515625, 0.7863546142578125, 0.7857490234375, 0.7838369140625, 0.7803174438476562, 0.7789570922851563, 0.7865230102539063, 0.7847035522460938, 0.7980538330078125, 0.7806293334960938, 0.7916286010742187, 0.7730145263671875, 0.7834500732421875]",tokens/s,80.32451100145924,kWh,2.23064743530215e-05,2.4600092574525966e-06,8.558268148147221e-06,3.332475175862131e-05,tokens/kWh,1890486.700585895,,s,819,10.188614595413219,0.012440310861310389,0.0002490865052194593,0.012426912307739258,0.012600223541259765,0.01267319393157959,0.01329150056838989,"[0.012099583625793458, 0.01241494369506836, 0.012588512420654297, 0.01240112018585205, 0.012310943603515624, 0.012381152153015136, 0.012483424186706542, 0.012439423561096192, 0.012439807891845704, 0.012334848403930664, 0.012407936096191407, 0.012344096183776855, 0.012552288055419922, 0.01252780818939209, 0.012666751861572265, 0.012570240020751954, 0.012529984474182129, 0.012507200241088867, 0.012444640159606933, 0.012498208045959473, 0.012359423637390137, 0.01234652805328369, 0.01237276840209961, 0.01233340835571289, 0.012414688110351563, 0.012406720161437988, 0.01236780834197998, 0.01242732810974121, 0.012629088401794433, 0.012823455810546875, 0.012461855888366699, 0.012508416175842285, 0.012266464233398437, 0.012304384231567383, 0.012249088287353516, 0.012166912078857423, 0.012212479591369629, 0.012175552368164063, 0.012170495986938477, 0.012187711715698242, 0.01208198356628418, 0.012168831825256347, 0.012191807746887207, 0.01236787223815918, 0.012310432434082032, 0.012431679725646972, 0.012441375732421874, 0.012605440139770508, 0.012555392265319824, 0.012467071533203125, 0.012541312217712403, 0.012507519721984862, 0.012536064147949218, 0.012517375946044922, 0.012529791831970215, 0.012523391723632812, 0.012505248069763183, 0.01250499153137207, 0.012520928382873536, 0.012511712074279785, 0.012472319602966308, 0.012472064018249511, 0.012431679725646972, 0.012390399932861328, 0.012529855728149413, 0.012481951713562011, 0.012566944122314454, 0.012478816032409667, 0.012492032051086425, 0.01244156837463379, 0.012364224433898926, 0.012445695877075195, 0.012496735572814941, 0.012365983963012696, 0.01239033603668213, 0.01242527961730957, 0.012502240180969239, 0.012512096405029297, 0.012527551651000977, 0.012429311752319335, 0.0123406400680542, 0.012409440040588379, 0.012367327690124512, 0.012310111999511719, 0.01241817569732666, 0.012514495849609375, 0.01259993553161621, 0.012489855766296387, 0.012544672012329101, 0.012456159591674804, 0.012489024162292481, 0.01248633575439453, 0.012490976333618163, 0.012444576263427735, 0.012442463874816895, 0.012496928215026855, 0.012484543800354004, 0.012482687950134278, 0.012469792366027832, 0.012468511581420899, 0.012533535957336426, 0.012495360374450683, 0.012520383834838868, 0.01256105613708496, 0.01261184024810791, 0.01245798397064209, 0.012478464126586914, 0.012468544006347657, 0.01244745635986328, 0.01242732810974121, 0.012455840110778809, 0.012381855964660644, 0.01240665626525879, 0.012422847747802734, 0.012542752265930176, 0.012470272064208985, 0.012445695877075195, 0.01244159984588623, 0.012498559951782227, 0.012544384002685547, 0.01249075222015381, 0.012515328407287597, 0.01247606372833252, 0.012673279762268067, 0.012562527656555175, 0.012398400306701661, 0.012230976104736328, 0.012460703849792481, 0.012439711570739746, 0.012395359992980958, 0.012407296180725098, 0.012433952331542968, 0.012308320045471192, 0.012740768432617188, 0.01238422393798828, 0.012326656341552734, 0.012394623756408691, 0.01235472011566162, 0.0123951997756958, 0.012480799674987792, 0.012590399742126465, 0.012722880363464355, 0.012565759658813477, 0.012407551765441895, 0.012391872406005859, 0.012313376426696778, 0.012182751655578613, 0.012098112106323243, 0.012120063781738282, 0.012134400367736817, 0.012152735710144042, 0.012134783744812012, 0.012247808456420898, 0.01237110424041748, 0.012313535690307617, 0.012350208282470703, 0.012406911849975586, 0.012391712188720703, 0.012362048149108887, 0.012583328247070312, 0.01244979190826416, 0.012582912445068359, 0.012668095588684081, 0.012554080009460448, 0.012534432411193848, 0.012550144195556641, 0.01254646396636963, 0.012550016403198243, 0.01254415988922119, 0.012543999671936035, 0.012529279708862305, 0.012517631530761719, 0.012543359756469726, 0.012670751571655273, 0.012612447738647461, 0.012627903938293456, 0.012521087646484375, 0.012552831649780273, 0.012588479995727538, 0.012580479621887207, 0.012519519805908203, 0.012563103675842286, 0.012543999671936035, 0.012580320358276368, 0.012728863716125489, 0.01263526439666748, 0.01262281608581543, 0.012574432373046875, 0.012531423568725586, 0.012290047645568849, 0.012462176322937012, 0.012463135719299316, 0.012468544006347657, 0.012445695877075195, 0.012425791740417481, 0.01245798397064209, 0.012533344268798829, 0.012491456031799316, 0.012449248313903809, 0.012468480110168456, 0.012505311965942383, 0.012500736236572265, 0.012574111938476563, 0.012548288345336915, 0.012580927848815917, 0.012502816200256348, 0.012482879638671875, 0.012454400062561035, 0.012465951919555664, 0.012504704475402832, 0.012545760154724122, 0.012546815872192384, 0.012511136054992676, 0.01246003246307373, 0.012453696250915528, 0.012533599853515626, 0.012517727851867676, 0.012415328025817872, 0.012461536407470703, 0.012581055641174316, 0.012601696014404296, 0.012569503784179687, 0.012527520179748536, 0.012497759819030762, 0.012456128120422363, 0.012453408241271972, 0.01233948802947998, 0.012423168182373047, 0.012456959724426269, 0.012368895530700684, 0.012313599586486817, 0.012312864303588867, 0.012772064208984375, 0.01235968017578125, 0.012310367584228515, 0.01225772762298584, 0.012312288284301759, 0.01225715160369873, 0.012304544448852539, 0.01226956844329834, 0.012247039794921874, 0.012254816055297851, 0.012303775787353515, 0.012583904266357422, 0.01236137580871582, 0.012314656257629394, 0.012402688026428223, 0.012337471961975098, 0.01230355167388916, 0.012245823860168457, 0.012369248390197754, 0.012298912048339843, 0.012154879570007325, 0.012371456146240235, 0.012446208000183106, 0.012318943977355957, 0.012351455688476563, 0.012325984001159668, 0.012347871780395508, 0.012255231857299804, 0.012279104232788086, 0.012281824111938476, 0.01230742359161377, 0.012263327598571778, 0.012251456260681153, 0.01233897590637207, 0.012322112083435059, 0.01243616008758545, 0.012281344413757325, 0.012274080276489258, 0.012236991882324218, 0.012179360389709473, 0.012253184318542481, 0.012269760131835937, 0.012254591941833495, 0.01227235221862793, 0.012291808128356934, 0.012300288200378418, 0.012324864387512208, 0.012312576293945313, 0.012230815887451172, 0.012326751708984375, 0.01255123233795166, 0.012458304405212402, 0.012412832260131837, 0.012317407608032226, 0.01236086368560791, 0.012335136413574218, 0.012321599960327148, 0.012401887893676758, 0.01236780834197998, 0.012306943893432617, 0.012269920349121093, 0.012339263916015626, 0.0124203519821167, 0.012366527557373047, 0.012378175735473633, 0.012380000114440917, 0.01230038356781006, 0.012783904075622559, 0.013473504066467286, 0.012815520286560059, 0.012485471725463868, 0.012489855766296387, 0.012450688362121582, 0.012441472053527832, 0.012368127822875977, 0.012377504348754884, 0.012380640029907226, 0.0123919038772583, 0.012488608360290527, 0.012423808097839356, 0.012396063804626465, 0.012465760231018067, 0.012370816230773926, 0.012053279876708985, 0.012380160331726075, 0.012476544380187988, 0.012263296127319336, 0.012367712020874024, 0.012306591987609864, 0.012342464447021485, 0.012546879768371582, 0.012445695877075195, 0.012391712188720703, 0.01235632038116455, 0.012451711654663087, 0.012365951538085937, 0.012396063804626465, 0.012357407569885254, 0.012417728424072266, 0.012414560317993165, 0.012222080230712891, 0.01216716766357422, 0.01214684772491455, 0.012087936401367188, 0.012084608078002929, 0.012038687705993652, 0.012152928352355957, 0.012201984405517579, 0.012240799903869629, 0.012249183654785157, 0.012352800369262696, 0.012343199729919433, 0.012421600341796875, 0.012239199638366699, 0.01225113582611084, 0.012237024307250977, 0.012164959907531738, 0.012828607559204102, 0.01220201587677002, 0.013653984069824219, 0.012327232360839845, 0.012299936294555664, 0.01226137638092041, 0.012282208442687988, 0.012515007972717285, 0.01236950397491455, 0.012324352264404297, 0.012329888343811036, 0.012511232376098632, 0.012644319534301758, 0.012396575927734375, 0.012443167686462402, 0.012503520011901855, 0.012470272064208985, 0.012422528266906738, 0.012432000160217286, 0.012402688026428223, 0.012361023902893066, 0.012348095893859863, 0.012371968269348145, 0.012283231735229492, 0.012260000228881835, 0.012353535652160644, 0.012332159996032715, 0.012218496322631837, 0.012311200141906738, 0.012294207572937012, 0.01248249626159668, 0.012377984046936035, 0.012375328063964844, 0.012345600128173827, 0.012415264129638672, 0.012390560150146484, 0.012357664108276368, 0.012365951538085937, 0.012346559524536133, 0.012417856216430665, 0.012467424392700195, 0.012437727928161621, 0.012438079833984376, 0.012486592292785644, 0.012488127708435059, 0.012420991897583007, 0.012487423896789552, 0.012530943870544434, 0.012603551864624023, 0.012525376319885253, 0.012520319938659668, 0.012822431564331055, 0.013252927780151367, 0.012793536186218262, 0.012637663841247558, 0.012466719627380371, 0.01246617603302002, 0.012523520469665527, 0.01253923225402832, 0.012450400352478028, 0.012503104209899902, 0.01255628776550293, 0.012471327781677246, 0.012603424072265624, 0.012560704231262206, 0.012581279754638672, 0.012495136260986328, 0.012515487670898438, 0.01254537582397461, 0.012515775680541992, 0.012480511665344238, 0.012382207870483398, 0.01248025608062744, 0.01241097640991211, 0.012423328399658203, 0.012414976119995117, 0.01237603187561035, 0.01243881607055664, 0.012401408195495605, 0.012461695671081543, 0.012367263793945312, 0.012380864143371582, 0.012371935844421388, 0.01242131233215332, 0.012343423843383789, 0.01236355209350586, 0.012341471672058105, 0.012365823745727538, 0.012381600379943849, 0.012372384071350098, 0.012400832176208496, 0.012638208389282226, 0.012328991889953613, 0.012606975555419921, 0.012609600067138672, 0.01240278434753418, 0.012382559776306153, 0.012357439994812012, 0.012363743782043458, 0.0124169921875, 0.012320799827575684, 0.012371552467346192, 0.012364640235900878, 0.012378047943115235, 0.012427103996276856, 0.012412927627563476, 0.012445311546325683, 0.012478848457336426, 0.012472448348999023, 0.01239641571044922, 0.012421119689941406, 0.012338848114013671, 0.012452192306518554, 0.012620927810668946, 0.012510080337524414, 0.012546079635620117, 0.012572640419006348, 0.012558303833007812, 0.012630047798156739, 0.012588640213012696, 0.012615903854370118, 0.012601375579833984, 0.012597408294677735, 0.012550144195556641, 0.012527071952819824, 0.012458527565002441, 0.012556223869323731, 0.012648480415344238, 0.012507200241088867, 0.012504416465759277, 0.012509568214416504, 0.012523743629455566, 0.01256601619720459, 0.012554656028747559, 0.012549983978271484, 0.012593279838562012, 0.012567999839782715, 0.012499456405639648, 0.01245631980895996, 0.012452768325805665, 0.012350272178649903, 0.012297344207763673, 0.012249919891357422, 0.012216511726379395, 0.012270976066589355, 0.012331456184387206, 0.0122772159576416, 0.01239311981201172, 0.012245408058166504, 0.01236342430114746, 0.012362815856933594, 0.012325344085693359, 0.012282431602478027, 0.012263168334960937, 0.012308287620544433, 0.01205452823638916, 0.012818431854248047, 0.012981792449951172, 0.0124235200881958, 0.013357343673706055, 0.012404640197753907, 0.012535840034484864, 0.012285856246948243, 0.012298015594482422, 0.01233737564086914, 0.012283488273620606, 0.012321184158325196, 0.012408415794372558, 0.012347807884216308, 0.012522879600524903, 0.012556927680969238, 0.01256828784942627, 0.012576543807983398, 0.013664768218994141, 0.012488703727722168, 0.012550399780273438, 0.012438431739807129, 0.01252233600616455, 0.0126113920211792, 0.012936767578125, 0.01260319995880127, 0.012589887619018554, 0.012640255928039551, 0.012595199584960937, 0.012571776390075683, 0.012598112106323242, 0.012705535888671874, 0.012550175666809082, 0.012551967620849609, 0.012612256050109863, 0.012598400115966796, 0.01304851245880127, 0.01276956844329834, 0.015601216316223144, 0.012673184394836426, 0.012630016326904296, 0.012533760070800782, 0.012572671890258789, 0.012576864242553712, 0.012586175918579101, 0.012524255752563477, 0.01268496036529541, 0.012529824256896973, 0.012558527946472168, 0.012465439796447753, 0.014080639839172363, 0.012463583946228027, 0.012462719917297364, 0.012570624351501464, 0.013017087936401368, 0.012619935989379882, 0.012543840408325195, 0.012543328285217285, 0.012552448272705078, 0.012547904014587403, 0.01251798439025879, 0.012408191680908204, 0.012456576347351074, 0.012217984199523926, 0.012511103630065919, 0.012423551559448243, 0.012394687652587891, 0.01236950397491455, 0.012406559944152832, 0.012351776123046875, 0.012304672241210938, 0.012322431564331054, 0.012355072021484375, 0.012395392417907715, 0.012375136375427247, 0.012358559608459472, 0.012378111839294433, 0.012354623794555665, 0.012394975662231446, 0.012495327949523925, 0.012518719673156738, 0.012509856224060058, 0.012531776428222657, 0.012476384162902832, 0.012389504432678222, 0.012379008293151856, 0.012408543586730957, 0.012398112297058105, 0.012400383949279786, 0.012431808471679687, 0.012454463958740234, 0.0124235200881958, 0.012523455619812012, 0.012326080322265625, 0.012372096061706543, 0.012310943603515624, 0.012464127540588378, 0.012969344139099122, 0.012442399978637695, 0.012386143684387207, 0.012390144348144531, 0.012336607933044433, 0.012727264404296876, 0.012269375801086425, 0.01232588768005371, 0.01232307243347168, 0.012269632339477539, 0.012282560348510742, 0.012214271545410157, 0.012426912307739258, 0.012257311820983887, 0.012244895935058593, 0.012198016166687012, 0.012229056358337402, 0.012285568237304688, 0.0122390718460083, 0.012298111915588378, 0.012322943687438965, 0.01251142406463623, 0.012365632057189942, 0.012268927574157715, 0.012226495742797851, 0.0125283203125, 0.012339200019836426, 0.012298527717590333, 0.01233033561706543, 0.012116000175476075, 0.012436927795410157, 0.012896224021911621, 0.01255894374847412, 0.012484767913818359, 0.012489855766296387, 0.012415712356567382, 0.012634112358093261, 0.012416959762573242, 0.012345727920532226, 0.012392224311828614, 0.012455455780029297, 0.012460543632507324, 0.012533632278442383, 0.012511296272277832, 0.012629055976867675, 0.012495743751525878, 0.01244979190826416, 0.01254745578765869, 0.012431200027465821, 0.01245263957977295, 0.012456992149353027, 0.012579360008239747, 0.012689824104309083, 0.012680959701538087, 0.012544287681579589, 0.012683391571044922, 0.013142111778259278, 0.012650176048278809, 0.012500896453857421, 0.01243769645690918, 0.012602751731872558, 0.012573087692260742, 0.012437727928161621, 0.012486656188964844, 0.012488703727722168, 0.012421152114868165, 0.012443327903747558, 0.01245417594909668, 0.01257695960998535, 0.012599167823791503, 0.013299967765808106, 0.015553248405456544, 0.012569664001464843, 0.01242080020904541, 0.012658528327941895, 0.012370559692382812, 0.012427007675170899, 0.012522687911987304, 0.013081503868103026, 0.012642208099365235, 0.012537471771240235, 0.012498559951782227, 0.012466943740844727, 0.012349439620971679, 0.01221225643157959, 0.012250271797180176, 0.012286687850952148, 0.012257439613342285, 0.012265439987182616, 0.012224032402038574, 0.012267968177795411, 0.012322815895080566, 0.011807328224182128, 0.012056575775146485, 0.012158880233764649, 0.01208944034576416, 0.012038175582885741, 0.012033568382263184, 0.012052767753601075, 0.01198095989227295, 0.012015616416931153, 0.011972415924072265, 0.01202195167541504, 0.012025088310241698, 0.01201638412475586, 0.011916799545288086, 0.011959103584289551, 0.011986623764038086, 0.011965567588806153, 0.01189151954650879, 0.011905088424682617, 0.012107904434204102, 0.012148608207702636, 0.01248256015777588, 0.013426719665527343, 0.012582880020141601, 0.012345343589782716, 0.01233897590637207, 0.012308608055114746, 0.012224703788757325, 0.012258272171020507, 0.012302847862243652, 0.012316608428955078, 0.012290080070495606, 0.012313055992126464, 0.012319968223571778, 0.012339008331298829, 0.0126430082321167, 0.0123722562789917, 0.012371968269348145, 0.012345151901245117, 0.012380352020263672, 0.012308095932006836, 0.012294431686401368, 0.012263360023498536, 0.012241056442260742, 0.012275712013244629, 0.012231712341308593, 0.01233619213104248, 0.012351391792297363, 0.012391712188720703, 0.01243996810913086, 0.012444191932678223, 0.012523296356201172, 0.012445919990539551, 0.012383168220520019, 0.012364255905151366, 0.012318464279174804, 0.01228867244720459, 0.012906047821044922, 0.01235308837890625, 0.012290783882141113, 0.012251263618469238, 0.012298239707946777, 0.012337151527404786, 0.012058464050292969, 0.012327072143554687, 0.012275967597961426, 0.012221535682678223, 0.012277631759643555, 0.012233535766601562, 0.012230015754699708, 0.012216256141662598, 0.01208569622039795, 0.012086784362792969, 0.01212492847442627, 0.012111104011535645, 0.012045023918151855, 0.012124159812927245, 0.01225119972229004, 0.012345343589782716, 0.012387424468994141, 0.01243836784362793, 0.01243068790435791, 0.012491104125976562, 0.012435839653015137, 0.012418111801147462, 0.012391231536865234, 0.012374079704284668, 0.012371744155883789, 0.012331551551818847, 0.012381024360656738, 0.01241097640991211, 0.012429120063781739, 0.012401599884033204, 0.012404735565185548, 0.012406111717224121, 0.012396927833557128, 0.012396832466125489, 0.012496031761169434, 0.012471136093139648, 0.012529664039611817, 0.01247436809539795, 0.012496031761169434, 0.012559200286865234, 0.012515711784362792, 0.012472000122070313, 0.012423104286193847, 0.012615232467651368, 0.012632224082946777, 0.012581151962280273, 0.012650272369384766, 0.0128307523727417, 0.012777664184570313, 0.012703583717346192, 0.012666399955749513, 0.012577407836914063, 0.012559712409973145, 0.012547776222229004, 0.012548480033874512, 0.012513888359069825, 0.012597151756286621, 0.012537407875061034, 0.012571040153503419, 0.012529600143432617, 0.012518752098083495, 0.012542143821716308, 0.01258351993560791]",tokens/s,80.38384339012136,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1491.501056,1546.584064,0.0,1168.113664,1154.613248,s,1,8.3148095703125,8.3148095703125,0.0,8.3148095703125,8.3148095703125,8.3148095703125,8.3148095703125,[8.3148095703125],,kWh,3.861054563332876e-05,4.251723312599966e-06,1.2329732086008294e-05,5.5192001031937024e-05,,MB,1591.00928,1796.145152,0.0,1388.314624,1334.065152,s,10,0.8245529479980468,0.08245529479980468,0.00039060264817304067,0.08240457534790038,0.0828208740234375,0.08305651702880859,0.08324503143310547,"[0.08329216003417969, 0.08210860443115234, 0.08276850891113281, 0.08228358459472657, 0.08231641387939453, 0.08249273681640625, 0.08180339050292969, 0.08251830291748047, 0.08223340606689453, 0.08273583984375]",tokens/s,3104.7126885125926,kWh,2.4979876080481665e-06,2.7548447445572607e-07,1.655366946324919e-06,4.428839028828811e-06,tokens/kWh,57802958.81914186,MB,1599.61088,1796.145152,0.0,1388.314624,1372.847616,s,10,14.184140502929688,1.4184140502929687,0.012717605394795299,1.4124118041992189,1.4413192993164063,1.441889324951172,1.4423453454589845,"[1.441192626953125, 1.41733544921875, 1.424362548828125, 1.4087288818359376, 1.4124964599609375, 1.4068421630859376, 1.40653955078125, 1.4424593505859375, 1.4123271484375, 1.4118563232421875]",tokens/s,44.41580368368994,kWh,4.158395035028572e-05,4.5863410420389415e-06,2.066892869087587e-05,6.683922008320053e-05,tokens/kWh,942560.369818476,,s,630,14.181650554656967,0.022510556435963464,0.000441573484618649,0.022355904579162596,0.023032106399536133,0.023229604244232177,0.024400959815979003,"[0.022883968353271486, 0.022775615692138672, 0.022372928619384766, 0.02328134346008301, 0.022628671646118165, 0.02272390365600586, 0.02265567970275879, 0.022644384384155274, 0.022833503723144532, 0.022494848251342774, 0.02241574478149414, 0.022583295822143554, 0.02249932861328125, 0.022908992767333984, 0.023009248733520508, 0.022877920150756837, 0.023116031646728517, 0.02322003173828125, 0.02311187171936035, 0.023651840209960938, 0.023056896209716796, 0.02309529685974121, 0.02294304084777832, 0.02310147285461426, 0.02301728057861328, 0.024599391937255858, 0.024070144653320313, 0.023216127395629883, 0.023259136199951173, 0.02327302360534668, 0.023402944564819336, 0.023218175888061524, 0.022996992111206056, 0.02312396812438965, 0.023259136199951173, 0.02307276725769043, 0.025007776260375977, 0.023136608123779295, 0.022815967559814455, 0.0225020809173584, 0.022383808135986328, 0.02230067253112793, 0.022448736190795897, 0.022437728881835938, 0.022378463745117188, 0.022277887344360352, 0.022255552291870116, 0.022253631591796875, 0.02218060874938965, 0.022359167098999023, 0.022243200302124025, 0.022194303512573243, 0.022563711166381836, 0.022208511352539064, 0.022335039138793946, 0.022378496170043945, 0.02266124725341797, 0.024013120651245116, 0.02289023971557617, 0.02280473518371582, 0.022786048889160155, 0.022952959060668944, 0.022753311157226563, 0.02301644706726074, 0.02279612731933594, 0.022786048889160155, 0.02294988822937012, 0.02255264091491699, 0.022235071182250977, 0.022177791595458983, 0.02229043197631836, 0.022593536376953126, 0.022623872756958006, 0.02232940864562988, 0.02225388717651367, 0.022257568359375, 0.022298719406127928, 0.022323200225830078, 0.022971935272216797, 0.0224116153717041, 0.022853759765625, 0.022228384017944337, 0.02239139175415039, 0.02236595153808594, 0.022364416122436524, 0.022529535293579102, 0.022729215621948243, 0.022755327224731444, 0.022562271118164064, 0.022388832092285156, 0.022342079162597655, 0.022296480178833008, 0.02247065544128418, 0.02235638427734375, 0.022286016464233397, 0.02224028778076172, 0.022258655548095703, 0.022404640197753907, 0.02227862358093262, 0.022374399185180666, 0.02251571273803711, 0.022420576095581055, 0.023034784317016603, 0.02244607925415039, 0.022343423843383788, 0.02227132797241211, 0.022188959121704103, 0.022254751205444335, 0.022262624740600586, 0.02230646324157715, 0.02229078483581543, 0.022312416076660156, 0.02226211166381836, 0.022220991134643556, 0.02251366424560547, 0.022529439926147463, 0.022677087783813478, 0.022727231979370117, 0.022767263412475584, 0.02290934371948242, 0.02287449645996094, 0.022814176559448243, 0.022614559173583983, 0.022638336181640625, 0.022840991973876953, 0.022696544647216797, 0.022761823654174805, 0.022866464614868163, 0.023133951187133787, 0.022542591094970702, 0.02263654327392578, 0.022433792114257813, 0.02242723274230957, 0.023140159606933594, 0.02424687957763672, 0.025407487869262696, 0.02250752067565918, 0.02267750358581543, 0.02267750358581543, 0.022466144561767577, 0.02253046417236328, 0.022842912673950194, 0.022785791397094725, 0.02378620719909668, 0.022408384323120117, 0.022247392654418944, 0.022202816009521485, 0.022208320617675782, 0.022264415740966798, 0.022228448867797852, 0.022065696716308595, 0.022181888580322266, 0.022411231994628907, 0.02440809631347656, 0.022855680465698244, 0.022275455474853517, 0.022387327194213866, 0.02217740821838379, 0.022122400283813477, 0.02222537612915039, 0.022192127227783204, 0.022598783493041993, 0.02252067184448242, 0.023130144119262695, 0.022355840682983397, 0.023351423263549803, 0.024786943435668944, 0.022586719512939453, 0.02228291130065918, 0.022366079330444336, 0.022209888458251954, 0.022217472076416014, 0.022179519653320313, 0.02222729682922363, 0.022106111526489256, 0.022106111526489256, 0.022194175720214843, 0.022514879226684572, 0.02235475158691406, 0.02232691192626953, 0.02224166488647461, 0.022228992462158204, 0.022435104370117188, 0.02245091247558594, 0.022202144622802733, 0.022455808639526367, 0.02236079978942871, 0.022337535858154296, 0.022275455474853517, 0.022552576065063477, 0.02245631980895996, 0.02229043197631836, 0.02243328094482422, 0.02217011260986328, 0.022390783309936522, 0.022234176635742186, 0.02214393615722656, 0.022216575622558594, 0.022159488677978515, 0.022279712677001955, 0.022272480010986327, 0.02224336051940918, 0.02218364715576172, 0.022210208892822266, 0.022238815307617187, 0.022316064834594727, 0.022187999725341797, 0.022181184768676757, 0.022153919219970702, 0.022237184524536133, 0.022147071838378905, 0.022197311401367187, 0.022188192367553712, 0.02226655960083008, 0.022278240203857422, 0.022312959671020507, 0.022340991973876952, 0.022227136611938477, 0.022311199188232423, 0.022413471221923827, 0.022147071838378905, 0.022304031372070314, 0.02323324775695801, 0.02351103973388672, 0.022396928787231447, 0.022261760711669923, 0.022554079055786134, 0.02232304000854492, 0.0223024959564209, 0.02263484764099121, 0.02215497589111328, 0.02224844741821289, 0.02221142387390137, 0.022231168746948242, 0.022216800689697266, 0.02222774314880371, 0.022278144836425783, 0.02220185661315918, 0.022161184310913087, 0.022171392440795898, 0.02208867263793945, 0.022404767990112304, 0.022303071975708008, 0.02246396827697754, 0.022314592361450194, 0.022388927459716795, 0.022864639282226564, 0.022550527572631835, 0.0226810245513916, 0.02312153625488281, 0.022612640380859375, 0.022675519943237306, 0.02293609619140625, 0.022534271240234376, 0.022812671661376953, 0.022734848022460938, 0.022807743072509764, 0.022569280624389648, 0.02236057662963867, 0.02228428840637207, 0.022380544662475587, 0.02233545684814453, 0.022393951416015623, 0.02222585678100586, 0.022371679306030273, 0.02229520034790039, 0.022417280197143556, 0.02248422431945801, 0.022621055603027344, 0.022996959686279298, 0.022722591400146486, 0.02252390480041504, 0.022364160537719727, 0.02224742317199707, 0.022351871490478514, 0.022318111419677735, 0.022352863311767578, 0.022171392440795898, 0.022233184814453126, 0.02237385559082031, 0.022264511108398437, 0.022242624282836913, 0.02220921516418457, 0.02227801513671875, 0.02235100746154785, 0.02218288040161133, 0.02227519989013672, 0.022331552505493166, 0.02224371147155762, 0.022357919692993163, 0.022319551467895507, 0.022358015060424806, 0.022296575546264647, 0.02227609634399414, 0.022405120849609376, 0.022280128479003906, 0.022296640396118166, 0.022261760711669923, 0.02225667190551758, 0.02222172737121582, 0.022300735473632812, 0.02230271911621094, 0.022466560363769532, 0.022486143112182617, 0.022637439727783204, 0.022380544662475587, 0.022543872833251953, 0.022311424255371092, 0.02237593650817871, 0.02243619155883789, 0.022775840759277344, 0.022409343719482423, 0.022288383483886717, 0.02306662368774414, 0.022550527572631835, 0.0224768009185791, 0.022357088088989258, 0.02243881607055664, 0.02226959991455078, 0.02234409523010254, 0.022258752822875975, 0.02229747200012207, 0.022132736206054687, 0.02219411277770996, 0.022327360153198243, 0.02229827117919922, 0.022258016586303712, 0.022382591247558595, 0.022209823608398436, 0.02227631950378418, 0.023505407333374022, 0.022734336853027344, 0.02234828758239746, 0.022292255401611328, 0.02248521614074707, 0.022220703125, 0.022324832916259765, 0.02221516799926758, 0.022218751907348632, 0.02225702476501465, 0.02223276710510254, 0.02211686325073242, 0.02212499237060547, 0.022165504455566407, 0.02225971221923828, 0.022423360824584963, 0.022300416946411133, 0.022218496322631835, 0.022205312728881835, 0.022292415618896486, 0.02246847915649414, 0.022203807830810548, 0.02239548873901367, 0.02224723243713379, 0.022436031341552733, 0.02213430404663086, 0.022295007705688475, 0.02215936088562012, 0.022222015380859376, 0.02226464080810547, 0.022247264862060547, 0.02241142463684082, 0.02251276779174805, 0.022524383544921874, 0.022315423965454103, 0.022278144836425783, 0.02227916717529297, 0.02233772850036621, 0.022348608016967773, 0.022204416275024414, 0.02225152015686035, 0.0224399356842041, 0.022355968475341798, 0.02222857666015625, 0.022685440063476562, 0.02230108833312988, 0.022335744857788085, 0.022274015426635742, 0.022337535858154296, 0.022419296264648437, 0.022288543701171875, 0.02225766372680664, 0.0222959041595459, 0.022289056777954102, 0.022306623458862303, 0.02239619255065918, 0.022319807052612304, 0.022438079833984374, 0.022249504089355467, 0.02230067253112793, 0.022187360763549803, 0.022196895599365236, 0.02234297561645508, 0.022249183654785155, 0.022603967666625976, 0.02241779136657715, 0.02238649559020996, 0.02240358352661133, 0.022175840377807617, 0.0221759033203125, 0.022234975814819338, 0.022204416275024414, 0.022409215927124023, 0.022200288772583007, 0.022195680618286133, 0.022104639053344727, 0.02212396812438965, 0.02225388717651367, 0.022100223541259765, 0.02213478469848633, 0.022417407989501953, 0.02242355155944824, 0.02229043197631836, 0.022122495651245116, 0.022286336898803712, 0.02224947166442871, 0.023358623504638673, 0.022686559677124022, 0.022351871490478514, 0.022351392745971678, 0.02268169593811035, 0.022258047103881837, 0.022191232681274414, 0.02223107147216797, 0.022283103942871092, 0.02229792022705078, 0.022344383239746093, 0.022216224670410158, 0.022256095886230468, 0.022359615325927736, 0.022346048355102538, 0.022358144760131836, 0.02227609634399414, 0.022253568649291993, 0.02231091117858887, 0.02246659278869629, 0.022317024230957033, 0.02225257682800293, 0.02235696029663086, 0.02237446403503418, 0.02234566307067871, 0.022310335159301756, 0.022362688064575195, 0.022379711151123048, 0.023225151062011718, 0.02478489685058594, 0.022353567123413087, 0.02247292709350586, 0.02226736068725586, 0.022384639739990234, 0.022495904922485353, 0.022478847503662108, 0.022670495986938478, 0.022686559677124022, 0.026043968200683595, 0.023142847061157226, 0.024383487701416014, 0.02275328063964844, 0.02280243110656738, 0.022683647155761717, 0.02246451187133789, 0.022357696533203124, 0.02231737518310547, 0.022276224136352538, 0.022463552474975584, 0.022343679428100584, 0.022423519134521484, 0.02236198425292969, 0.02231395149230957, 0.022306400299072264, 0.022262176513671874, 0.022358015060424806, 0.022908672332763672, 0.022225151062011717, 0.022382272720336913, 0.022355487823486328, 0.022855680465698244, 0.022528064727783202, 0.022604511260986327, 0.022816768646240236, 0.02287820816040039, 0.02285977554321289, 0.0228286075592041, 0.023067071914672853, 0.023119871139526366, 0.022996639251708983, 0.023184831619262696, 0.023116640090942383, 0.02337798309326172, 0.023250944137573244, 0.023222080230712892, 0.023189695358276367, 0.023220224380493162, 0.023193151473999023, 0.023140800476074218, 0.02314035224914551, 0.023031808853149413, 0.0230645751953125, 0.02331443214416504, 0.023328800201416016, 0.02333692741394043, 0.024162303924560546, 0.02351923179626465, 0.023011264801025392, 0.023147584915161133, 0.022755712509155274, 0.022620576858520508, 0.022452384948730468, 0.022826528549194335, 0.02290928077697754, 0.022496639251708986, 0.0223874568939209, 0.02249728012084961, 0.022355615615844728, 0.022369632720947264, 0.022172672271728516, 0.02227609634399414, 0.022181631088256836, 0.02245043182373047, 0.022451616287231444, 0.022602336883544922, 0.022378496170043945, 0.022425600051879883, 0.022458368301391602, 0.022343679428100584, 0.022512832641601564, 0.022276927947998047, 0.02226563262939453, 0.022411136627197265, 0.02228873634338379, 0.022285600662231446, 0.022278879165649416, 0.022312959671020507, 0.022218719482421875, 0.02222697639465332, 0.022318496704101562, 0.022311519622802735, 0.022251487731933594, 0.02218377685546875, 0.022349119186401367, 0.022256511688232422, 0.02223468780517578, 0.022096319198608397, 0.02228428840637207, 0.02230179214477539, 0.022287040710449218, 0.0222589111328125, 0.022162431716918944, 0.022220800399780274, 0.022228992462158204, 0.02246451187133789, 0.02233263969421387, 0.022315807342529297, 0.022304767608642577, 0.02233296012878418, 0.022170080184936523, 0.02267955207824707, 0.0243056640625, 0.022999040603637694, 0.02231491279602051, 0.022304224014282226, 0.022551168441772462, 0.022258880615234376, 0.022278976440429688, 0.02251087951660156, 0.02230659294128418, 0.02228495979309082, 0.02249318313598633, 0.02240716743469238, 0.022392192840576173, 0.02232588768005371, 0.022392831802368163, 0.02227609634399414, 0.024000511169433594, 0.022401023864746093, 0.022353919982910156, 0.02224742317199707, 0.02232729530334473, 0.02242473602294922, 0.022266752243041994, 0.022249439239501952, 0.02230067253112793, 0.022335487365722655, 0.02275846481323242, 0.022360832214355468, 0.022363391876220703, 0.022344640731811524, 0.02225702476501465, 0.022260032653808593, 0.02222105598449707, 0.02230806350708008, 0.022258527755737306, 0.02214694404602051, 0.022241312026977537, 0.022275264739990235, 0.02226700782775879, 0.022187423706054688, 0.02218022346496582, 0.02249318313598633, 0.022268159866333008, 0.022181312561035157, 0.022305088043212892, 0.022245216369628906, 0.022288543701171875, 0.022231039047241212, 0.02251087951660156, 0.022242015838623046, 0.022273855209350588, 0.02231100845336914, 0.022851680755615233, 0.022468608856201173, 0.02234163284301758, 0.022269952774047853, 0.022300640106201173, 0.022312032699584962, 0.02226483154296875, 0.022200096130371095, 0.02236128044128418, 0.02227020835876465, 0.022468927383422852, 0.02255036735534668, 0.022633056640625, 0.022695903778076173, 0.022629695892333983, 0.02271504020690918, 0.022581600189208986, 0.02269264030456543, 0.022758272171020506, 0.0226345272064209, 0.02259552001953125]",tokens/s,44.423602004007925,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1847.76704,2718.892032,0.0,2340.421632,2285.568,s,1,9.1549111328125,9.1549111328125,0.0,9.1549111328125,9.1549111328125,9.1549111328125,9.1549111328125,[9.1549111328125],,kWh,6.062554752914669e-05,6.676919272620522e-06,2.004196047800111e-05,8.734442727976833e-05,,MB,1862.26688,3087.990784,0.0,2680.160256,2578.855936,s,10,0.8394956512451173,0.08394956512451172,0.0005466569905256003,0.08386230087280273,0.08469899291992188,0.0848067756652832,0.08489300186157227,"[0.08491455841064453, 0.08333392333984375, 0.08359808349609375, 0.08373212432861328, 0.08372908782958985, 0.08399247741699219, 0.08308509063720704, 0.08467504119873047, 0.08435814666748047, 0.08407711791992188]",tokens/s,3049.4499836932773,kWh,2.601100544247896e-06,2.8680764648819035e-07,1.7284355027078933e-06,4.616343693443979e-06,tokens/kWh,55455143.074283026,MB,1862.26688,3087.990784,0.0,2680.160256,2578.858496,s,10,16.271067260742186,1.6271067260742185,0.00914318288080362,1.6311510620117189,1.6338862548828126,1.63465595703125,1.6352717187499999,"[1.63228515625, 1.628584228515625, 1.6121324462890625, 1.6354256591796874, 1.606869384765625, 1.6318067626953126, 1.6337152099609376, 1.6274884033203125, 1.630495361328125, 1.6322646484375]",tokens/s,38.71903360144203,kWh,4.652391645158327e-05,5.1313133475030975e-06,2.424697416649375e-05,7.590220396558012e-05,tokens/kWh,830015.4239074405,,s,630,16.268711450576784,0.025823351508852035,0.00041652314541038346,0.025816512107849122,0.026103797721862793,0.026413397789001463,0.02720422792434693,"[0.027130592346191407, 0.02615091133117676, 0.02595430374145508, 0.025999359130859375, 0.025994943618774413, 0.0259116153717041, 0.02613590431213379, 0.025985696792602538, 0.02609699249267578, 0.025886592864990236, 0.02596329689025879, 0.02595840072631836, 0.02592131233215332, 0.026223007202148436, 0.026103647232055664, 0.026085439682006835, 0.026139743804931642, 0.025955135345458985, 0.02582307243347168, 0.026020000457763672, 0.02592665672302246, 0.025818111419677735, 0.026755071640014647, 0.02574336051940918, 0.025810304641723632, 0.02568780708312988, 0.02572198486328125, 0.025695072174072266, 0.025461376190185545, 0.02553875160217285, 0.02553455924987793, 0.025527839660644532, 0.025780704498291014, 0.025591808319091795, 0.02540652847290039, 0.0254835205078125, 0.025446144104003907, 0.0254486083984375, 0.02541423988342285, 0.02560985565185547, 0.0255534725189209, 0.026009599685668947, 0.025882623672485353, 0.02589286422729492, 0.02591744041442871, 0.02609766387939453, 0.025968639373779297, 0.025862207412719728, 0.025884063720703124, 0.02599171257019043, 0.025982112884521485, 0.026026304244995118, 0.02599103927612305, 0.02594643211364746, 0.025958751678466795, 0.025827199935913085, 0.025922943115234375, 0.025962400436401366, 0.02603094482421875, 0.02588057518005371, 0.025790464401245116, 0.026869760513305665, 0.02595430374145508, 0.026584415435791015, 0.025816736221313478, 0.02659868812561035, 0.025992927551269532, 0.025860095977783205, 0.02594508743286133, 0.025967615127563477, 0.026025983810424806, 0.026017663955688476, 0.026310400009155275, 0.02603046417236328, 0.02614476776123047, 0.026019584655761718, 0.02608768081665039, 0.02616227149963379, 0.026016672134399413, 0.02591744041442871, 0.026275840759277344, 0.02594927978515625, 0.02614159965515137, 0.026858816146850584, 0.02607174491882324, 0.025951744079589844, 0.0260183048248291, 0.026068159103393555, 0.025972864151000977, 0.02591801643371582, 0.026009376525878907, 0.025995616912841798, 0.026032127380371094, 0.026105152130126954, 0.025846464157104492, 0.025802751541137696, 0.02698854446411133, 0.025800703048706054, 0.025829376220703124, 0.025899007797241212, 0.02591334342956543, 0.02571673583984375, 0.02574867248535156, 0.025879360198974608, 0.02625859260559082, 0.025893280029296875, 0.02575814437866211, 0.025540607452392578, 0.0256135368347168, 0.02570524787902832, 0.025540607452392578, 0.02555084800720215, 0.02556662368774414, 0.02546745681762695, 0.025151039123535158, 0.02510483169555664, 0.02505116844177246, 0.025298143386840822, 0.0252271671295166, 0.025381343841552734, 0.02521708869934082, 0.025149791717529298, 0.025333152770996094, 0.025598560333251953, 0.025438047409057616, 0.02524736022949219, 0.026045343399047852, 0.025817087173461914, 0.025727136611938477, 0.025640480041503905, 0.025370752334594727, 0.025132383346557617, 0.026252031326293945, 0.025604192733764648, 0.02525491142272949, 0.025426944732666015, 0.025452192306518556, 0.02529520034790039, 0.025140960693359374, 0.02519798469543457, 0.02533670425415039, 0.025274368286132814, 0.025070911407470704, 0.025174720764160156, 0.025184255599975586, 0.025251840591430662, 0.025167552947998047, 0.025205055236816407, 0.02547427177429199, 0.025376895904541015, 0.025207359313964842, 0.02539299201965332, 0.025374528884887695, 0.025464448928833008, 0.025338687896728516, 0.02531449508666992, 0.02606982421875, 0.02524729537963867, 0.02531372833251953, 0.02554409599304199, 0.025609983444213866, 0.02569664001464844, 0.025520608901977538, 0.02566534423828125, 0.025587327957153322, 0.025835615158081054, 0.025774560928344726, 0.02577177619934082, 0.025796863555908205, 0.026169343948364256, 0.025710687637329102, 0.026701728820800782, 0.025863487243652342, 0.02570515251159668, 0.02565449523925781, 0.025823007583618163, 0.025705472946166992, 0.02573107147216797, 0.025757696151733397, 0.02593756866455078, 0.02565974426269531, 0.025881599426269532, 0.025712799072265625, 0.02582601547241211, 0.02579849624633789, 0.02560233688354492, 0.025599199295043944, 0.025835552215576173, 0.02581385612487793, 0.026593408584594726, 0.025968639373779297, 0.02602406311035156, 0.025956575393676757, 0.025945375442504883, 0.025853376388549804, 0.02585696029663086, 0.025882976531982422, 0.02589263916015625, 0.025935712814331054, 0.025949888229370117, 0.025846111297607423, 0.02591526412963867, 0.025829183578491212, 0.02594972801208496, 0.025908000946044923, 0.025871679306030272, 0.02589766311645508, 0.027058176040649414, 0.02635366439819336, 0.026094976425170897, 0.026055295944213866, 0.026009599685668947, 0.02586953544616699, 0.025951007843017578, 0.026015743255615235, 0.02578963279724121, 0.025930559158325196, 0.025870336532592773, 0.026048511505126954, 0.025753664016723632, 0.02578201675415039, 0.025714176177978516, 0.025793216705322267, 0.025967775344848634, 0.02596268844604492, 0.025889440536499022, 0.025956352233886718, 0.02603545570373535, 0.026270399093627928, 0.025720895767211913, 0.02582703971862793, 0.025536800384521486, 0.02549760055541992, 0.025618431091308593, 0.025651199340820312, 0.026524927139282225, 0.028223455429077147, 0.025776927947998046, 0.025806848526000976, 0.025748512268066407, 0.02573155212402344, 0.025725439071655275, 0.025640960693359374, 0.02567884826660156, 0.02575667190551758, 0.026731840133666994, 0.025780096054077148, 0.025795391082763672, 0.025792415618896485, 0.02578374481201172, 0.025765695571899415, 0.025817535400390626, 0.026429439544677736, 0.025906463623046876, 0.025955039978027342, 0.02593120002746582, 0.026473215103149414, 0.025994335174560547, 0.02584217643737793, 0.025720800399780273, 0.025725183486938478, 0.02558361625671387, 0.025664735794067382, 0.025588512420654297, 0.025536512374877928, 0.025293855667114257, 0.025068511962890627, 0.024983552932739257, 0.02483344078063965, 0.025037408828735352, 0.024844287872314453, 0.025153535842895508, 0.02489139175415039, 0.02508185577392578, 0.024860607147216798, 0.024853952407836916, 0.02493503952026367, 0.025184192657470704, 0.025089696884155275, 0.02523334312438965, 0.025330144882202147, 0.02622604751586914, 0.0255863037109375, 0.025708351135253906, 0.025469120025634766, 0.02616659164428711, 0.02546963119506836, 0.025517087936401367, 0.025256927490234376, 0.0253986873626709, 0.02538147163391113, 0.02556694412231445, 0.025475360870361327, 0.025556991577148438, 0.02551398468017578, 0.02548284721374512, 0.025500064849853517, 0.02569024085998535, 0.025558464050292967, 0.025868383407592774, 0.025604448318481444, 0.025750688552856445, 0.025752416610717775, 0.025715839385986327, 0.02556403160095215, 0.02569625663757324, 0.025552896499633788, 0.025425312042236328, 0.025354400634765625, 0.025426431655883788, 0.025457632064819335, 0.025504735946655272, 0.025510047912597655, 0.02561756706237793, 0.025304864883422852, 0.026129600524902343, 0.025505887985229493, 0.02648297691345215, 0.025583360671997072, 0.025551519393920898, 0.025404800415039064, 0.02548601531982422, 0.0255446720123291, 0.025665536880493164, 0.02559494400024414, 0.02587487983703613, 0.025646944046020508, 0.025651872634887694, 0.025657344818115234, 0.025673728942871094, 0.02571878433227539, 0.02639379119873047, 0.029094240188598634, 0.025950687408447266, 0.02590444755554199, 0.025733823776245116, 0.02592972755432129, 0.026036224365234374, 0.025994815826416016, 0.02589961624145508, 0.025886560440063478, 0.025802400588989256, 0.025772384643554688, 0.025757728576660158, 0.0259849910736084, 0.02582476806640625, 0.025778688430786133, 0.028792383193969727, 0.02750035285949707, 0.026251039505004882, 0.02592854309082031, 0.02578553581237793, 0.025801183700561524, 0.02567612838745117, 0.025792512893676758, 0.025753599166870117, 0.02667910385131836, 0.026197471618652345, 0.025895647048950195, 0.02579987144470215, 0.02586092758178711, 0.02561347198486328, 0.025645919799804687, 0.025677919387817383, 0.02576153564453125, 0.025596063613891603, 0.025660640716552736, 0.02557360076904297, 0.025577152252197265, 0.025394144058227538, 0.02536639976501465, 0.0254997444152832, 0.025577407836914062, 0.025572383880615234, 0.02558051109313965, 0.02546227264404297, 0.0258155517578125, 0.025570943832397462, 0.026572799682617186, 0.025797983169555665, 0.02591606330871582, 0.025763423919677734, 0.02592527961730957, 0.025749984741210936, 0.02581737518310547, 0.02588431930541992, 0.025839679718017577, 0.025808895111083984, 0.02602627182006836, 0.025704095840454102, 0.025831647872924805, 0.025835615158081054, 0.027063520431518554, 0.02600796890258789, 0.025956031799316406, 0.026026079177856445, 0.02605072021484375, 0.025820863723754882, 0.025847808837890625, 0.026225408554077147, 0.025978496551513672, 0.026597343444824218, 0.025949760437011717, 0.025858495712280275, 0.02591584014892578, 0.025817087173461914, 0.02590015983581543, 0.025811840057373046, 0.02598636817932129, 0.025864736557006836, 0.02601558494567871, 0.026269535064697265, 0.026018239974975585, 0.025728832244873046, 0.025866239547729493, 0.025839839935302734, 0.025831424713134765, 0.025651199340820312, 0.025654367446899414, 0.025772960662841796, 0.025665536880493164, 0.02560358428955078, 0.02573091125488281, 0.025664159774780274, 0.025680063247680664, 0.02578518486022949, 0.025846271514892577, 0.025811424255371095, 0.025878240585327148, 0.025696544647216796, 0.027234304428100587, 0.026108192443847655, 0.02586515235900879, 0.025692960739135743, 0.025886720657348632, 0.02569606399536133, 0.02591084861755371, 0.026179807662963867, 0.026077600479125978, 0.025707551956176758, 0.025945056915283204, 0.026597408294677733, 0.025973791122436522, 0.025908191680908202, 0.026081279754638673, 0.025804800033569338, 0.02632441520690918, 0.02590287971496582, 0.026030847549438477, 0.025909120559692383, 0.026159263610839843, 0.025944063186645508, 0.025935007095336915, 0.02598179244995117, 0.026005504608154296, 0.02595155143737793, 0.02590585517883301, 0.02584137535095215, 0.026077472686767578, 0.025827327728271485, 0.025881887435913086, 0.02579961585998535, 0.02594803237915039, 0.025865888595581053, 0.02585420799255371, 0.025777952194213867, 0.027056352615356445, 0.02593791961669922, 0.02596611213684082, 0.025778848648071288, 0.02550377655029297, 0.025658336639404297, 0.025586015701293947, 0.026386463165283203, 0.025657791137695313, 0.02553798484802246, 0.025498176574707033, 0.025568479537963866, 0.02552239990234375, 0.02592416000366211, 0.02600134468078613, 0.02545055961608887, 0.02553152084350586, 0.02574425506591797, 0.025380704879760744, 0.025443872451782226, 0.025544704437255858, 0.025641599655151368, 0.025462591171264648, 0.025604223251342772, 0.025540000915527345, 0.02572764778137207, 0.025614336013793947, 0.025818527221679686, 0.02567843246459961, 0.025856000900268555, 0.025579456329345704, 0.025870399475097658, 0.025796192169189453, 0.02574172782897949, 0.02592518424987793, 0.025851743698120117, 0.025805408477783204, 0.02575564765930176, 0.02671615982055664, 0.026029855728149413, 0.02589104080200195, 0.026058752059936522, 0.026259456634521484, 0.02608332824707031, 0.025845760345458983, 0.025860095977783205, 0.025781343460083008, 0.025907712936401366, 0.025885087966918945, 0.0259051513671875, 0.02581033515930176, 0.025796192169189453, 0.0256409912109375, 0.02576278305053711, 0.02570854377746582, 0.02578563117980957, 0.02575164794921875, 0.025751487731933594, 0.025782976150512695, 0.025847808837890625, 0.02578220748901367, 0.02577004814147949, 0.02578166389465332, 0.02582307243347168, 0.025807167053222658, 0.025850303649902345, 0.02578179168701172, 0.025817440032958983, 0.025770111083984373, 0.025769535064697265, 0.026024383544921877, 0.025845056533813478, 0.025757568359375, 0.02586419105529785, 0.02690336036682129, 0.02608687973022461, 0.025736831665039064, 0.02582819175720215, 0.025916671752929686, 0.025908031463623048, 0.02884351921081543, 0.026077056884765627, 0.02596713638305664, 0.026081375122070313, 0.025884159088134767, 0.025849376678466797, 0.02591638374328613, 0.025790464401245116, 0.025866176605224608, 0.025704000473022463, 0.025772544860839845, 0.025965599060058592, 0.025736160278320312, 0.02570035171508789, 0.025560224533081054, 0.025445215225219725, 0.025221120834350585, 0.02519196891784668, 0.025239744186401368, 0.025332000732421874, 0.025419647216796876, 0.026289888381958008, 0.025775968551635744, 0.025670080184936522, 0.025735040664672852, 0.02572915267944336, 0.02584137535095215, 0.025897247314453125, 0.025886720657348632, 0.026193920135498046, 0.02687385559082031, 0.02580588722229004, 0.025768896102905274, 0.02587001609802246, 0.02567731285095215, 0.02583635139465332, 0.02570854377746582, 0.025913087844848633, 0.025790143966674804, 0.025786943435668945, 0.02573516845703125, 0.02595430374145508, 0.02598684883117676, 0.025984415054321287, 0.025899839401245118, 0.02595235252380371, 0.02555244827270508, 0.025874784469604492, 0.025753599166870117, 0.02607513618469238, 0.025860095977783205, 0.02582271957397461, 0.02582969665527344, 0.025859487533569335, 0.02558380889892578, 0.02570844841003418, 0.02554742431640625, 0.02785487937927246, 0.02584480094909668, 0.02577872085571289, 0.02581670379638672, 0.025979679107666017, 0.02574745559692383, 0.025816320419311523, 0.02575391960144043, 0.02582486343383789, 0.025723743438720702, 0.02597887992858887, 0.026787296295166015, 0.02590508842468262, 0.025729631423950194, 0.025847808837890625, 0.02590243148803711, 0.02595439910888672, 0.025823392868041993, 0.025868703842163086, 0.025856000900268555, 0.025888832092285156, 0.025802688598632814, 0.025831392288208008, 0.026034208297729493, 0.025937471389770508, 0.025848255157470704, 0.025849855422973633]",tokens/s,38.72464035728314,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,gemv,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,869.863424,556.72832,0.0,178.25792,176.546816,s,1,7.713306640625,7.713306640625,0.0,7.713306640625,7.713306640625,7.713306640625,7.713306640625,[7.713306640625],,kWh,2.018954481249201e-05,2.2169501482602328e-06,6.545560791992866e-06,2.895205575274511e-05,,MB,1167.09376,661.58592,0.0,253.755392,221.106688,s,16,0.22666860866546631,0.014166788041591645,7.366024324280652e-05,0.014159088134765626,0.014224016189575195,0.01426743197441101,0.01435408148765564,"[0.014181344032287598, 0.014150176048278808, 0.014187104225158691, 0.01410108757019043, 0.014125760078430177, 0.014212160110473633, 0.014231328010559083, 0.014052319526672363, 0.014131360054016114, 0.014070048332214355, 0.014132672309875489, 0.014129983901977538, 0.014375743865966796, 0.014202816009521484, 0.014216704368591309, 0.014168000221252442]",tokens/s,18070.43341429412,kWh,4.245650116763654e-07,4.6821627602398355e-08,2.809361097616254e-07,7.523227490403892e-07,tokens/kWh,340279488.19377834,MB,1201.364992,676.265984,0.0,268.435456,221.109248,s,16,10.183209472656252,0.6364505920410156,0.006485617121469593,0.6356436767578124,0.6450568542480469,0.646190658569336,0.6476012664794921,"[0.6382002563476562, 0.6479539184570312, 0.6319490966796875, 0.6264183959960937, 0.6369337768554687, 0.6389736328125, 0.6244791259765625, 0.6310694580078124, 0.6333386840820312, 0.6341722412109375, 0.6456029052734376, 0.639483642578125, 0.6445108032226563, 0.6432012329101563, 0.6343535766601562, 0.6325687255859375]",tokens/s,98.98647402929906,kWh,1.8477347904209184e-05,2.0375971334138364e-06,7.2614310352382455e-06,2.777637607286127e-05,tokens/kWh,2268114.4521784373,,s,1008,10.174912675857541,0.010094159400652326,0.00023526298569966127,0.01006612777709961,0.010324982166290284,0.010412306880950928,0.01069016888618469,"[0.009861120223999023, 0.010414079666137695, 0.009979904174804688, 0.00999443244934082, 0.009987903594970702, 0.010186752319335938, 0.010055071830749511, 0.010114912033081055, 0.010099136352539062, 0.010152095794677735, 0.010123200416564941, 0.010078847885131836, 0.010112607955932617, 0.01022976016998291, 0.01031987190246582, 0.010203071594238281, 0.010119071960449219, 0.010082464218139648, 0.010083711624145508, 0.01004361629486084, 0.01006015968322754, 0.01005951976776123, 0.010055968284606934, 0.010016575813293457, 0.010135744094848633, 0.010137439727783203, 0.010311840057373046, 0.01031372833251953, 0.010389535903930665, 0.010439680099487305, 0.01033340835571289, 0.010251615524291992, 0.010205087661743164, 0.01019545555114746, 0.010227007865905762, 0.010268447875976562, 0.010138848304748534, 0.010229439735412597, 0.010141247749328614, 0.010085984230041504, 0.01012822437286377, 0.010097920417785644, 0.010217696189880372, 0.010148320198059082, 0.010129216194152833, 0.010099295616149903, 0.010123968124389648, 0.010046431541442871, 0.010053631782531738, 0.010096351623535156, 0.01026863956451416, 0.009992511749267578, 0.010031328201293946, 0.010026592254638672, 0.009978048324584961, 0.009887807846069336, 0.009973695755004883, 0.010045760154724121, 0.010106240272521973, 0.010027327537536622, 0.00999014377593994, 0.009934847831726074, 0.010046527862548829, 0.009910079956054687, 0.01032032012939453, 0.0103088321685791, 0.010349311828613282, 0.010326527595520019, 0.010358431816101073, 0.010336383819580078, 0.010327232360839845, 0.01022432041168213, 0.010190848350524903, 0.010346495628356933, 0.010219648361206054, 0.010196864128112793, 0.010554431915283203, 0.010563872337341308, 0.010399616241455079, 0.010332415580749512, 0.010281791687011719, 0.010198752403259277, 0.01021939182281494, 0.01028825569152832, 0.010473952293395997, 0.010359328269958496, 0.010303263664245605, 0.010283231735229492, 0.010337823867797852, 0.0102423677444458, 0.01025654411315918, 0.010231807708740234, 0.010190848350524903, 0.01010483169555664, 0.010465567588806152, 0.010211039543151855, 0.010157952308654785, 0.010056896209716797, 0.010228832244873047, 0.010055392265319825, 0.010091808319091797, 0.010249343872070313, 0.010220800399780273, 0.010250271797180176, 0.010377152442932129, 0.010487584114074707, 0.010420960426330566, 0.010375167846679688, 0.010466783523559571, 0.010422271728515625, 0.010435232162475586, 0.010432383537292481, 0.010298751831054687, 0.010276512145996093, 0.010095583915710448, 0.010268671989440918, 0.010164223670959472, 0.010092543601989747, 0.010117407798767089, 0.010167072296142578, 0.01025324821472168, 0.010211647987365723, 0.010186431884765625, 0.010287103652954101, 0.010282527923583985, 0.010229472160339356, 0.009875455856323241, 0.010094592094421387, 0.011318943977355956, 0.010265055656433106, 0.010109888076782227, 0.010261407852172852, 0.010049216270446778, 0.01019878387451172, 0.010219552040100098, 0.010252287864685058, 0.010283583641052247, 0.010244128227233886, 0.010175647735595704, 0.010064703941345215, 0.009984000205993653, 0.00993391990661621, 0.009972640037536621, 0.00988588809967041, 0.010098496437072753, 0.009881216049194336, 0.009777183532714845, 0.009850751876831055, 0.009832927703857422, 0.009841823577880859, 0.009816512107849121, 0.009830304145812988, 0.00983296012878418, 0.009891839981079101, 0.009851103782653808, 0.009885312080383301, 0.009861215591430664, 0.009910335540771484, 0.009914591789245605, 0.009835488319396972, 0.009877856254577638, 0.009834400177001953, 0.009929344177246094, 0.009834752082824708, 0.009954943656921386, 0.009838239669799805, 0.009841055870056152, 0.009830016136169434, 0.009788127899169921, 0.009835583686828613, 0.009785311698913574, 0.00984124755859375, 0.009768992424011231, 0.009850624084472657, 0.009799648284912109, 0.009819552421569825, 0.010087295532226563, 0.009842144012451173, 0.009804320335388184, 0.009795167922973632, 0.009779616355895996, 0.009825375556945801, 0.010572064399719238, 0.013265536308288575, 0.010495776176452637, 0.009938400268554687, 0.009855520248413087, 0.009866496086120606, 0.009823007583618164, 0.00959718418121338, 0.009935199737548828, 0.009860639572143554, 0.009914496421813964, 0.009828351974487304, 0.009793536186218262, 0.00991641616821289, 0.009865216255187988, 0.009873408317565918, 0.009887743949890136, 0.009887743949890136, 0.009973759651184083, 0.010074111938476562, 0.009979968070983886, 0.009965503692626953, 0.009919615745544433, 0.009878399848937989, 0.009861023902893066, 0.00987564754486084, 0.009848735809326171, 0.009885696411132813, 0.009814016342163086, 0.009862208366394043, 0.009872320175170899, 0.009747648239135742, 0.009795968055725097, 0.009869631767272949, 0.009906304359436036, 0.009934847831726074, 0.010026080131530762, 0.009939743995666503, 0.009968768119812012, 0.010019840240478516, 0.0101396484375, 0.010088095664978028, 0.010088735580444336, 0.010030143737792969, 0.009984416007995605, 0.009937151908874512, 0.009976351737976074, 0.00992204761505127, 0.009888256072998047, 0.00987936019897461, 0.009834495544433594, 0.009840895652770996, 0.009884767532348633, 0.00983516788482666, 0.009951231956481933, 0.00991427230834961, 0.010010335922241211, 0.009986207962036133, 0.010010784149169921, 0.010004672050476074, 0.010059552192687988, 0.009930944442749024, 0.009956992149353028, 0.010073920249938965, 0.01005350399017334, 0.010084287643432617, 0.010075072288513183, 0.010046751976013183, 0.01001638412475586, 0.010025792121887207, 0.009690655708312988, 0.010018783569335938, 0.009960255622863769, 0.010053343772888183, 0.009957440376281738, 0.010059743881225587, 0.01009996795654297, 0.01018329620361328, 0.01012339210510254, 0.010123167991638184, 0.010166367530822755, 0.010168319702148437, 0.010145792007446289, 0.010141759872436524, 0.010081536293029785, 0.010014944076538086, 0.00998265552520752, 0.009842656135559081, 0.009916288375854492, 0.010153599739074707, 0.010182975769042969, 0.010069600105285645, 0.010280896186828613, 0.010145440101623535, 0.010197823524475097, 0.01009455966949463, 0.0101048641204834, 0.01005891227722168, 0.010107744216918945, 0.010026335716247559, 0.009987903594970702, 0.00990511989593506, 0.009917632102966308, 0.00987615966796875, 0.00999833583831787, 0.009973504066467286, 0.009924863815307617, 0.009909664154052734, 0.009908576011657715, 0.009847040176391602, 0.009885120391845702, 0.01007260799407959, 0.009881631851196289, 0.009901663780212403, 0.009922975540161133, 0.009930751800537109, 0.009945088386535645, 0.010000384330749512, 0.010071519851684571, 0.0099901123046875, 0.01005615997314453, 0.0100414400100708, 0.011681792259216308, 0.010293439865112304, 0.010350111961364747, 0.010420767784118653, 0.010315520286560059, 0.010378463745117188, 0.010541152000427247, 0.01032431983947754, 0.010314111709594726, 0.010307552337646485, 0.01035654354095459, 0.009958208084106445, 0.01022383975982666, 0.010217184066772462, 0.010540575981140136, 0.010324095726013184, 0.010307807922363282, 0.010330240249633789, 0.01032102394104004, 0.010195551872253418, 0.01031942367553711, 0.01023692798614502, 0.01034937572479248, 0.010135519981384277, 0.01014470386505127, 0.010149888038635254, 0.010059231758117676, 0.010079839706420898, 0.010078944206237794, 0.00998243236541748, 0.010070048332214355, 0.009999808311462402, 0.010107135772705078, 0.00998198413848877, 0.010114751815795898, 0.01101036834716797, 0.010404959678649902, 0.01118835163116455, 0.01003996753692627, 0.010096768379211425, 0.010348544120788575, 0.010207136154174805, 0.010116607666015624, 0.01010371208190918, 0.010158047676086426, 0.01009763240814209, 0.010316703796386719, 0.010350432395935059, 0.010317888259887695, 0.010193023681640625, 0.01027462387084961, 0.009965567588806153, 0.010042688369750977, 0.010007519721984864, 0.009975520133972168, 0.009852288246154785, 0.009939583778381347, 0.009948639869689942, 0.009957056045532226, 0.009851743698120117, 0.010210463523864746, 0.009913311958312989, 0.009926527976989745, 0.009943360328674317, 0.009894720077514648, 0.00983948802947998, 0.009881119728088378, 0.009872896194458008, 0.00997001552581787, 0.009964159965515137, 0.010047264099121093, 0.010073568344116211, 0.009992608070373535, 0.00995475196838379, 0.009736191749572755, 0.010069408416748048, 0.010011136054992676, 0.010172415733337402, 0.010296480178833008, 0.010412351608276367, 0.010320799827575684, 0.010171680450439454, 0.01008499240875244, 0.010004063606262208, 0.010023136138916016, 0.010212960243225098, 0.009865632057189941, 0.009903776168823241, 0.00991049575805664, 0.009969792366027831, 0.009881600379943848, 0.009796607971191406, 0.009792799949645997, 0.009809632301330567, 0.009801792144775391, 0.009767999649047851, 0.009752639770507813, 0.009798399925231934, 0.010280832290649414, 0.0098920316696167, 0.009856351852416992, 0.009790111541748046, 0.009803775787353516, 0.009875455856323241, 0.009797632217407226, 0.009987968444824219, 0.009846912384033204, 0.00981606388092041, 0.009791487693786622, 0.009852928161621094, 0.00987980842590332, 0.009807616233825684, 0.01001471996307373, 0.009775103569030762, 0.009802911758422852, 0.009771424293518067, 0.009803199768066406, 0.009833215713500976, 0.009825887680053711, 0.009839263916015625, 0.009959615707397462, 0.009814911842346192, 0.009830495834350587, 0.00981283187866211, 0.009780287742614746, 0.009751168251037598, 0.009760576248168945, 0.009798144340515137, 0.009844736099243164, 0.00976089572906494, 0.009920384407043457, 0.009911359786987305, 0.009825216293334962, 0.009846879959106445, 0.00994489574432373, 0.009947232246398926, 0.009981951713562011, 0.009586591720581055, 0.009936991691589356, 0.009971808433532715, 0.009973759651184083, 0.009965567588806153, 0.009904288291931152, 0.00998588752746582, 0.009958975791931153, 0.009912032127380371, 0.009917152404785157, 0.009954784393310547, 0.009982144355773926, 0.010090911865234375, 0.00989583969116211, 0.009985119819641113, 0.01002182388305664, 0.009960800170898438, 0.009951231956481933, 0.00992527961730957, 0.009977984428405761, 0.009943103790283203, 0.010172063827514649, 0.010049759864807129, 0.009993887901306153, 0.01000271987915039, 0.010123231887817382, 0.010366815567016602, 0.009945152282714843, 0.01003651237487793, 0.009945631980895996, 0.010010687828063965, 0.010076416015625, 0.010082559585571289, 0.009977791786193848, 0.010041152000427246, 0.009930751800537109, 0.010017024040222169, 0.009959168434143066, 0.009968864440917968, 0.010025152206420898, 0.010009247779846192, 0.009965184211730958, 0.01005190372467041, 0.00999785614013672, 0.00996953582763672, 0.009964287757873535, 0.009999296188354493, 0.009957951545715332, 0.01020963191986084, 0.010037247657775878, 0.010076160430908204, 0.010074272155761718, 0.010032863616943359, 0.010086624145507813, 0.010049728393554688, 0.01006982421875, 0.010056768417358398, 0.010089311599731445, 0.010051839828491211, 0.010041088104248046, 0.010048704147338867, 0.010153087615966797, 0.010026880264282226, 0.009779680252075195, 0.010229984283447265, 0.01021343994140625, 0.010170304298400878, 0.010073344230651855, 0.010033408164978027, 0.009996800422668458, 0.010057727813720703, 0.010012672424316407, 0.010040448188781739, 0.010168800354003906, 0.009985535621643067, 0.010013888359069825, 0.010174176216125489, 0.010280960083007813, 0.010116288185119629, 0.01008518409729004, 0.00996947193145752, 0.010144288063049317, 0.009989983558654785, 0.009981887817382813, 0.00996947193145752, 0.009957056045532226, 0.009927103996276856, 0.00993785572052002, 0.009903103828430175, 0.009922335624694825, 0.009916576385498047, 0.009952832221984863, 0.010025312423706054, 0.009955648422241212, 0.010092063903808594, 0.009941568374633789, 0.009959168434143066, 0.010026528358459472, 0.010041824340820312, 0.00998307228088379, 0.009987071990966797, 0.010039551734924317, 0.010110624313354492, 0.010174015998840331, 0.010096799850463868, 0.010041631698608399, 0.009973759651184083, 0.010051103591918945, 0.010122847557067872, 0.010112992286682129, 0.01010700798034668, 0.01018992042541504, 0.010104512214660645, 0.010210368156433105, 0.010130304336547852, 0.01004963207244873, 0.010090016365051269, 0.010090944290161132, 0.010068256378173828, 0.010272480010986329, 0.010137248039245605, 0.00990447998046875, 0.00992972755432129, 0.009947135925292968, 0.009962207794189453, 0.00992636775970459, 0.009659616470336915, 0.00990287971496582, 0.009897983551025391, 0.009928671836853028, 0.009817279815673829, 0.009830368041992188, 0.009819007873535157, 0.009740608215332031, 0.009842368125915527, 0.009857024192810059, 0.009827584266662597, 0.009788448333740235, 0.009805536270141602, 0.009774592399597168, 0.009938528060913086, 0.010105759620666503, 0.010258560180664062, 0.009937984466552734, 0.00994591999053955, 0.01010812759399414, 0.009948287963867188, 0.009870495796203613, 0.010078495979309082, 0.00993667221069336, 0.010001184463500977, 0.010025952339172363, 0.009907999992370605, 0.00989247989654541, 0.009916223526000977, 0.010693087577819824, 0.010110495567321777, 0.010797087669372558, 0.011152095794677734, 0.010643168449401856, 0.01012451171875, 0.010148639678955079, 0.010169471740722656, 0.010208127975463867, 0.010249919891357422, 0.01011900806427002, 0.010138015747070312, 0.010233920097351074, 0.01012707233428955, 0.010223936080932618, 0.010166080474853515, 0.010301600456237793, 0.01016755199432373, 0.010113375663757324, 0.010069984436035156, 0.010186304092407226, 0.010043807983398437, 0.010044032096862793, 0.010116640090942383, 0.010000800132751465, 0.009969599723815918, 0.00998192024230957, 0.009940447807312012, 0.009947680473327636, 0.01005568027496338, 0.009977855682373048, 0.010038304328918456, 0.01003974437713623, 0.010002976417541503, 0.009648384094238282, 0.010227295875549316, 0.010117216110229492, 0.010055808067321778, 0.01007817554473877, 0.01004748821258545, 0.009969568252563477, 0.010036607742309571, 0.010410240173339844, 0.012155263900756837, 0.011288352012634277, 0.010651391983032226, 0.010070143699645996, 0.010055871963500976, 0.010213536262512207, 0.010379263877868652, 0.010612735748291016, 0.01030784034729004, 0.01034768009185791, 0.010402463912963867, 0.010422207832336427, 0.01023795223236084, 0.010166367530822755, 0.010167360305786133, 0.010134495735168458, 0.010264320373535156, 0.010514752388000488, 0.010232640266418456, 0.010228672027587891, 0.010432191848754883, 0.010406496047973633, 0.01044480037689209, 0.010198816299438477, 0.010166080474853515, 0.010348735809326172, 0.010114720344543457, 0.01031817626953125, 0.01028332805633545, 0.010159808158874512, 0.010197152137756347, 0.01020911979675293, 0.010173600196838379, 0.010132575988769531, 0.010145183563232421, 0.010202816009521484, 0.010084320068359375, 0.010043647766113281, 0.01008841609954834, 0.009973407745361328, 0.009968447685241699, 0.01006287956237793, 0.009987199783325195, 0.010469311714172364, 0.010004287719726562, 0.009936991691589356, 0.010004480361938477, 0.010119168281555176, 0.010288288116455078, 0.010297727584838868, 0.010171072006225586, 0.010130687713623047, 0.010068511962890625, 0.01003110408782959, 0.009707615852355958, 0.010123167991638184, 0.010076160430908204, 0.010195232391357422, 0.010052831649780273, 0.009981535911560058, 0.010019743919372558, 0.010123264312744141, 0.010067968368530274, 0.010024448394775391, 0.01006438446044922, 0.009959424018859863, 0.009953056335449219, 0.010029631614685059, 0.010182016372680664, 0.010113311767578124, 0.009998271942138672, 0.01013152027130127, 0.010057567596435547, 0.010106304168701171, 0.010031071662902832, 0.009974464416503907, 0.009913663864135742, 0.00996224021911621, 0.009977855682373048, 0.010282688140869141, 0.010264896392822265, 0.010178560256958008, 0.010192288398742675, 0.010123680114746094, 0.010102975845336913, 0.010309632301330567, 0.010233856201171876, 0.010194944381713868, 0.01022156810760498, 0.01028003215789795, 0.010249055862426758, 0.010230912208557128, 0.010164640426635741, 0.010094431877136231, 0.010158816337585449, 0.010149855613708497, 0.01017420768737793, 0.010315711975097657, 0.010154623985290528, 0.010159808158874512, 0.010102784156799317, 0.010186880111694337, 0.010162015914916993, 0.010233983993530274, 0.010207263946533204, 0.010339776039123534, 0.010243840217590332, 0.010232159614562988, 0.010223679542541504, 0.010170304298400878, 0.010109248161315918, 0.010104415893554687, 0.010227359771728516, 0.0101046724319458, 0.010439616203308105, 0.01027609634399414, 0.01044159984588623, 0.009976544380187989, 0.010237792015075683, 0.010256511688232422, 0.010419936180114745, 0.010379584312438964, 0.010550559997558594, 0.010304384231567383, 0.010344287872314454, 0.01039568042755127, 0.010283200263977051, 0.01026643180847168, 0.010202752113342286, 0.010147744178771973, 0.010226112365722656, 0.010288352012634278, 0.010314528465270995, 0.010201087951660156, 0.0101910400390625, 0.010270591735839843, 0.010205120086669921, 0.010248224258422851, 0.010226655960083007, 0.010971872329711915, 0.010174752235412598, 0.010110591888427735, 0.010094976425170899, 0.01012268829345703, 0.009974335670471192, 0.010055135726928711, 0.009966015815734863, 0.010018143653869629, 0.010044159889221192, 0.010084351539611817, 0.010036800384521484, 0.010027520179748535, 0.009913311958312989, 0.009974111557006836, 0.01014844799041748, 0.010090784072875976, 0.010149567604064942, 0.010231871604919434, 0.010392736434936524, 0.010314751625061035, 0.01015497589111328, 0.010331007957458496, 0.010320896148681641, 0.010259455680847167, 0.01026153564453125, 0.010345439910888671, 0.010292767524719238, 0.010311327934265136, 0.010279071807861329, 0.01035536003112793, 0.010358624458312989, 0.010423871994018554, 0.01024675178527832, 0.010225664138793946, 0.010098688125610352, 0.010235903739929199, 0.010170368194580079, 0.010110976219177246, 0.010164511680603028, 0.010188223838806153, 0.009944319725036621, 0.010488287925720215, 0.010380831718444825, 0.01035542392730713, 0.010346752166748047, 0.010420000076293946, 0.010294367790222168, 0.010204416275024414, 0.010092191696166992, 0.010035136222839355, 0.01000864028930664, 0.009928095817565917, 0.009942975997924805, 0.009953760147094726, 0.010049856185913086, 0.009955039978027344, 0.009959584236145019, 0.010001888275146484, 0.009970208168029786, 0.009941023826599121, 0.010055904388427734, 0.010137184143066406, 0.010170528411865234, 0.010106271743774414, 0.010056320190429688, 0.010069727897644043, 0.010080672264099122, 0.009998175621032715, 0.009975808143615723, 0.010051360130310058, 0.010092767715454101, 0.010151488304138184, 0.010239935874938965, 0.010401439666748048, 0.010414560317993165, 0.010181280136108398, 0.010233119964599609, 0.010262944221496583, 0.010207263946533204, 0.010553471565246582, 0.010388895988464356, 0.010472991943359374, 0.010353599548339844, 0.010283007621765136, 0.010309344291687011, 0.010209440231323242, 0.010189184188842773, 0.010272512435913086, 0.010291199684143066, 0.010553407669067382, 0.010279935836791992, 0.010269632339477539, 0.010452128410339356, 0.010348575592041016, 0.010249119758605957, 0.010272671699523925, 0.010489055633544922, 0.010276991844177246, 0.010268896102905274, 0.010243552207946777, 0.010173407554626465, 0.010169695854187012, 0.010139936447143555, 0.009957088470458984, 0.010216383934020997, 0.01022156810760498, 0.010188608169555665, 0.010209471702575683, 0.01020041561126709, 0.01024182415008545, 0.01015283203125, 0.010108927726745605, 0.010202848434448242, 0.010403679847717286, 0.010285504341125488, 0.010283200263977051, 0.010321696281433106, 0.010254015922546387, 0.010267007827758788, 0.010247296333312988, 0.010238816261291504, 0.01023795223236084, 0.010160127639770507, 0.010412223815917969, 0.010519424438476562, 0.009982912063598633, 0.009995391845703125, 0.009971776008605958, 0.0099268798828125, 0.009943296432495117, 0.00992086410522461, 0.009904224395751953, 0.00997977638244629, 0.009976096153259277, 0.009964384078979493, 0.01011631965637207, 0.010073599815368652, 0.010031295776367187, 0.010046527862548829, 0.01002387237548828, 0.010090496063232422, 0.010090496063232422, 0.00998140811920166, 0.01003600025177002, 0.010057696342468262, 0.009977631568908692, 0.01002086353302002, 0.009967776298522949, 0.009953120231628417, 0.00987679958343506, 0.009894880294799804, 0.009997983932495118, 0.009982015609741211, 0.009920512199401856, 0.009989472389221191, 0.009924960136413575, 0.009879776000976562, 0.009941087722778321, 0.009861120223999023, 0.009914048194885254, 0.009863167762756348, 0.009879487991333008, 0.009930399894714356, 0.009847455978393555, 0.009875200271606445, 0.009910592079162598, 0.009927935600280761, 0.010320896148681641, 0.010408639907836914, 0.01033625602722168, 0.0101397123336792, 0.01012342357635498, 0.01040886402130127, 0.009882495880126954, 0.009893695831298828, 0.010173824310302735, 0.009904543876647949, 0.009979392051696777, 0.009921440124511719, 0.009943039894104003, 0.009926655769348144, 0.009946240425109863, 0.009927552223205566, 0.009932703971862793, 0.00983356761932373, 0.010013728141784669, 0.00999830436706543, 0.009975968360900879, 0.009983648300170898, 0.009927071571350098, 0.009922335624694825, 0.010004192352294923, 0.009997759819030762, 0.010265695571899413, 0.010173791885375977, 0.01008681583404541, 0.010175616264343261, 0.010047552108764648, 0.01007481575012207, 0.010160032272338868, 0.010088352203369141, 0.010101056098937989, 0.010162176132202149, 0.010196991920471191, 0.010112544059753419, 0.010031776428222656, 0.010000191688537597, 0.009996288299560547, 0.009928704261779785, 0.009963520050048828, 0.009957056045532226, 0.010084735870361327, 0.009988032341003417, 0.009965151786804198, 0.010004511833190918, 0.009938336372375489, 0.009948127746582032, 0.010036319732666015, 0.009919391632080079, 0.009904128074645996, 0.009862239837646485, 0.010056608200073243, 0.01008841609954834, 0.009953439712524414, 0.009928288459777833, 0.010012351989746093, 0.009976415634155274, 0.010035167694091797, 0.010067551612854005]",tokens/s,99.06718928327761,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 1013, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 839, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 467, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gptj/modeling_gptj.py"", line 397, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 199, in _flash_attention_forward attn_output_unpad = flash_attn_varlen_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 1124, in flash_attn_varlen_func return FlashAttnVarlenFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 620, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_varlen_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 90, in _flash_attn_varlen_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.varlen_fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3798, in from_pretrained model = cls(config, *model_args, **model_kwargs) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ self.model = DeciCoderModel(config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ self.self_attn = DeciCoderAttention(config=config) File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ self._init_rope() File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1709, in __getattr__ raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4357.275648,4562.28864,0.0,4183.81824,4182.069248,s,1,10.0476240234375,10.0476240234375,0.0,10.0476240234375,10.0476240234375,10.0476240234375,10.0476240234375,[10.0476240234375],,kWh,9.00516337291691e-05,9.925974722610089e-06,2.9063078805999232e-05,0.00012904068725777844,,MB,1646.116864,4715.380736,0.0,4307.550208,4281.174016,s,10,4.240579620361328,0.4240579620361328,0.00973513499326288,0.4238329620361328,0.43182900390625,0.439088330078125,0.444895791015625,"[0.4058084716796875, 0.418193603515625, 0.42089401245117186, 0.44634765625, 0.41901727294921876, 0.42609442138671877, 0.4232594299316406, 0.4263424377441406, 0.4302158203125, 0.424406494140625]",tokens/s,603.6910585779474,kWh,1.2225260772569512e-05,1.3473687572735923e-06,8.141916235750024e-06,2.1714545765593125e-05,tokens/kWh,11789332.49460986,MB,1650.315264,4730.0608,0.0,4322.230272,4281.176576,s,10,25.663558593750004,2.5663558593749998,0.012182986936632783,2.57035546875,2.5787344970703123,2.580027038574219,2.5810610717773437,"[2.57439501953125, 2.548342529296875, 2.56631591796875, 2.55032080078125, 2.564474853515625, 2.54906787109375, 2.581319580078125, 2.574916259765625, 2.57595849609375, 2.578447265625]",tokens/s,24.54842720656159,kWh,7.52577669924302e-05,8.299891720148061e-06,4.9471185410250034e-05,0.00013302884412282824,tokens/kWh,473581.5034356821,,s,630,25.659648288726828,0.04072960045829652,0.0006686577631998439,0.04065139198303223,0.04114664154052734,0.0414242654800415,0.043639575233459474,"[0.04099299240112304, 0.04076748657226562, 0.0403513298034668, 0.040348064422607424, 0.040390655517578124, 0.04044976043701172, 0.04084288024902344, 0.040583839416503904, 0.040483905792236326, 0.04080841445922852, 0.040733566284179686, 0.04058486557006836, 0.04051555252075195, 0.04062460708618164, 0.04030054473876953, 0.040328670501708984, 0.041082431793212894, 0.040731616973876957, 0.04061721420288086, 0.04047743988037109, 0.040271713256835935, 0.04044809722900391, 0.04028422546386719, 0.04057843017578125, 0.040628864288330076, 0.04068460845947266, 0.04068038558959961, 0.040687614440917966, 0.040904705047607424, 0.04079372787475586, 0.040857982635498044, 0.04057702255249023, 0.04013804626464844, 0.040585536956787106, 0.04038899230957031, 0.04036812973022461, 0.044386302947998044, 0.040521728515625, 0.04073174285888672, 0.04049382400512695, 0.04045225524902344, 0.04076544189453125, 0.040882080078125, 0.04057097625732422, 0.040286209106445314, 0.04068924713134765, 0.0414744644165039, 0.040427520751953126, 0.04086579132080078, 0.04296636962890625, 0.04254991912841797, 0.04094976043701172, 0.04068153762817383, 0.04109436798095703, 0.04098735809326172, 0.04190127944946289, 0.04121049499511719, 0.0405667839050293, 0.041056190490722656, 0.04073494338989258, 0.043607585906982424, 0.040876510620117185, 0.04141670227050781, 0.040854686737060546, 0.04049292755126953, 0.040171966552734376, 0.04034000015258789, 0.03998310470581055, 0.040302593231201174, 0.04025139236450195, 0.040443904876708986, 0.039964672088623046, 0.04060160064697266, 0.04005449676513672, 0.04046169662475586, 0.04043571090698242, 0.040393630981445314, 0.04053401565551758, 0.04067327880859375, 0.04053561782836914, 0.04061743927001953, 0.04064969635009766, 0.040687614440917966, 0.04075929641723633, 0.04046950531005859, 0.040313152313232424, 0.04031900787353516, 0.04016604614257813, 0.040122207641601564, 0.04009795379638672, 0.04019200134277344, 0.040136703491210936, 0.040385601043701175, 0.04028067016601562, 0.04058534240722656, 0.04008572769165039, 0.04019171142578125, 0.040245025634765626, 0.04042598342895508, 0.04023651123046875, 0.04015977478027344, 0.04037129592895508, 0.0407213134765625, 0.04067327880859375, 0.041166847229003906, 0.040754814147949216, 0.04069209671020508, 0.040541248321533205, 0.040721343994140624, 0.04042918395996094, 0.040647041320800784, 0.04050124740600586, 0.04046761703491211, 0.040309600830078125, 0.04012236785888672, 0.04023455810546875, 0.04039081573486328, 0.0405261116027832, 0.04043881607055664, 0.040522720336914064, 0.040687614440917966, 0.04068966293334961, 0.04067728042602539, 0.040742687225341793, 0.040640064239501957, 0.040653568267822265, 0.04099398422241211, 0.04044800186157226, 0.04026655960083008, 0.04081011199951172, 0.040527328491210934, 0.04061385726928711, 0.04032198333740234, 0.04034956741333008, 0.04076300811767578, 0.04047513580322266, 0.04068556976318359, 0.04076748657226562, 0.04049100875854492, 0.04095795059204101, 0.04068124771118164, 0.04091507339477539, 0.04103177642822266, 0.04173551940917969, 0.040768161773681644, 0.040687614440917966, 0.04060979080200195, 0.04059340667724609, 0.04070348739624023, 0.04064230346679688, 0.04059734344482422, 0.04052470397949219, 0.0408675537109375, 0.04050758361816406, 0.040502750396728515, 0.0406165771484375, 0.040574977874755856, 0.04029849624633789, 0.04055219268798828, 0.040470592498779295, 0.0402413444519043, 0.04036198425292969, 0.04030809783935547, 0.04056537628173828, 0.040531967163085936, 0.040529918670654294, 0.040650718688964846, 0.04053763198852539, 0.04080275344848633, 0.04061916732788086, 0.040428447723388675, 0.04046438217163086, 0.040542209625244144, 0.04080361557006836, 0.040780513763427735, 0.040515583038330076, 0.04065033721923828, 0.04058563232421875, 0.0456736946105957, 0.04099555206298828, 0.040759361267089844, 0.04069375991821289, 0.04068966293334961, 0.040769535064697264, 0.040837120056152344, 0.04097359848022461, 0.04070883178710937, 0.04064460754394531, 0.04068486404418945, 0.04122544097900391, 0.04059017562866211, 0.04028416061401367, 0.040118270874023435, 0.03983740615844727, 0.040118560791015626, 0.040114177703857425, 0.04040003204345703, 0.040219009399414064, 0.040522144317626956, 0.044087360382080075, 0.04056883239746094, 0.0404128303527832, 0.04004079818725586, 0.039798782348632815, 0.039847934722900394, 0.04000153732299805, 0.039907329559326174, 0.040019168853759765, 0.039772544860839844, 0.039874561309814455, 0.03970294570922851, 0.03973222351074219, 0.03986735916137695, 0.03997494506835937, 0.039858177185058595, 0.03974553680419922, 0.039923583984375, 0.03997708892822266, 0.03990300750732422, 0.03969660949707031, 0.03993971252441406, 0.03968447875976563, 0.03996604919433594, 0.04004726409912109, 0.040288257598876956, 0.04018380737304687, 0.039858081817626956, 0.04016537475585937, 0.04063036727905273, 0.04063961410522461, 0.040573406219482425, 0.04217897415161133, 0.04079526519775391, 0.04056972885131836, 0.04063772964477539, 0.040677921295166015, 0.04031507110595703, 0.04016742324829101, 0.04041849517822266, 0.040756031036376955, 0.040316001892089844, 0.04020316696166992, 0.04023471832275391, 0.04056092834472656, 0.04045004653930664, 0.04056063842773437, 0.04100505447387695, 0.04126271820068359, 0.041216129302978514, 0.0410994873046875, 0.04076752090454101, 0.04764665603637695, 0.0408870735168457, 0.0409354248046875, 0.044077056884765625, 0.040892414093017575, 0.04066918563842774, 0.040768959045410155, 0.04060422515869141, 0.04060774230957031, 0.040515583038330076, 0.040359584808349606, 0.04047407913208008, 0.0401847038269043, 0.04207820892333984, 0.04047872161865235, 0.04040703964233398, 0.04056790542602539, 0.04019292831420898, 0.040298336029052736, 0.04082854461669922, 0.04355945587158203, 0.04052550506591797, 0.04044800186157226, 0.04027628707885742, 0.040307807922363284, 0.04035676956176758, 0.04023436737060547, 0.04010841751098633, 0.040048736572265625, 0.04021878433227539, 0.04014460754394531, 0.04010604858398437, 0.04014435195922852, 0.040686336517333985, 0.040738815307617186, 0.04076508712768555, 0.04229974365234375, 0.04100710296630859, 0.040949825286865235, 0.04104595184326172, 0.04077699279785156, 0.04058595275878906, 0.04066099166870117, 0.04074905776977539, 0.04057692718505859, 0.040597599029541014, 0.04060160064697266, 0.040683135986328126, 0.04040073776245117, 0.040280609130859374, 0.040787967681884765, 0.04067932891845703, 0.04069116973876953, 0.04039539337158203, 0.040431167602539064, 0.040470977783203126, 0.04036713409423828, 0.04043056106567383, 0.04029206466674805, 0.040519966125488284, 0.04089782333374024, 0.04095049667358398, 0.04093743896484375, 0.04056659317016602, 0.04129014587402344, 0.04072652816772461, 0.040720382690429685, 0.040738815307617186, 0.04065894317626953, 0.040795551300048825, 0.04112847900390625, 0.040642623901367185, 0.040643840789794924, 0.040825599670410155, 0.040736766815185545, 0.04103987121582031, 0.040420574188232423, 0.04053433609008789, 0.040444385528564455, 0.04039833450317383, 0.040544063568115234, 0.04047123336791992, 0.04053942489624023, 0.040544319152832034, 0.04041779327392578, 0.04059337615966797, 0.04062432098388672, 0.04055622482299805, 0.04060150527954102, 0.040547870635986326, 0.04041558456420898, 0.04052431869506836, 0.040320510864257815, 0.040405502319335936, 0.04021247863769531, 0.03996057510375976, 0.04013596725463867, 0.04001251220703125, 0.04041318511962891, 0.040019966125488284, 0.04008345413208008, 0.040383937835693356, 0.04028473663330078, 0.04034307098388672, 0.04022630310058594, 0.040285152435302736, 0.040352832794189455, 0.04075411224365234, 0.04010531234741211, 0.04061983871459961, 0.03988966369628906, 0.040185951232910154, 0.040030208587646485, 0.04005068969726563, 0.04032716751098633, 0.0401899528503418, 0.04045388793945313, 0.04029254531860352, 0.040321086883544924, 0.04050124740600586, 0.040204032897949216, 0.040081375122070315, 0.04038684844970703, 0.040564735412597655, 0.040836673736572265, 0.04063686370849609, 0.04072998428344726, 0.04122224044799805, 0.04087174224853515, 0.04069660949707031, 0.040787967681884765, 0.04343804931640625, 0.04214700698852539, 0.04054307174682617, 0.040537792205810545, 0.040487232208251955, 0.04087398529052735, 0.04053305435180664, 0.0409804801940918, 0.0404284782409668, 0.04140415954589844, 0.040936958312988284, 0.041293567657470706, 0.04073497772216797, 0.04078054428100586, 0.04114636611938476, 0.04106668853759766, 0.04201852798461914, 0.04207388687133789, 0.04095827102661133, 0.04076697540283203, 0.040540672302246096, 0.040574943542480466, 0.04081462478637695, 0.04078182220458984, 0.0406418571472168, 0.040468544006347654, 0.040604286193847657, 0.04063987350463867, 0.04069807815551758, 0.04055231857299805, 0.04050998306274414, 0.04066707229614258, 0.04618656158447266, 0.041025535583496094, 0.040567806243896484, 0.040399360656738284, 0.04029087829589844, 0.0422808952331543, 0.040648704528808595, 0.04089993667602539, 0.0407374382019043, 0.040986625671386716, 0.040546302795410154, 0.04086067199707031, 0.040799232482910154, 0.040854686737060546, 0.04072739028930664, 0.04140031814575195, 0.04088761520385742, 0.04054822540283203, 0.040690494537353517, 0.04067737579345703, 0.04038246536254883, 0.04083907318115235, 0.04056892776489258, 0.04052896118164063, 0.040840129852294925, 0.04074086380004883, 0.040836639404296875, 0.04127043151855469, 0.04026230239868164, 0.04004617691040039, 0.04057148742675781, 0.04025328063964844, 0.04026377487182617, 0.04096006393432617, 0.04070367813110352, 0.040955520629882815, 0.040904705047607424, 0.04114912033081055, 0.04098777770996094, 0.04095616149902344, 0.04092911911010742, 0.04118403244018555, 0.04270284652709961, 0.04107385635375976, 0.04081903839111328, 0.04087424087524414, 0.04071855926513672, 0.04085097503662109, 0.04062460708618164, 0.040595169067382815, 0.04101705551147461, 0.04064716720581055, 0.040745025634765626, 0.041223838806152345, 0.041655937194824216, 0.04113248062133789, 0.041070655822753904, 0.04078976058959961, 0.04148681640625, 0.04094473648071289, 0.04082575988769531, 0.04088576126098633, 0.0408089599609375, 0.04131430435180664, 0.04141993713378906, 0.04292204666137695, 0.04059011077880859, 0.04065206527709961, 0.040893150329589845, 0.04141424179077149, 0.040610206604003905, 0.040376319885253906, 0.04057907104492187, 0.04045759963989258, 0.040645248413085935, 0.040776927947998046, 0.040715038299560545, 0.040476673126220705, 0.04062412643432617, 0.0405852165222168, 0.04048076629638672, 0.040531967163085936, 0.040720382690429685, 0.04074291229248047, 0.040905887603759766, 0.04070896148681641, 0.040709983825683596, 0.04045225524902344, 0.040720382690429685, 0.0405931510925293, 0.04134003067016601, 0.040792831420898436, 0.04067136001586914, 0.04076892852783203, 0.04091494369506836, 0.0404376335144043, 0.04088905715942383, 0.04075462341308594, 0.040722782135009766, 0.0406036491394043, 0.040623870849609375, 0.04082710266113281, 0.04079232025146484, 0.040613056182861325, 0.04066387176513672, 0.040892353057861326, 0.04120787048339844, 0.041059486389160155, 0.041427806854248045, 0.04118233489990234, 0.040768383026123046, 0.040761119842529295, 0.0414681282043457, 0.04094486236572266, 0.04093417739868164, 0.040787967681884765, 0.04103577423095703, 0.041084129333496096, 0.041175838470458984, 0.040787582397460935, 0.04093705749511719, 0.041401119232177735, 0.040632320404052735, 0.04073583984375, 0.04072259140014649, 0.04100771331787109, 0.04110316848754883, 0.04105865478515625, 0.04094748687744141, 0.04100527954101563, 0.04057702255249023, 0.04117913436889648, 0.041027454376220705, 0.04105023956298828, 0.04093952178955078, 0.04084326553344726, 0.04074700927734375, 0.04077881622314453, 0.04068652725219726, 0.04084051132202148, 0.04099142456054688, 0.040812545776367185, 0.04061798477172852, 0.040764480590820315, 0.04083808135986328, 0.040853023529052734, 0.0406258544921875, 0.04068022537231445, 0.04072652816772461, 0.04070713424682617, 0.04085036849975586, 0.041095169067382815, 0.04085747146606445, 0.04129443359375, 0.04072857666015625, 0.04048486328125, 0.040796161651611325, 0.0406036491394043, 0.040783199310302734, 0.04070467376708985, 0.040748638153076173, 0.04072284698486328, 0.04085964965820312, 0.04078387069702148, 0.04085075378417969, 0.04069036865234375, 0.04116070556640625, 0.04066089630126953, 0.040581214904785154, 0.040897567749023436, 0.040653087615966796, 0.04106924819946289, 0.04173587036132813, 0.04215225601196289, 0.04080543899536133, 0.04059996795654297, 0.04066678237915039, 0.04084415817260742, 0.04085760116577149, 0.04125491333007812, 0.040690719604492186, 0.040778720855712894, 0.040738815307617186, 0.04037587356567383, 0.040673023223876954, 0.04062627029418946, 0.04074911880493164, 0.04069839859008789, 0.0406828498840332, 0.04088611221313477, 0.04094806289672852, 0.040914913177490235, 0.04126092910766602, 0.041097854614257814, 0.041455615997314454, 0.040963680267333984, 0.04129219055175781, 0.04134092712402344, 0.04219638442993164, 0.040944225311279295, 0.040630271911621094, 0.04078579330444336, 0.04084953689575195, 0.041284862518310546, 0.04091494369506836, 0.040794017791748044, 0.04064547348022461, 0.04052374267578125, 0.040554527282714845, 0.040564735412597655, 0.040417278289794925, 0.04067737579345703, 0.04070809555053711, 0.04365264129638672, 0.04082457733154297, 0.0409463996887207]",tokens/s,24.552168171252035,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 1141, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 944, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 677, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/llama/modeling_llama.py"", line 500, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 989, in forward outputs = self.gpt_neox( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 880, in forward outputs = layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 647, in forward attention_layer_outputs = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neox/modeling_gpt_neox.py"", line 369, in forward attn_weights = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 1054, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 856, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 596, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/qwen2/modeling_qwen2.py"", line 426, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 551, in from_pretrained model_class = get_class_from_dynamic_module( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 502, in get_class_from_dynamic_module final_module = get_cached_module_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 327, in get_cached_module_file modules_needed = check_imports(resolved_module_file) File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 182, in check_imports raise ImportError( ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,,,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 96, in run self.run_text_generation_memory_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 200, in run_text_generation_memory_tracking _ = backend.prefill(self.inputs, prefill_kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 426, in prefill return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1989, in generate result = self._sample( File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2932, in _sample outputs = self(**model_inputs, return_dict=True) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 864, in forward transformer_outputs = self.transformer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 732, in forward outputs = block( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 459, in forward attn_outputs = self.attn( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 411, in forward return self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/gpt_neo/modeling_gpt_neo.py"", line 358, in forward attn_output = _flash_attention_forward( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_flash_attention_utils.py"", line 214, in _flash_attention_forward attn_output = flash_attn_func( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 880, in flash_attn_func return FlashAttnFunc.apply( File ""/usr/local/lib/python3.10/dist-packages/torch/autograd/function.py"", line 598, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 546, in forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = _flash_attn_forward( File ""/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py"", line 52, in _flash_attn_forward out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( RuntimeError: FlashAttention only supports Ampere GPUs or newer. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model self.load_transformers_model_with_no_weights() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 133, in load_transformers_model_with_no_weights self.load_transformers_model_from_pretrained() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 85, in load_transformers_model_from_pretrained self.pretrained_model = self.automodel_loader.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained return model_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 4000, in from_pretrained dispatch_model(model, **device_map_kwargs) File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 494, in dispatch_model model.to(device) File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2871, in to return super().to(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to return self._apply(convert) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply module._apply(fn) [Previous line repeated 2 more times] File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply self._buffers[key] = fn(buf) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert return t.to( torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 70.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4203.692032,5925.43744,0.0,5530.189824,5138.859008,s,1,12.0731376953125,12.0731376953125,0.0,12.0731376953125,12.0731376953125,12.0731376953125,12.0731376953125,[12.0731376953125],,kWh,0.00014106460684169329,1.555255706320597e-05,6.053282620399614e-05,0.0002171499901088954,,MB,1346.072576,5944.311808,0.0,5534.384128,4845.009408,s,10,2.1223248291015624,0.21223248291015628,0.0006155042745064004,0.2121497802734375,0.2128214599609375,0.21313273010253905,0.21338174621582032,"[0.21275228881835936, 0.21145152282714844, 0.21251513671875, 0.2122291259765625, 0.21344400024414062, 0.2127508087158203, 0.2120704345703125, 0.21166184997558593, 0.21206002807617189, 0.21138963317871093]",tokens/s,1206.2244030211516,kWh,6.256342691046083e-06,6.899540493929816e-07,4.159814202554111e-06,1.1106110942993175e-05,tokens/kWh,23050373.016623784,MB,1355.706368,5944.311808,0.0,5534.384128,5014.504448,s,10,19.151216674804687,1.9151216674804687,0.007566801084310754,1.9153659057617187,1.9236949584960936,1.9251369689941407,1.926290577392578,"[1.9125977783203125, 1.918134033203125, 1.92337451171875, 1.9195284423828125, 1.922497802734375, 1.9265789794921875, 1.902783447265625, 1.9088680419921875, 1.9096376953125, 1.9072159423828126]",tokens/s,32.896082306291646,kWh,5.530638489020411e-05,6.10016108972048e-06,3.6190023632844804e-05,9.759656961276938e-05,tokens/kWh,645514.4914412769,,s,630,19.147601362228393,0.030393018035283163,0.00043964754739872025,0.030350192070007324,0.030686198234558107,0.030910873889923095,0.032269759674072265,"[0.030111743927001954, 0.03027020835876465, 0.03071379280090332, 0.03030841636657715, 0.030068735122680663, 0.029963552474975588, 0.030040800094604494, 0.02991836738586426, 0.029935903549194336, 0.029993152618408202, 0.029962623596191406, 0.029997055053710937, 0.030203903198242187, 0.029853696823120116, 0.029997055053710937, 0.0299233283996582, 0.029970016479492188, 0.03009987258911133, 0.029996768951416015, 0.029968671798706055, 0.030044160842895507, 0.02992259216308594, 0.02990358352661133, 0.030365695953369142, 0.030026815414428712, 0.029924287796020507, 0.029881696701049804, 0.02996291160583496, 0.029847551345825195, 0.03037593650817871, 0.03012326431274414, 0.03005721664428711, 0.03009324836730957, 0.029945440292358398, 0.03094099235534668, 0.030613279342651366, 0.030752767562866212, 0.030438047409057617, 0.030339296340942384, 0.030453760147094725, 0.03043328094482422, 0.03059097671508789, 0.030914560317993164, 0.030317760467529296, 0.03035219192504883, 0.030464000701904297, 0.032626686096191404, 0.03220684814453125, 0.030823871612548827, 0.030476863861083985, 0.03043328094482422, 0.030664831161499023, 0.030436735153198242, 0.03016329574584961, 0.0300501766204834, 0.030085407257080077, 0.030236671447753907, 0.03025712013244629, 0.030279647827148436, 0.030145631790161134, 0.032142303466796876, 0.03215520095825195, 0.030665151596069334, 0.031100128173828127, 0.030376031875610353, 0.030333568572998047, 0.03298099136352539, 0.03464191818237305, 0.030572256088256835, 0.030384416580200194, 0.03064944076538086, 0.030303136825561523, 0.030513151168823242, 0.030337024688720703, 0.03073023986816406, 0.03046735954284668, 0.030585567474365236, 0.03040208053588867, 0.030433759689331055, 0.030470144271850585, 0.030676992416381835, 0.030463903427124024, 0.030350976943969727, 0.03015113639831543, 0.03026063919067383, 0.030040672302246094, 0.03006195259094238, 0.030019264221191406, 0.03032159996032715, 0.029902847290039062, 0.030099456787109374, 0.030369760513305664, 0.030090816497802736, 0.02996067237854004, 0.03039961624145508, 0.029934080123901367, 0.030137887954711916, 0.03006265640258789, 0.030348064422607422, 0.030129695892333986, 0.030308191299438476, 0.0302475528717041, 0.030225887298583984, 0.030269983291625977, 0.030364959716796876, 0.030122623443603516, 0.030242912292480467, 0.030113023757934572, 0.030255199432373047, 0.030081695556640625, 0.030390272140502928, 0.030162784576416017, 0.031088800430297853, 0.030289920806884765, 0.030410751342773438, 0.03058483123779297, 0.03037183952331543, 0.030263263702392577, 0.030572576522827147, 0.030259103775024415, 0.03019759941101074, 0.030613759994506835, 0.03020595169067383, 0.03015852737426758, 0.030529216766357423, 0.030356096267700194, 0.03056928062438965, 0.03031180763244629, 0.03024140739440918, 0.03038969612121582, 0.030655040740966796, 0.030468095779418947, 0.030488576889038086, 0.030186592102050783, 0.030479263305664063, 0.030461727142333986, 0.03057072067260742, 0.030486528396606444, 0.03033907127380371, 0.030273536682128906, 0.030465599060058593, 0.030517696380615234, 0.030280927658081054, 0.030517791748046873, 0.03057481575012207, 0.03056630325317383, 0.030379167556762697, 0.03073302459716797, 0.03033113670349121, 0.030498815536499024, 0.030693376541137695, 0.030467391967773438, 0.030349407196044922, 0.030277408599853516, 0.030374496459960938, 0.030388448715209963, 0.030449663162231445, 0.03056025505065918, 0.030265344619750976, 0.030341119766235353, 0.030308351516723633, 0.03054080009460449, 0.030527679443359376, 0.030509536743164062, 0.03051136016845703, 0.0305828800201416, 0.030494720458984374, 0.03044918441772461, 0.030603744506835937, 0.030301664352416994, 0.030540063858032228, 0.030445823669433592, 0.03073023986816406, 0.030957536697387697, 0.030633087158203124, 0.0305447998046875, 0.030683008193969727, 0.030642303466796875, 0.030565759658813477, 0.03045235252380371, 0.030962976455688476, 0.03043190383911133, 0.030562368392944336, 0.030648319244384766, 0.030541696548461915, 0.03199180793762207, 0.030679168701171874, 0.030472192764282226, 0.030715904235839843, 0.03075984001159668, 0.03059712028503418, 0.030496768951416016, 0.030487936019897462, 0.03031923294067383, 0.030514976501464845, 0.03051968002319336, 0.030489824295043946, 0.03124083137512207, 0.03050271987915039, 0.030428800582885742, 0.030476512908935546, 0.03075926399230957, 0.030524768829345704, 0.030716575622558594, 0.030598560333251954, 0.030528095245361327, 0.030494144439697266, 0.030455968856811524, 0.03044598388671875, 0.03057663917541504, 0.030512992858886718, 0.03046620750427246, 0.030471839904785157, 0.030439231872558595, 0.0304051513671875, 0.030347488403320313, 0.0303470401763916, 0.03042889595031738, 0.030332704544067383, 0.030433792114257813, 0.030316511154174806, 0.03055619239807129, 0.030390272140502928, 0.030312448501586913, 0.03035955238342285, 0.030373664855957033, 0.03033830451965332, 0.030558847427368165, 0.03045984077453613, 0.030433696746826173, 0.030516799926757814, 0.030269887924194334, 0.030166816711425782, 0.030505184173583985, 0.030373664855957033, 0.030480607986450196, 0.030416448593139647, 0.030368160247802735, 0.03035753631591797, 0.030810111999511718, 0.030310400009155275, 0.030365695953369142, 0.03040377616882324, 0.030417728424072265, 0.03061350440979004, 0.030361600875854492, 0.030398464202880858, 0.03031603240966797, 0.030466560363769532, 0.030402559280395508, 0.030273120880126955, 0.030325151443481444, 0.030507007598876954, 0.030537727355957032, 0.030351327896118163, 0.030394399642944336, 0.03010748863220215, 0.030150815963745116, 0.030119935989379884, 0.030257152557373046, 0.03213478469848633, 0.030421375274658203, 0.030488576889038086, 0.030382080078125, 0.03051215934753418, 0.030374399185180666, 0.030853599548339845, 0.032107967376708985, 0.030583040237426758, 0.030409023284912108, 0.030296064376831053, 0.030485792160034178, 0.030391008377075195, 0.030265344619750976, 0.030251007080078125, 0.030310144424438478, 0.030330528259277345, 0.030099103927612305, 0.030154783248901366, 0.02994451141357422, 0.030168544769287108, 0.030397184371948244, 0.030201343536376952, 0.030189088821411133, 0.030267488479614257, 0.03003481674194336, 0.03074662399291992, 0.03017318344116211, 0.030287872314453124, 0.030224384307861327, 0.030311424255371092, 0.030409568786621093, 0.0302873592376709, 0.03179952049255371, 0.030422847747802736, 0.030447263717651368, 0.03040470314025879, 0.030659231185913086, 0.0305296630859375, 0.030359647750854493, 0.030535680770874023, 0.030402559280395508, 0.03208396911621094, 0.030612512588500975, 0.0306428165435791, 0.030719648361206053, 0.030622047424316408, 0.030695039749145506, 0.030767839431762697, 0.030653472900390624, 0.03061356735229492, 0.03075142478942871, 0.030472415924072266, 0.030642175674438478, 0.03036774444580078, 0.030435392379760742, 0.030281728744506835, 0.030298112869262695, 0.030162464141845702, 0.030256959915161134, 0.03004182434082031, 0.03035545539855957, 0.030292224884033204, 0.030651071548461913, 0.030357503890991212, 0.03013599967956543, 0.03015488052368164, 0.030148799896240235, 0.030093311309814453, 0.030363616943359376, 0.030388191223144533, 0.030367807388305666, 0.030338144302368163, 0.030413728713989258, 0.03038617515563965, 0.03061459159851074, 0.030147008895874024, 0.03045743942260742, 0.030383007049560547, 0.032734783172607425, 0.03169148826599121, 0.030531295776367186, 0.030488576889038086, 0.03058278465270996, 0.030636032104492186, 0.03043328094482422, 0.03022355270385742, 0.030589183807373046, 0.030279487609863282, 0.03058127975463867, 0.03035772705078125, 0.03110019111633301, 0.030381792068481444, 0.031214591979980468, 0.03341299057006836, 0.030363775253295897, 0.030689279556274415, 0.030549760818481445, 0.030476543426513673, 0.03040480041503906, 0.03111302375793457, 0.030889951705932617, 0.03028995132446289, 0.030504959106445313, 0.030463008880615233, 0.03046067237854004, 0.030439008712768556, 0.032295455932617186, 0.03132425689697266, 0.030584735870361326, 0.03040835189819336, 0.03035385513305664, 0.030187519073486328, 0.03016908836364746, 0.030121984481811522, 0.030240480422973632, 0.03037392044067383, 0.030737695693969728, 0.030601184844970705, 0.03056057548522949, 0.030744512557983397, 0.03058937644958496, 0.030879072189331055, 0.030685855865478517, 0.030629888534545898, 0.030621696472167968, 0.030549375534057618, 0.030614080429077147, 0.03070572853088379, 0.030707712173461913, 0.030502912521362304, 0.030300096511840822, 0.03045792007446289, 0.030325759887695314, 0.030354080200195314, 0.030138303756713867, 0.03027395248413086, 0.029941024780273436, 0.03006892776489258, 0.02993779182434082, 0.030180927276611327, 0.029942623138427736, 0.029995008468627928, 0.029981887817382813, 0.030022464752197265, 0.03015065574645996, 0.030127519607543944, 0.030013408660888672, 0.03005299186706543, 0.030041919708251954, 0.02998700714111328, 0.029913087844848633, 0.02999075126647949, 0.029989023208618164, 0.030013439178466796, 0.02984137535095215, 0.030012895584106445, 0.030450239181518554, 0.029998336791992188, 0.030595584869384764, 0.030294271469116212, 0.02992742347717285, 0.029977983474731445, 0.029700191497802734, 0.029899295806884767, 0.030192800521850586, 0.029997919082641603, 0.029870080947875976, 0.02998681640625, 0.03001740837097168, 0.030025856018066406, 0.030323808670043945, 0.030186336517333986, 0.029966400146484374, 0.030269439697265626, 0.02998886489868164, 0.030216192245483397, 0.029929471969604493, 0.03000115203857422, 0.030109407424926758, 0.030103519439697267, 0.03058278465270996, 0.030396127700805665, 0.030263584136962892, 0.03027916717529297, 0.03021670341491699, 0.030006303787231445, 0.03005721664428711, 0.030121984481811522, 0.030309823989868163, 0.03015555191040039, 0.03029747200012207, 0.030239360809326172, 0.030218271255493163, 0.030117855072021485, 0.03035545539855957, 0.030353279113769532, 0.030174911499023436, 0.030254592895507814, 0.030229440689086916, 0.03030406379699707, 0.030234783172607423, 0.03050048065185547, 0.030351392745971678, 0.030359935760498048, 0.03153494453430176, 0.03038419151306152, 0.03043132781982422, 0.030774911880493163, 0.030474143981933592, 0.030265567779541015, 0.030646528244018555, 0.030522687911987305, 0.03060505676269531, 0.031136703491210938, 0.030646047592163085, 0.030492895126342772, 0.030332927703857423, 0.03118489646911621, 0.030777183532714844, 0.030232128143310548, 0.03016969680786133, 0.030136320114135744, 0.030135616302490235, 0.0309418888092041, 0.03007279968261719, 0.030005279541015624, 0.029941631317138673, 0.029935743331909178, 0.02998182487487793, 0.029864095687866212, 0.02985443115234375, 0.02983526420593262, 0.029890560150146486, 0.030318592071533205, 0.030068735122680663, 0.0299683837890625, 0.030003200531005858, 0.030011392593383788, 0.03036947250366211, 0.03003004837036133, 0.030107135772705077, 0.029987424850463868, 0.03007036781311035, 0.03045327949523926, 0.030236671447753907, 0.030155231475830078, 0.03245859146118164, 0.03033660888671875, 0.030024255752563477, 0.030182655334472657, 0.030153120040893554, 0.03018544006347656, 0.030271232604980467, 0.030103168487548827, 0.030098432540893554, 0.03013222312927246, 0.030058496475219725, 0.030325759887695314, 0.03008358383178711, 0.030177791595458983, 0.03039641571044922, 0.030144512176513674, 0.030094688415527343, 0.030398719787597655, 0.030152416229248045, 0.030263168334960937, 0.03021238327026367, 0.030152767181396485, 0.030220767974853516, 0.030274688720703127, 0.03040656089782715, 0.030340063095092774, 0.030588287353515625, 0.030394016265869142, 0.03026019287109375, 0.030478271484375, 0.03053984069824219, 0.03086329650878906, 0.030373760223388672, 0.030612895965576172, 0.030649120330810548, 0.030397663116455077, 0.03019024085998535, 0.030354848861694338, 0.030249696731567383, 0.030195711135864257, 0.03016703987121582, 0.030170848846435547, 0.030165056228637695, 0.03015292739868164, 0.030208000183105467, 0.0301711368560791, 0.030066688537597655, 0.030289920806884765, 0.030162399291992187, 0.030050559997558592, 0.030440864562988282, 0.02984815979003906, 0.029804832458496095, 0.029726720809936522, 0.031883264541625975, 0.03082035255432129, 0.030238719940185548, 0.03007279968261719, 0.03004128074645996, 0.030151519775390625, 0.030527488708496094, 0.030224384307861327, 0.030353408813476562, 0.030468095779418947, 0.030312448501586913, 0.030200895309448243, 0.03041289520263672, 0.030544736862182616, 0.03042483139038086, 0.030314559936523437, 0.03032048034667969, 0.030363359451293946, 0.030118528366088866, 0.03025820732116699, 0.030196704864501954, 0.03020956802368164, 0.030077407836914063, 0.030193119049072265, 0.03006217575073242, 0.030053312301635743, 0.030906368255615234, 0.03016499137878418, 0.030170175552368166, 0.029983680725097658, 0.02998886489868164, 0.0301711368560791, 0.030263296127319338, 0.03037539291381836, 0.030380319595336915, 0.030226688385009765, 0.030731264114379882, 0.03039468765258789, 0.030456512451171876, 0.030529535293579102, 0.030395584106445314, 0.030328832626342773, 0.030462783813476564, 0.03057663917541504, 0.030527488708496094, 0.030404287338256834, 0.030144832611083985, 0.03027276802062988, 0.030313215255737304, 0.03034275245666504, 0.030339296340942384, 0.030138559341430664, 0.030513151168823242, 0.03026896095275879, 0.030047903060913084, 0.03033171272277832, 0.030033151626586915, 0.03007753562927246, 0.02992483139038086, 0.030057151794433593, 0.030052352905273437, 0.03018137550354004, 0.0301560001373291, 0.0300501766204834, 0.03047222328186035, 0.030079872131347656, 0.030093311309814453, 0.03006224060058594, 0.029860191345214844]",tokens/s,32.90229350830191,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-0.5B,Qwen/Qwen1.5-0.5B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1339.273216,1148.125184,0.0,752.877568,710.554112,s,1,9.06067578125,9.06067578125,0.0,9.06067578125,9.06067578125,9.06067578125,9.06067578125,[9.06067578125],,kWh,4.953646999583725e-05,5.456865979846261e-06,1.666501333200121e-05,7.165834930768472e-05,,MB,1380.43392,1429.143552,0.0,1019.215872,949.09696,s,10,0.2942716503143311,0.02942716503143311,0.00022323368126339786,0.029456704139709473,0.029709081268310546,0.029712220954895018,0.029714732704162597,"[0.02908736038208008, 0.029266624450683593, 0.02917398452758789, 0.029661823272705078, 0.02954355239868164, 0.029525312423706054, 0.02971536064147949, 0.02938809585571289, 0.029201152801513672, 0.029708383560180664]",tokens/s,8699.44487437201,kWh,8.653325942554742e-07,9.542787400274307e-08,5.681242217574055e-07,1.528884690015623e-06,tokens/kWh,167442320.3213475,MB,1411.227648,1437.53216,0.0,1027.60448,949.09952,s,10,14.7783876953125,1.47783876953125,0.012717323493488161,1.4748740844726562,1.4969176635742187,1.4986131286621094,1.4999695007324219,"[1.465909423828125, 1.487019775390625, 1.4622545166015626, 1.4743558349609376, 1.475392333984375, 1.50030859375, 1.4965408935546876, 1.4825546875, 1.4701060791015625, 1.463945556640625]",tokens/s,42.629819503234934,kWh,4.354316999407808e-05,4.802410100368045e-06,1.7512890243041744e-05,6.585847033748786e-05,tokens/kWh,956596.7699699098,,s,630,14.776034406661996,0.02345402286771744,0.0004678781912243277,0.023406271934509275,0.023798483657836916,0.023929944038391115,0.025378705310821544,"[0.02317145538330078, 0.023236608505249022, 0.023126047134399415, 0.023090911865234376, 0.023014816284179687, 0.023077728271484375, 0.023382015228271484, 0.02467020797729492, 0.023322336196899413, 0.023066335678100586, 0.023068256378173828, 0.023079103469848632, 0.023062944412231445, 0.02304787254333496, 0.02313222312927246, 0.023034496307373045, 0.02304956817626953, 0.02298908805847168, 0.023044479370117187, 0.023377920150756838, 0.02310495948791504, 0.02314687919616699, 0.02325654411315918, 0.023218624114990233, 0.023033760070800782, 0.023341440200805665, 0.023050048828125, 0.023063743591308594, 0.02334617614746094, 0.023126016616821288, 0.023224319458007812, 0.023160991668701173, 0.02316476821899414, 0.02305638313293457, 0.023698720932006836, 0.023110368728637695, 0.023233888626098632, 0.023124639511108398, 0.02306252861022949, 0.023969280242919923, 0.023087615966796874, 0.023363136291503907, 0.02329439926147461, 0.02344927978515625, 0.023587039947509766, 0.0241112003326416, 0.023644159317016602, 0.02357855987548828, 0.023468095779418944, 0.02367692756652832, 0.023396127700805663, 0.02362758445739746, 0.023465599060058594, 0.02331119918823242, 0.023117759704589843, 0.023085279464721678, 0.0230664005279541, 0.02309939193725586, 0.023205791473388672, 0.023170528411865236, 0.02310403251647949, 0.023055551528930664, 0.02319388771057129, 0.025132352828979493, 0.024007328033447267, 0.023142208099365236, 0.023210016250610352, 0.023070272445678712, 0.02308134460449219, 0.023132511138916016, 0.023209856033325194, 0.023405664443969725, 0.02321878433227539, 0.02308127975463867, 0.022891584396362304, 0.02290598487854004, 0.022908512115478515, 0.02299888038635254, 0.023011711120605467, 0.02288640022277832, 0.02293350410461426, 0.023472127914428712, 0.025620479583740235, 0.02332467269897461, 0.023330976486206054, 0.02352943992614746, 0.023367551803588866, 0.023373823165893554, 0.023383071899414062, 0.023323776245117188, 0.023493568420410157, 0.024407167434692383, 0.02378838348388672, 0.023742591857910156, 0.027173215866088868, 0.0237957763671875, 0.02355036735534668, 0.023366975784301757, 0.0232159366607666, 0.023216991424560546, 0.02326265525817871, 0.02343382453918457, 0.023395360946655272, 0.023634784698486327, 0.02388595199584961, 0.023769216537475588, 0.023877504348754883, 0.023607295989990236, 0.0236810245513916, 0.02367487907409668, 0.023668512344360352, 0.023748735427856445, 0.02376006317138672, 0.023824480056762694, 0.023689888000488282, 0.023719776153564454, 0.02391276741027832, 0.02393087959289551, 0.023842815399169923, 0.02393414306640625, 0.023944000244140624, 0.023554048538208007, 0.023588863372802735, 0.023379968643188476, 0.023971839904785155, 0.023398143768310547, 0.023480319976806642, 0.0229171199798584, 0.023529504776000975, 0.022869983673095704, 0.02287820816040039, 0.023367679595947266, 0.02339801597595215, 0.022933759689331056, 0.023593088150024415, 0.022965824127197266, 0.022795808792114257, 0.022799264907836913, 0.022814399719238283, 0.022935552597045897, 0.02283344078063965, 0.023304224014282227, 0.02288435173034668, 0.02278121566772461, 0.022868703842163086, 0.023011327743530274, 0.022796287536621093, 0.02305366325378418, 0.022975135803222656, 0.022882368087768553, 0.022757408142089843, 0.022945632934570314, 0.022964544296264648, 0.023022592544555662, 0.02297318458557129, 0.023085056304931642, 0.023082080841064452, 0.02289142417907715, 0.02294988822937012, 0.023019392013549803, 0.023131391525268555, 0.023349151611328126, 0.023194271087646483, 0.02331875228881836, 0.023220191955566405, 0.023195199966430664, 0.023343679428100585, 0.023236608505249022, 0.023218175888061524, 0.02325836753845215, 0.023541631698608397, 0.023339616775512696, 0.023406879425048828, 0.02332054328918457, 0.023735519409179687, 0.02461948776245117, 0.023928800582885743, 0.023382144927978514, 0.023486719131469727, 0.02337366485595703, 0.023539871215820313, 0.023298240661621093, 0.02380985641479492, 0.023310111999511718, 0.023339231491088866, 0.023396352767944335, 0.023333919525146483, 0.0234935359954834, 0.02352521514892578, 0.023863296508789062, 0.023743839263916017, 0.023628448486328123, 0.023662591934204103, 0.02360691261291504, 0.023711679458618164, 0.023652799606323244, 0.023738367080688477, 0.023756799697875978, 0.02371583938598633, 0.02372425651550293, 0.02371708869934082, 0.02373075294494629, 0.023746559143066406, 0.02385305595397949, 0.023570432662963867, 0.02344550323486328, 0.023468032836914062, 0.023334527969360353, 0.023372159957885743, 0.02330419158935547, 0.023189504623413085, 0.02327756881713867, 0.023117279052734373, 0.023043615341186523, 0.02304921531677246, 0.022982656478881838, 0.023136255264282226, 0.023289279937744142, 0.023705312728881836, 0.023285888671875, 0.023255775451660157, 0.023250143051147462, 0.022958848953247072, 0.023107616424560547, 0.023052288055419923, 0.023072383880615235, 0.023005855560302733, 0.023137472152709962, 0.023271455764770507, 0.023640575408935546, 0.023326719284057617, 0.023373823165893554, 0.023471296310424803, 0.02328454399108887, 0.02345564842224121, 0.023334304809570314, 0.023296703338623048, 0.023244800567626952, 0.023464960098266603, 0.02330294418334961, 0.023417055130004884, 0.023402496337890624, 0.023444896697998048, 0.02325926399230957, 0.023286239624023437, 0.02321343994140625, 0.02341337585449219, 0.023267328262329103, 0.02328371238708496, 0.02326323127746582, 0.023371776580810546, 0.023750879287719726, 0.023793664932250977, 0.023723167419433595, 0.023555168151855467, 0.023559328079223632, 0.0236343994140625, 0.02379952049255371, 0.024199167251586915, 0.023945152282714845, 0.023740896224975584, 0.02360691261291504, 0.023716224670410156, 0.02370560073852539, 0.023721376419067384, 0.023691871643066405, 0.023741952896118163, 0.023687679290771483, 0.023416831970214845, 0.02338819122314453, 0.023413856506347655, 0.023558080673217772, 0.02348700714111328, 0.02352783966064453, 0.023338464736938475, 0.023236896514892576, 0.023181343078613283, 0.023082399368286134, 0.022969152450561522, 0.02307449531555176, 0.023085535049438475, 0.02314019203186035, 0.02312735939025879, 0.023029632568359375, 0.022984575271606446, 0.023086015701293944, 0.022937599182128905, 0.023077024459838866, 0.023221824645996095, 0.023001279830932617, 0.023040096282958986, 0.023013376235961915, 0.023195232391357422, 0.023187871932983398, 0.023090463638305664, 0.02329248046875, 0.02319545555114746, 0.023404895782470705, 0.02353561592102051, 0.023602176666259765, 0.023654783248901367, 0.02366054344177246, 0.023454208374023438, 0.023224447250366213, 0.023326208114624023, 0.023238367080688476, 0.02331523132324219, 0.023338207244873045, 0.02357107162475586, 0.023419008255004883, 0.023457632064819336, 0.02363376045227051, 0.02369980812072754, 0.023704671859741212, 0.02367990493774414, 0.023982912063598632, 0.023767040252685546, 0.02367283248901367, 0.023817344665527342, 0.023664928436279296, 0.023730783462524413, 0.02369126319885254, 0.023807136535644532, 0.023767904281616212, 0.023736480712890626, 0.02377507209777832, 0.023744543075561522, 0.023680959701538086, 0.023650335311889648, 0.02379481506347656, 0.023645055770874022, 0.023661983489990233, 0.023683679580688476, 0.02389561653137207, 0.023673280715942383, 0.02364975929260254, 0.02377987289428711, 0.02373222351074219, 0.023748607635498048, 0.023738304138183595, 0.023799840927124023, 0.023688608169555665, 0.024232576370239258, 0.02368511962890625, 0.02374678421020508, 0.023551040649414063, 0.023628480911254884, 0.023742496490478517, 0.02365644836425781, 0.02382601547241211, 0.02377564811706543, 0.023743871688842774, 0.023613632202148436, 0.02386089515686035, 0.023902048110961915, 0.023872447967529298, 0.023666688919067383, 0.02384092712402344, 0.023738208770751952, 0.023586816787719726, 0.023619583129882812, 0.023737760543823243, 0.023648799896240233, 0.023586879730224608, 0.024459264755249024, 0.02366464042663574, 0.023686559677124023, 0.023734880447387696, 0.023787519454956055, 0.023777280807495117, 0.023819583892822266, 0.024060415267944335, 0.027090335845947267, 0.023963775634765625, 0.023654752731323243, 0.023760831832885743, 0.023578079223632812, 0.023777408599853514, 0.023635551452636717, 0.026352415084838866, 0.024042751312255858, 0.023329376220703125, 0.023359199523925782, 0.023655872344970703, 0.023469024658203125, 0.023701536178588868, 0.02365644836425781, 0.023704992294311524, 0.023681631088256837, 0.023830528259277343, 0.023762943267822266, 0.023797760009765623, 0.023834367752075196, 0.0238287353515625, 0.02368953514099121, 0.023764896392822265, 0.023774848937988282, 0.025479328155517577, 0.023715167999267577, 0.023869472503662108, 0.023683040618896485, 0.023740800857543945, 0.023559839248657226, 0.02375529670715332, 0.023938655853271484, 0.02389596748352051, 0.023798368453979493, 0.023786815643310547, 0.023610048294067383, 0.02366208076477051, 0.02362201690673828, 0.02364166450500488, 0.026810943603515627, 0.023846015930175782, 0.023659040451049804, 0.023546207427978517, 0.023611391067504883, 0.023537664413452147, 0.023576448440551758, 0.02380112075805664, 0.023543840408325196, 0.02344428825378418, 0.023556095123291015, 0.02345779228210449, 0.023764480590820314, 0.024295583724975586, 0.023822687149047853, 0.023564287185668945, 0.023492448806762694, 0.023418367385864256, 0.02348646354675293, 0.02353753662109375, 0.023382816314697265, 0.023333984375, 0.023215007781982423, 0.023286912918090822, 0.023335807800292967, 0.023590911865234376, 0.023217504501342773, 0.02325980758666992, 0.023297695159912108, 0.023316352844238282, 0.02369430351257324, 0.023160928726196288, 0.023456703186035155, 0.023466976165771484, 0.023586816787719726, 0.02346598434448242, 0.023613439559936524, 0.023773151397705077, 0.023831615447998045, 0.023699615478515627, 0.023706432342529296, 0.02375267219543457, 0.02358415985107422, 0.023638656616210937, 0.023571903228759766, 0.023636064529418944, 0.023607776641845702, 0.023603200912475586, 0.023717439651489258, 0.02368115234375, 0.023900480270385743, 0.02367852783203125, 0.023697599411010743, 0.02348876762390137, 0.023616640090942383, 0.02352547264099121, 0.023438112258911133, 0.02332819175720215, 0.02324502372741699, 0.023039743423461913, 0.023020000457763673, 0.023199167251586914, 0.02288096046447754, 0.023060480117797853, 0.02306252861022949, 0.023109792709350586, 0.022951776504516602, 0.02309939193725586, 0.02471731185913086, 0.024356576919555666, 0.023531808853149414, 0.0232193603515625, 0.023532384872436522, 0.02350694465637207, 0.023272960662841798, 0.02346031951904297, 0.02322003173828125, 0.023435487747192382, 0.023418079376220702, 0.023251007080078125, 0.02333283233642578, 0.023489280700683592, 0.023491775512695313, 0.02330291175842285, 0.023273536682128906, 0.023215551376342774, 0.02326585578918457, 0.026685440063476562, 0.024603679656982423, 0.02335968017578125, 0.02330307197570801, 0.023185279846191405, 0.023202175140380858, 0.023189952850341797, 0.02337980842590332, 0.024762367248535155, 0.02369126319885254, 0.023716896057128907, 0.023649248123168945, 0.023715232849121092, 0.02361782455444336, 0.023650623321533202, 0.023622848510742187, 0.023598047256469728, 0.0236596794128418, 0.023494560241699217, 0.02361369514465332, 0.023732351303100585, 0.02361180877685547, 0.023609344482421874, 0.02371379280090332, 0.023690528869628906, 0.02357232093811035, 0.023599071502685545, 0.02333590316772461, 0.023291839599609374, 0.02329190444946289, 0.023371776580810546, 0.023184480667114257, 0.02317932891845703, 0.023011199951171873, 0.023100383758544923, 0.023101152420043944, 0.02303932762145996, 0.022924448013305666, 0.022990623474121095, 0.022844959259033203, 0.02294950485229492, 0.022952800750732423, 0.0229965763092041, 0.022861888885498047, 0.022987104415893553, 0.023044095993041993, 0.022943744659423827, 0.022927007675170898, 0.022965824127197266, 0.02305718421936035, 0.023094688415527344, 0.02359766387939453, 0.023175167083740233, 0.023107168197631835, 0.023281791687011718, 0.02340892791748047, 0.023162879943847657, 0.02308006477355957, 0.023212095260620118, 0.022958528518676757, 0.023078367233276366, 0.0230960636138916, 0.023122079849243166, 0.02345779228210449, 0.02369945526123047, 0.023629472732543944, 0.023630176544189453, 0.023600927352905275, 0.02382681655883789, 0.02376153564453125, 0.023654399871826173, 0.0237076473236084, 0.023617536544799804, 0.02377462387084961, 0.023581151962280274, 0.0236627197265625, 0.02374390411376953, 0.023738975524902343, 0.02369126319885254, 0.023463455200195313, 0.02326188850402832, 0.02343497657775879, 0.023358848571777342, 0.023343807220458986, 0.023342975616455076, 0.02332480049133301, 0.023236608505249022, 0.023220224380493162, 0.023011327743530274, 0.022924383163452147, 0.023034624099731445, 0.02332054328918457, 0.023392671585083007, 0.023652128219604492, 0.023131519317626952, 0.023063167572021485, 0.022898496627807616, 0.02309872055053711, 0.022985727310180663, 0.02335750389099121, 0.02323843193054199, 0.023554048538208007, 0.023201791763305665, 0.023233983993530275, 0.02313007926940918, 0.02294371223449707, 0.023199392318725587, 0.0229385929107666, 0.023277183532714844, 0.022931840896606444, 0.022960256576538086, 0.02356825637817383, 0.023244800567626952, 0.022974464416503908, 0.022996768951416016, 0.023230495452880858, 0.02297056007385254, 0.02301686477661133, 0.022899295806884764, 0.022974464416503908, 0.022999040603637694, 0.022975839614868165, 0.022910816192626953, 0.023233343124389648, 0.02287820816040039, 0.022988128662109374, 0.022862112045288086, 0.0229748477935791, 0.022902624130249023, 0.02289004707336426, 0.023021440505981445]",tokens/s,42.63660889392322,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3731.329024,5369.69216,0.0,4974.444544,4685.071872,s,1,11.1715478515625,11.1715478515625,0.0,11.1715478515625,11.1715478515625,11.1715478515625,11.1715478515625,[11.1715478515625],,kWh,0.00012451798164166424,1.3727608633347855e-05,5.261337542400088e-05,0.00019085896569901296,,MB,1988.276224,5403.246592,0.0,4993.318912,4233.624576,s,10,1.9220595092773436,0.19220595092773438,0.00023303240069167826,0.19226694488525392,0.19240954589843748,0.1924179229736328,0.19242462463378907,"[0.19225372314453126, 0.19157609558105468, 0.19242630004882813, 0.19224838256835938, 0.19240768432617186, 0.1921829376220703, 0.19234197998046876, 0.19228518676757814, 0.19205705261230469, 0.19228016662597655]",tokens/s,1331.9046510492847,kWh,5.68030962131394e-06,6.264281311712178e-07,3.7552861238846522e-06,1.006202387636981e-05,tokens/kWh,25442197.627974622,MB,1998.278656,5403.246592,0.0,4993.318912,4336.176128,s,10,21.814504150390626,2.1814504150390626,0.024596799761330265,2.18764794921875,2.2065200439453125,2.2078297241210936,2.208877468261719,"[2.176327392578125, 2.205752197265625, 2.209139404296875, 2.198985595703125, 2.167893310546875, 2.163714111328125, 2.136791748046875, 2.150702880859375, 2.20622900390625, 2.198968505859375]",tokens/s,28.87986798401369,kWh,6.44792890078529e-05,7.111966963577139e-06,3.8065941777515405e-05,0.00010965719774894542,tokens/kWh,574517.6905234739,,s,630,21.811262943267828,0.03462105229090131,0.0006365128838513587,0.034709230422973636,0.03513799514770508,0.035320075225830076,0.036336495742797854,"[0.034657791137695314, 0.03510147094726562, 0.035061153411865234, 0.034299678802490234, 0.03433555221557617, 0.0346662712097168, 0.0344005126953125, 0.03419084930419922, 0.034439136505126956, 0.03428403091430664, 0.03415033721923828, 0.034528865814208984, 0.03439379119873047, 0.03421878433227539, 0.03442496109008789, 0.03454553604125977, 0.0344917106628418, 0.03459756851196289, 0.03441196823120117, 0.034748992919921874, 0.03480924987792969, 0.03475219345092773, 0.03456911849975586, 0.03450259017944336, 0.03487468719482422, 0.0345747184753418, 0.03452876663208008, 0.03502735900878906, 0.034482177734375, 0.03420848083496094, 0.03463529586791992, 0.03438582229614258, 0.034328895568847655, 0.034375679016113284, 0.034293441772460936, 0.034229984283447264, 0.03427388763427734, 0.03424380874633789, 0.03431878280639648, 0.03425497436523438, 0.03456991958618164, 0.03460889434814453, 0.03442768096923828, 0.03477673721313477, 0.0345214729309082, 0.034662559509277345, 0.03448320007324219, 0.034349918365478516, 0.034522911071777344, 0.03487353515625, 0.034461441040039065, 0.034713855743408205, 0.03460931015014648, 0.034522815704345705, 0.03444483184814453, 0.03478144073486328, 0.03474166488647461, 0.03471868896484375, 0.034678943634033205, 0.03520259094238281, 0.034642238616943356, 0.034565567016601566, 0.03453580856323242, 0.035050815582275394, 0.0350728645324707, 0.03494390487670898, 0.034964000701904294, 0.03513148880004883, 0.03518086242675781, 0.03522355270385742, 0.03498521423339844, 0.035007232666015624, 0.03561872100830078, 0.03513148880004883, 0.03505692672729492, 0.03503081512451172, 0.03492345428466797, 0.034848159790039065, 0.034869537353515626, 0.03470070266723633, 0.03519811248779297, 0.03518848037719727, 0.03495935821533203, 0.03494876861572266, 0.035037281036376954, 0.034828544616699215, 0.03522076797485352, 0.03496124649047851, 0.034853759765625, 0.035077278137207034, 0.035126113891601564, 0.03509036636352539, 0.03490208053588867, 0.035053184509277344, 0.0348983039855957, 0.03505718231201172, 0.035375457763671875, 0.03511439895629883, 0.03496137619018555, 0.03522051239013672, 0.035344318389892576, 0.035106494903564454, 0.03503318405151367, 0.035016384124755856, 0.034995967864990235, 0.03510476684570313, 0.03502342224121094, 0.03476591873168945, 0.034845375061035154, 0.035302879333496094, 0.03490291213989258, 0.035522430419921876, 0.0348671989440918, 0.034740222930908206, 0.03491356658935547, 0.03467142486572266, 0.03465107345581055, 0.034697952270507815, 0.03463808059692383, 0.03468921661376953, 0.03466400146484375, 0.03486044692993164, 0.035044193267822266, 0.034952545166015626, 0.0349967041015625, 0.03521865463256836, 0.03520716857910156, 0.035127296447753906, 0.03501055908203125, 0.03508224105834961, 0.03529913711547852, 0.03512118530273438, 0.035379711151123046, 0.03526006317138672, 0.03510441589355469, 0.03473648071289062, 0.034899585723876955, 0.03485561752319336, 0.034762081146240235, 0.034707809448242186, 0.03492195129394531, 0.03464041519165039, 0.03478291320800781, 0.034892097473144534, 0.034840286254882814, 0.034947006225585935, 0.03482803344726562, 0.03480201721191406, 0.035111167907714846, 0.03474985504150391, 0.03483286285400391, 0.03503936004638672, 0.03503308868408203, 0.03524726486206055, 0.03498223876953125, 0.03503279876708985, 0.03512400054931641, 0.034866943359375, 0.03495552062988281, 0.03515830230712891, 0.03488470458984375, 0.03538988876342773, 0.03513772964477539, 0.0352852783203125, 0.03497955322265625, 0.034920448303222655, 0.035052574157714844, 0.03509747314453125, 0.03486265563964844, 0.035844894409179685, 0.034995967864990235, 0.034816001892089846, 0.03489014434814453, 0.035090015411376956, 0.03492428970336914, 0.03510201644897461, 0.03503721618652344, 0.03536374282836914, 0.03494911956787109, 0.03491020965576172, 0.03502489471435547, 0.035334144592285156, 0.03510198211669922, 0.03685772705078125, 0.035224414825439455, 0.035092479705810545, 0.03523763275146485, 0.035066112518310544, 0.034991329193115234, 0.0350777587890625, 0.035180927276611325, 0.035018272399902343, 0.03497385787963867, 0.03511328125, 0.03501628875732422, 0.03501916885375977, 0.035151966094970705, 0.03522361755371094, 0.03485270309448242, 0.035435966491699215, 0.03516473770141602, 0.03478704071044922, 0.03489820861816406, 0.034903968811035156, 0.034994174957275394, 0.03496764755249023, 0.03487744140625, 0.03501260757446289, 0.034990367889404295, 0.03496540832519531, 0.03494236755371094, 0.034904449462890626, 0.03491206359863281, 0.035117279052734374, 0.035133377075195316, 0.03508595275878906, 0.034877120971679686, 0.03481472015380859, 0.03470678329467773, 0.03470998382568359, 0.03485244750976563, 0.03463024139404297, 0.034721694946289065, 0.03468854522705078, 0.034807777404785155, 0.03483299255371094, 0.03468288040161133, 0.03480902481079102, 0.03494380950927734, 0.03493465423583984, 0.03462892913818359, 0.03472681427001953, 0.03473516845703125, 0.03502985763549805, 0.034812030792236326, 0.03462684631347656, 0.0346855354309082, 0.034854911804199216, 0.034708606719970704, 0.03482304000854492, 0.03486848068237305, 0.034875553131103514, 0.03477155303955078, 0.034983615875244144, 0.035020896911621094, 0.03480352020263672, 0.03478160095214844, 0.03489382553100586, 0.03494460678100586, 0.03489628982543945, 0.03458857727050781, 0.03484630584716797, 0.03506988906860352, 0.034912254333496096, 0.03480112075805664, 0.03493116760253906, 0.0348529281616211, 0.03486515045166016, 0.03518838500976563, 0.035359073638916015, 0.03637350463867187, 0.03528396987915039, 0.034968929290771486, 0.03520169448852539, 0.03498393630981445, 0.034890975952148434, 0.035608993530273435, 0.03492287826538086, 0.034887680053710936, 0.037615135192871095, 0.03499657440185547, 0.03495334243774414, 0.034568321228027346, 0.03449843215942383, 0.03452102279663086, 0.03442694473266601, 0.03426044845581055, 0.03410383987426758, 0.034149856567382814, 0.033927711486816406, 0.03390089416503906, 0.034018913269042966, 0.033905025482177734, 0.033994430541992186, 0.033877246856689455, 0.03415036773681641, 0.03413897705078125, 0.03401523208618164, 0.03393734359741211, 0.03383030319213867, 0.03384707260131836, 0.03395462417602539, 0.0340912971496582, 0.03401084899902344, 0.03382799911499024, 0.033829631805419924, 0.033783584594726565, 0.033818943023681644, 0.03375500869750977, 0.033901920318603514, 0.03382966232299805, 0.03383622360229492, 0.03379692840576172, 0.03393526458740234, 0.033833183288574216, 0.03386150360107422, 0.03395993423461914, 0.03385139083862305, 0.03380601501464844, 0.033976638793945316, 0.03413212966918945, 0.03431164932250977, 0.03432182312011719, 0.03422073745727539, 0.0342256965637207, 0.03465030288696289, 0.034551551818847656, 0.034345409393310544, 0.034385921478271485, 0.03475408172607422, 0.034484703063964846, 0.03437577438354492, 0.03438310241699219, 0.034687393188476565, 0.03460275268554688, 0.03478374481201172, 0.03779116821289062, 0.03495951843261719, 0.03472864151000977, 0.03481958389282227, 0.03483168029785156, 0.03502127838134766, 0.034920833587646485, 0.034848831176757813, 0.03551798248291015, 0.03491635131835937, 0.03466083145141602, 0.03464176177978515, 0.03455123138427734, 0.034400993347167966, 0.034557472229003905, 0.03437641525268555, 0.034153759002685545, 0.03399932861328125, 0.034007392883300784, 0.03410835266113281, 0.033907424926757815, 0.033931102752685544, 0.034285728454589846, 0.03390630340576172, 0.033834590911865234, 0.033880062103271484, 0.03375795364379883, 0.03400278472900391, 0.03398851013183594, 0.034248992919921874, 0.03426639938354492, 0.03424739074707031, 0.03408246231079102, 0.033997150421142576, 0.03403571319580078, 0.03413638305664062, 0.033936351776123044, 0.03403062438964844, 0.03393535995483398, 0.03382854461669922, 0.03379974365234375, 0.034107231140136716, 0.03377008056640625, 0.033955520629882815, 0.03390284729003906, 0.03381033706665039, 0.033839263916015626, 0.03381248092651367, 0.03393539047241211, 0.03403158569335937, 0.03391241455078125, 0.03452048110961914, 0.03415225601196289, 0.03381577682495117, 0.033892833709716796, 0.03376710510253906, 0.033872222900390624, 0.03379657745361328, 0.03380223846435547, 0.0337448959350586, 0.033920352935791015, 0.033802913665771483, 0.033860897064208986, 0.03370687866210938, 0.03379315185546875, 0.03367536163330078, 0.034025310516357425, 0.033710880279541014, 0.033897663116455076, 0.033735488891601564, 0.03384320068359375, 0.033716224670410154, 0.03394355010986328, 0.0337570571899414, 0.033775039672851566, 0.033726337432861325, 0.03386028671264649, 0.03377577590942383, 0.033991935729980466, 0.03378153610229492, 0.03394831848144531, 0.033871200561523436, 0.03384415817260742, 0.033742847442626955, 0.033783584594726565, 0.03364476776123047, 0.03378176116943359, 0.03367353439331055, 0.03386272048950195, 0.03390118408203125, 0.0339865608215332, 0.033832351684570314, 0.03395955276489258, 0.033810497283935544, 0.03487017440795898, 0.03407014465332031, 0.03392755126953125, 0.03387731170654297, 0.033767681121826175, 0.03374111938476562, 0.03397644805908203, 0.03391392135620117, 0.033792961120605466, 0.033736671447753906, 0.03390985488891601, 0.03505456161499024, 0.03408924865722656, 0.03459449768066406, 0.0339060173034668, 0.0338438720703125, 0.033939456939697264, 0.03512351989746094, 0.03388716888427734, 0.03383987045288086, 0.03388620758056641, 0.03426019287109375, 0.03397481536865234, 0.033840320587158204, 0.03404019165039063, 0.03496966552734375, 0.03562060928344726, 0.03401507186889648, 0.033872608184814454, 0.033859329223632814, 0.03422854232788086, 0.0340398063659668, 0.03400908660888672, 0.03393929672241211, 0.033938785552978516, 0.0338430061340332, 0.03387289428710937, 0.03388729476928711, 0.03388102340698242, 0.03385945510864258, 0.0337204475402832, 0.033764671325683594, 0.033694400787353515, 0.03375718307495117, 0.03417411041259766, 0.03389321517944336, 0.03378598403930664, 0.03414412689208984, 0.033773822784423826, 0.033940895080566406, 0.03378006362915039, 0.03377129745483398, 0.03376316833496094, 0.033693790435791016, 0.03376380920410156, 0.034727489471435544, 0.03447529602050781, 0.033891521453857425, 0.03388095855712891, 0.03391766357421875, 0.03387615966796875, 0.0340766716003418, 0.03410943984985351, 0.03433478546142578, 0.03451017761230469, 0.0338166389465332, 0.03390963363647461, 0.033861343383789065, 0.034175167083740236, 0.034297119140625, 0.03443699264526367, 0.034328670501708985, 0.034880001068115236, 0.0343551025390625, 0.03444543838500977, 0.034609119415283204, 0.03460710525512695, 0.034590049743652346, 0.03458899307250977, 0.03451068878173828, 0.03438153457641602, 0.03435411071777344, 0.034305889129638674, 0.03471260833740234, 0.03517875289916992, 0.03493072128295899, 0.03492262268066406, 0.03503091049194336, 0.03478720092773437, 0.034756736755371095, 0.03490313720703125, 0.03503177642822266, 0.03497334289550781, 0.03487161636352539, 0.03480188751220703, 0.03479321670532227, 0.034928417205810545, 0.03593593597412109, 0.03483337783813477, 0.03501571273803711, 0.034429153442382815, 0.03569926452636719, 0.03606099319458008, 0.03472198486328125, 0.035871967315673825, 0.03470985412597656, 0.034966209411621096, 0.034823070526123046, 0.034859073638916015, 0.036114944458007815, 0.037464351654052735, 0.03471769714355469, 0.03478511810302734, 0.034621601104736326, 0.034653888702392575, 0.03624297714233399, 0.034826656341552735, 0.0345849609375, 0.03468288040161133, 0.03465539169311523, 0.03478524780273438, 0.035003326416015626, 0.03495135879516602, 0.034751873016357425, 0.034746753692626954, 0.034788734436035154, 0.0348084487915039, 0.034723838806152346, 0.034909278869628906, 0.03495935821533203, 0.03524291229248047, 0.03472326278686524, 0.03474256134033203, 0.03491167831420899, 0.03503327941894531, 0.035151840209960934, 0.034726593017578126, 0.03475580978393555, 0.03538000106811524, 0.034911552429199216, 0.03485766220092774, 0.034947071075439456, 0.03485935974121094, 0.03494198226928711, 0.03497564697265625, 0.03515055847167969, 0.034936832427978515, 0.03504537582397461, 0.03501398468017578, 0.03514038467407227, 0.034834110260009765, 0.03465804672241211, 0.03484415817260742, 0.035513278961181644, 0.034813438415527344, 0.03475711822509766, 0.03465625762939453, 0.034826526641845705, 0.03494063949584961, 0.03479100799560547, 0.034621856689453126, 0.0346255989074707, 0.034670528411865235, 0.034697216033935545, 0.03459408187866211, 0.03464662551879883, 0.03471334457397461, 0.036245887756347656, 0.03476460647583008, 0.03477875137329101, 0.03465039825439453, 0.034834720611572265, 0.035259521484375, 0.03517734527587891, 0.034885631561279294, 0.03660377502441406, 0.042235328674316404, 0.0347283821105957, 0.03441196823120117, 0.03442361450195312, 0.03434086227416992, 0.035374752044677736, 0.03471571350097656, 0.03549196624755859, 0.03411574554443359, 0.03463372802734375, 0.03429171371459961, 0.034225505828857423, 0.033992862701416014, 0.03452915191650391, 0.03444755172729492, 0.03469337463378906, 0.034508800506591795, 0.03469126510620117, 0.03464022445678711, 0.03484169769287109, 0.034632160186767576, 0.03459414291381836, 0.034630401611328125, 0.03478307342529297, 0.03473177719116211, 0.034866657257080075, 0.03461215972900391, 0.03481372833251953, 0.03475254440307617, 0.03482156753540039, 0.03461628723144531, 0.03453449630737305, 0.03451996612548828, 0.034746143341064455]",tokens/s,28.884159603167465,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Deci/DeciCoder-1b contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Deci/DeciCoder-1b. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2166.718464,2964.258816,0.0,2569.0112,2295.745536,s,1,11.1279580078125,11.1279580078125,0.0,11.1279580078125,11.1279580078125,11.1279580078125,11.1279580078125,[11.1279580078125],,kWh,8.049906672503463e-05,8.872083881687885e-06,3.1210302745998675e-05,0.00012058145335272118,,MB,2131.853312,2981.036032,0.0,2571.108352,2282.97216,s,10,1.0005862045288085,0.10005862045288086,0.0002416026830852026,0.09998668670654297,0.10033361968994141,0.10044057159423828,0.10052613311767578,"[0.10054752349853516, 0.09991827392578125, 0.10029766082763672, 0.09975218963623046, 0.10008512115478516, 0.09992655944824219, 0.09990780639648437, 0.10004681396484374, 0.10030985260009766, 0.09979440307617188]",tokens/s,2558.5001955983826,kWh,2.975335105260814e-06,3.2812588964280005e-07,1.9688595997777816e-06,5.272320594681396e-06,tokens/kWh,48555469.15304189,MB,2145.849344,2981.036032,0.0,2571.108352,2390.926848,s,10,18.799360107421872,1.8799360107421872,0.006567797939365876,1.8798137207031251,1.8885275390624998,1.8895375122070313,1.8903454907226562,"[1.8782554931640625, 1.8793629150390625, 1.8883031005859374, 1.8768994140625, 1.872440185546875, 1.88361767578125, 1.86726904296875, 1.8802645263671875, 1.8824002685546875, 1.8905474853515625]",tokens/s,33.511778932905266,kWh,5.4661805717654654e-05,6.028934780673097e-06,2.725338600022186e-05,8.794412649854961e-05,tokens/kWh,716363.9291026331,,s,630,18.796350204467796,0.029835476515028207,0.0005983430423490145,0.029766528129577634,0.030310063362121584,0.030632141304016113,0.031643901500701906,"[0.030085248947143553, 0.029583232879638672, 0.029198335647583007, 0.02931679916381836, 0.029225055694580077, 0.029317344665527344, 0.02921401596069336, 0.029323200225830078, 0.029260543823242186, 0.029469791412353515, 0.02990572738647461, 0.029855840682983397, 0.029741056442260744, 0.029847423553466798, 0.029919359207153322, 0.029820928573608397, 0.02989414405822754, 0.029972736358642577, 0.030054079055786134, 0.02980512046813965, 0.031064064025878906, 0.02976153564453125, 0.02956819152832031, 0.02992620849609375, 0.029839519500732423, 0.030032928466796876, 0.029621055603027344, 0.029650016784667967, 0.029678688049316407, 0.0295032958984375, 0.029908992767333983, 0.029560831069946288, 0.029489152908325194, 0.029691904067993165, 0.029730815887451172, 0.029847328186035155, 0.029926719665527343, 0.03020240020751953, 0.02942223930358887, 0.029333215713500976, 0.02976742362976074, 0.029165184020996094, 0.029265792846679687, 0.029497791290283203, 0.029544160842895507, 0.02988912010192871, 0.029908319473266602, 0.03003664016723633, 0.029949951171875, 0.029896703720092774, 0.029853567123413086, 0.029939775466918946, 0.029836576461791994, 0.03012588882446289, 0.02983228874206543, 0.03363622283935547, 0.030794815063476564, 0.030011871337890624, 0.030013824462890627, 0.029782112121582032, 0.02975129508972168, 0.029582399368286133, 0.029332416534423828, 0.03008950424194336, 0.030223487854003906, 0.02970899200439453, 0.02986582374572754, 0.029860191345214844, 0.029863616943359376, 0.02967788887023926, 0.02999225616455078, 0.029788576126098632, 0.0298253116607666, 0.030033023834228515, 0.03035628890991211, 0.029812896728515625, 0.029628320693969725, 0.029677568435668947, 0.02965839958190918, 0.029483552932739257, 0.029447456359863282, 0.029877151489257812, 0.029326751708984376, 0.029239200592041017, 0.029747903823852537, 0.02914303970336914, 0.029288448333740235, 0.0291778564453125, 0.02940025520324707, 0.029926464080810546, 0.029871871948242187, 0.02970572853088379, 0.029803007125854493, 0.029767679214477538, 0.029775871276855468, 0.030281503677368163, 0.030458080291748048, 0.03000060844421387, 0.029930015563964844, 0.029560831069946288, 0.02973014450073242, 0.029724735260009766, 0.029626976013183592, 0.02963046455383301, 0.03023251152038574, 0.029461952209472658, 0.029513696670532226, 0.029413856506347657, 0.029448352813720702, 0.02932649612426758, 0.029716800689697266, 0.029676095962524414, 0.02971785545349121, 0.029509504318237303, 0.03016783905029297, 0.030033567428588866, 0.029991264343261718, 0.030118976593017578, 0.030094272613525392, 0.02994380760192871, 0.030119327545166014, 0.02994867134094238, 0.030297952651977537, 0.030095039367675783, 0.031193344116210938, 0.0310413761138916, 0.030129407882690428, 0.030309343338012697, 0.032150558471679684, 0.029980863571166992, 0.029887264251708984, 0.0297574405670166, 0.029665599822998046, 0.029699775695800783, 0.029628416061401368, 0.029665279388427734, 0.029933631896972655, 0.030123071670532228, 0.030094144821166992, 0.03025632095336914, 0.03012188720703125, 0.030086111068725586, 0.030674943923950194, 0.03030611228942871, 0.030195903778076173, 0.03039967918395996, 0.030130176544189452, 0.03195167922973633, 0.03177004814147949, 0.03033964729309082, 0.030338752746582032, 0.030285888671875, 0.030490432739257813, 0.030300256729125976, 0.030142431259155274, 0.03044384002685547, 0.02969759941101074, 0.029743616104125976, 0.029548511505126954, 0.029746591567993166, 0.029438783645629883, 0.02929030418395996, 0.029298688888549803, 0.029460479736328125, 0.029659135818481445, 0.029471967697143556, 0.029178688049316406, 0.029284128189086912, 0.029315263748168945, 0.029471872329711914, 0.029698335647583007, 0.029837919235229493, 0.02977110481262207, 0.030045984268188476, 0.02981977653503418, 0.029849599838256836, 0.029838815689086914, 0.03002217674255371, 0.02982863998413086, 0.029892095565795897, 0.02987241554260254, 0.029946559906005858, 0.03013804817199707, 0.02986342430114746, 0.029653663635253905, 0.029558944702148437, 0.029310976028442383, 0.029703519821166993, 0.029508256912231447, 0.030160959243774415, 0.030202144622802733, 0.030005247116088866, 0.029933088302612303, 0.029571744918823244, 0.029478208541870117, 0.029190656661987304, 0.029443840026855468, 0.029624576568603515, 0.029888511657714844, 0.029828447341918946, 0.029809024810791014, 0.029731103897094727, 0.029644735336303712, 0.029750751495361327, 0.029972192764282226, 0.0306712646484375, 0.031049503326416015, 0.030140256881713866, 0.03026620864868164, 0.02990835189819336, 0.030366016387939454, 0.030270816802978516, 0.03101590347290039, 0.029638656616210936, 0.029618175506591796, 0.02934169578552246, 0.029640480041503905, 0.02933366394042969, 0.029259103775024414, 0.02925788879394531, 0.030773279190063476, 0.029911808013916016, 0.02962156867980957, 0.029792032241821288, 0.029617984771728514, 0.029362752914428712, 0.031062240600585937, 0.029593727111816407, 0.02985487937927246, 0.02917433547973633, 0.029239551544189453, 0.02920243263244629, 0.029418655395507812, 0.02922991943359375, 0.029826879501342773, 0.02961987113952637, 0.029676063537597656, 0.029664543151855467, 0.029814592361450197, 0.02968796730041504, 0.029827840805053712, 0.02974028778076172, 0.029827840805053712, 0.0295280647277832, 0.029663232803344725, 0.029491071701049806, 0.029993087768554687, 0.030443519592285157, 0.029928447723388672, 0.029484031677246093, 0.02932067108154297, 0.02922140884399414, 0.029472543716430665, 0.030872543334960936, 0.03060736083984375, 0.030689376831054688, 0.02974710464477539, 0.029437952041625977, 0.02920355224609375, 0.029173664093017578, 0.029076480865478517, 0.029206527709960937, 0.029144800186157227, 0.029233024597167968, 0.029114784240722655, 0.029507583618164062, 0.02961974334716797, 0.029769983291625977, 0.029713951110839843, 0.029706943511962892, 0.02970591926574707, 0.02970195198059082, 0.02967193603515625, 0.029710111618041993, 0.029522144317626953, 0.029548032760620117, 0.029778432846069337, 0.02940928077697754, 0.02947715187072754, 0.02964796829223633, 0.02971526336669922, 0.029382463455200195, 0.02915705680847168, 0.029378879547119142, 0.029170719146728516, 0.029237600326538087, 0.029221120834350585, 0.029501535415649413, 0.030172767639160155, 0.030245567321777345, 0.030287679672241212, 0.030163135528564453, 0.030652416229248046, 0.03022233581542969, 0.02998054313659668, 0.03015465545654297, 0.030274015426635742, 0.030367359161376953, 0.03029209518432617, 0.03015065574645996, 0.030454111099243165, 0.029914783477783202, 0.02969331169128418, 0.029909631729125977, 0.029621919631958007, 0.0298110408782959, 0.029617855072021484, 0.029526336669921875, 0.029259967803955077, 0.029345504760742186, 0.029325408935546873, 0.030076927185058593, 0.029222911834716796, 0.029394559860229492, 0.029737632751464845, 0.030038015365600586, 0.030072832107543947, 0.029962175369262694, 0.030449600219726564, 0.03038630485534668, 0.03036742401123047, 0.030045536041259764, 0.030057439804077147, 0.029677568435668947, 0.029569023132324217, 0.02955264091491699, 0.029934783935546876, 0.030255935668945313, 0.0296812801361084, 0.029543807983398437, 0.02978508758544922, 0.02969148826599121, 0.029986623764038087, 0.029956703186035157, 0.029669376373291017, 0.029597183227539063, 0.029983232498168946, 0.03004377555847168, 0.029957727432250978, 0.030054912567138672, 0.0300546875, 0.02983526420593262, 0.029861055374145507, 0.02989084815979004, 0.030095903396606446, 0.029834592819213867, 0.0299465274810791, 0.029872127532958984, 0.0297609920501709, 0.02942416000366211, 0.029459808349609377, 0.029678367614746095, 0.03013996887207031, 0.02974083137512207, 0.029759584426879884, 0.029883968353271485, 0.029614368438720704, 0.02948975944519043, 0.02936800003051758, 0.0294050235748291, 0.029413856506347657, 0.030033344268798827, 0.029803071975708008, 0.029933216094970704, 0.029972095489501954, 0.02988310432434082, 0.029998367309570312, 0.029862783432006837, 0.02997587203979492, 0.029763647079467773, 0.029921760559082033, 0.030057600021362305, 0.029989023208618164, 0.029456256866455078, 0.03273779296875, 0.029587808609008788, 0.02976563262939453, 0.029737056732177733, 0.029645023345947267, 0.02952556800842285, 0.02945510482788086, 0.02941939163208008, 0.029394559860229492, 0.03005900764465332, 0.030144800186157228, 0.02975062370300293, 0.03127539253234863, 0.0316682243347168, 0.029464799880981444, 0.029251359939575197, 0.02935980796813965, 0.02940883255004883, 0.029600704193115234, 0.02950124740600586, 0.029359455108642577, 0.029311647415161134, 0.0293919677734375, 0.02939792060852051, 0.029566719055175782, 0.029745248794555663, 0.02936412811279297, 0.029284000396728516, 0.029137504577636718, 0.029298688888549803, 0.029719680786132813, 0.029369216918945312, 0.029211872100830077, 0.029487136840820313, 0.029492992401123047, 0.029477632522583008, 0.02922710418701172, 0.02931727981567383, 0.0291843204498291, 0.02930246353149414, 0.029169824600219725, 0.029331296920776368, 0.02924473571777344, 0.029831584930419923, 0.029538591384887694, 0.030424383163452147, 0.029377056121826173, 0.029310367584228517, 0.029333248138427734, 0.02927622413635254, 0.02944000053405762, 0.0298853759765625, 0.029875232696533204, 0.029969375610351564, 0.02990825653076172, 0.029983583450317382, 0.02995712089538574, 0.03043008041381836, 0.030110879898071288, 0.030100320816040037, 0.029831167221069335, 0.029730815887451172, 0.029609024047851564, 0.029567935943603515, 0.030072832107543947, 0.02960963249206543, 0.029482656478881836, 0.030195072174072267, 0.030063072204589845, 0.0297043514251709, 0.02994380760192871, 0.030019872665405272, 0.03011555290222168, 0.02977177619934082, 0.03092230415344238, 0.029642784118652343, 0.029543872833251952, 0.029543392181396483, 0.02936832046508789, 0.029239456176757814, 0.02945212745666504, 0.0292938232421875, 0.029304927825927734, 0.02918671989440918, 0.02951366424560547, 0.029510751724243164, 0.02927631950378418, 0.029223167419433593, 0.02925609588623047, 0.029228799819946288, 0.029395360946655274, 0.02918604850769043, 0.031028608322143554, 0.029380640029907226, 0.02972313690185547, 0.029298784255981446, 0.029265920639038087, 0.029138751983642578, 0.029325504302978516, 0.02927568054199219, 0.029395423889160156, 0.029183935165405274, 0.029254880905151368, 0.02918684768676758, 0.029564992904663086, 0.02959564781188965, 0.0296376953125, 0.029434816360473633, 0.030011392593383788, 0.02958736038208008, 0.029698495864868165, 0.029296287536621092, 0.02936627197265625, 0.029203903198242186, 0.029497919082641603, 0.029418943405151367, 0.029855615615844728, 0.02991276741027832, 0.030143487930297853, 0.02991663932800293, 0.030113855361938478, 0.030249088287353516, 0.031141504287719727, 0.0313863353729248, 0.030803359985351563, 0.03851068878173828, 0.03006719970703125, 0.030251007080078125, 0.02997452735900879, 0.02997452735900879, 0.029560224533081055, 0.02951420783996582, 0.02946249580383301, 0.02939091110229492, 0.029276287078857422, 0.030003200531005858, 0.029709440231323242, 0.03000831985473633, 0.030301504135131836, 0.030228191375732422, 0.029833280563354492, 0.029932319641113283, 0.02980659294128418, 0.030270463943481447, 0.029850336074829103, 0.029861440658569337, 0.02988051223754883, 0.029539936065673827, 0.029411808013916015, 0.029499040603637696, 0.030976736068725585, 0.031403392791748044, 0.029716512680053712, 0.029505632400512696, 0.029166240692138672, 0.029353919982910155, 0.029355167388916015, 0.029335712432861327, 0.030500831604003905, 0.029557472229003907, 0.02954217529296875, 0.02983900833129883, 0.029892255783081054, 0.030061311721801758, 0.029940191268920897, 0.03137251281738281, 0.030804128646850587, 0.030271488189697264, 0.030097408294677733, 0.030144832611083985, 0.030078975677490235, 0.030187295913696288, 0.03158435249328613, 0.03023052787780762, 0.029703807830810548, 0.029869951248168946, 0.029751840591430663, 0.029884511947631837, 0.02935807991027832, 0.029362239837646485, 0.029264032363891603, 0.02976915168762207, 0.029820255279541016, 0.0297740478515625, 0.029489023208618164, 0.02946345520019531, 0.02955299186706543, 0.029916000366210938, 0.029719711303710938, 0.029836959838867187, 0.029792255401611328, 0.02980624008178711, 0.02968726348876953, 0.030058496475219725, 0.029822975158691405, 0.0294150390625, 0.02950387191772461, 0.02989023971557617, 0.030142784118652344, 0.02983065605163574, 0.029726688385009765, 0.02975119972229004, 0.030175775527954102, 0.030308448791503906, 0.030334848403930664, 0.0303187198638916, 0.030316543579101563, 0.030291263580322265, 0.030335647583007812, 0.030335008621215822, 0.030283775329589844, 0.03026918411254883, 0.030382335662841795, 0.030130048751831055, 0.03030847930908203, 0.0301977596282959, 0.030289920806884765, 0.03026652717590332, 0.03044233512878418, 0.030398527145385743, 0.030152191162109376, 0.030237024307250976, 0.03035935974121094, 0.030349599838256837, 0.030451488494873048, 0.030513568878173827, 0.030404415130615235, 0.030298112869262695, 0.03015884780883789, 0.02994937515258789, 0.03004195213317871, 0.02995248031616211, 0.02988368034362793, 0.029749887466430664, 0.030007232666015626, 0.030136735916137695, 0.031016960144042968, 0.030116031646728516, 0.030105152130126954, 0.030177087783813478, 0.030249408721923828, 0.02996396827697754, 0.0296342716217041, 0.029513919830322265, 0.0296595516204834, 0.02968560028076172, 0.029726879119873047, 0.02941244888305664, 0.029422464370727538, 0.029306304931640624, 0.029297088623046873, 0.029221023559570312, 0.029294048309326172, 0.02928220748901367, 0.02932150459289551, 0.029618528366088866]",tokens/s,33.517145251435736,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,7361.441792,9673.048064,0.0,9277.800448,8679.633408,s,1,14.699318359375,14.699318359375,0.0,14.699318359375,14.699318359375,14.699318359375,14.699318359375,[14.699318359375],,kWh,0.00022458091174167217,2.4765545809885814e-05,0.00010117424760598681,0.00035052070515754477,,MB,1702.240256,9687.728128,0.0,9277.800448,8206.575616,s,10,3.9106325683593752,0.3910632568359375,0.0018663111923173303,0.39105384826660156,0.3933038970947265,0.393489030456543,0.39363713714599613,"[0.3874965209960938, 0.391135498046875, 0.389873779296875, 0.39326275634765623, 0.389359375, 0.3922892150878906, 0.39097219848632814, 0.39278826904296876, 0.3936741638183594, 0.3897807922363281]",tokens/s,654.6255510458235,kWh,1.1371556762821183e-05,1.254069628126514e-06,7.523809437846008e-06,2.014943582879371e-05,tokens/kWh,12705070.364013562,MB,1709.735936,9687.728128,0.0,9277.800448,8480.92416,s,10,27.260906982421876,2.7260906982421877,0.004611978211197152,2.72476708984375,2.7334384521484374,2.7347204467773434,2.7357460424804687,"[2.73600244140625, 2.722218505859375, 2.724154296875, 2.7224716796875, 2.72430078125, 2.7252333984375, 2.733153564453125, 2.727108154296875, 2.72086767578125, 2.725396484375]",tokens/s,23.11001612698472,kWh,7.990248834301077e-05,8.813380497793079e-06,5.313982242635492e-05,0.00014185569126715872,tokens/kWh,444113.30583382276,,s,630,27.257424415588368,0.04326575304061648,0.0005153213214338492,0.04324404716491699,0.04369912796020508,0.04383510227203369,0.044511854667663574,"[0.044332801818847654, 0.0430247688293457, 0.04291785430908203, 0.0427848014831543, 0.04272111892700195, 0.042909854888916014, 0.04286873626708984, 0.042907230377197264, 0.04295644760131836, 0.04301491165161133, 0.043015167236328124, 0.043033695220947264, 0.04298160171508789, 0.042995391845703126, 0.043118495941162106, 0.043006046295166016, 0.043020286560058595, 0.04312473678588867, 0.04309961700439453, 0.04318832015991211, 0.04314156723022461, 0.04345644760131836, 0.04329177474975586, 0.04335647964477539, 0.043412094116210935, 0.04341964721679688, 0.04326604843139648, 0.04335030364990235, 0.04332073593139649, 0.043227455139160154, 0.0434442253112793, 0.04352329635620117, 0.043420448303222656, 0.04374489593505859, 0.043909503936767576, 0.04348435211181641, 0.043379360198974606, 0.04346428680419922, 0.04323142242431641, 0.043342529296875, 0.04341292953491211, 0.043450366973876955, 0.04362060928344726, 0.04398899078369141, 0.044434593200683596, 0.043754112243652346, 0.043730335235595705, 0.04370105743408203, 0.04369612884521484, 0.04380876922607422, 0.043665279388427736, 0.04361804962158203, 0.043573631286621096, 0.043640830993652346, 0.043586784362792966, 0.043488033294677736, 0.04379852676391602, 0.043597824096679685, 0.043816543579101565, 0.04371836853027344, 0.04365382385253906, 0.044771007537841793, 0.043883968353271484, 0.043439903259277345, 0.04291609573364258, 0.042816062927246094, 0.04277199935913086, 0.04269481658935547, 0.04281788635253906, 0.04276633453369141, 0.04280316925048828, 0.04298873519897461, 0.042793502807617186, 0.04288262557983399, 0.042912513732910156, 0.04285356903076172, 0.04270703887939453, 0.042787105560302734, 0.042775104522705075, 0.04285222244262695, 0.04289945602416992, 0.0429035530090332, 0.04287692642211914, 0.04310015869140625, 0.04300799942016602, 0.04298956680297852, 0.04299980926513672, 0.04294451141357422, 0.042869888305664065, 0.043047264099121095, 0.04336080169677734, 0.04326604843139648, 0.04319027328491211, 0.04314726257324219, 0.04316310501098633, 0.043264545440673825, 0.04316364669799805, 0.04320460891723633, 0.043251361846923825, 0.043211105346679685, 0.043243518829345705, 0.04333286285400391, 0.04340403366088867, 0.04332748794555664, 0.04335558319091797, 0.043407745361328125, 0.04358777618408203, 0.043574783325195314, 0.043582977294921874, 0.04373811340332031, 0.04374262237548828, 0.043651649475097656, 0.043569183349609374, 0.043466625213623045, 0.043407360076904294, 0.043606143951416015, 0.043482208251953126, 0.0436069450378418, 0.04355481719970703, 0.04350960159301758, 0.043408576965332034, 0.043464702606201173, 0.043453407287597653, 0.04361795043945312, 0.04360022354125977, 0.043720703125, 0.04343603134155274, 0.042866687774658206, 0.04262236785888672, 0.04265635299682617, 0.04250848007202149, 0.04276416015625, 0.04268025588989258, 0.04278681564331055, 0.042782718658447266, 0.042970912933349606, 0.04290380859375, 0.04292809677124024, 0.042872127532958985, 0.04294112014770508, 0.042913791656494144, 0.042976734161376956, 0.04295100784301758, 0.04301433563232422, 0.043071487426757815, 0.04294595336914062, 0.042930782318115236, 0.04305817413330078, 0.04298425674438477, 0.043122879028320314, 0.0433144645690918, 0.04326268768310547, 0.043466110229492184, 0.043270496368408205, 0.043190559387207034, 0.043283870697021484, 0.04346303939819336, 0.043251361846923825, 0.04322361755371094, 0.043081729888916016, 0.04319968032836914, 0.04326688003540039, 0.043200511932373044, 0.04324966430664062, 0.043210750579833986, 0.04319232177734375, 0.0433889274597168, 0.04354457473754883, 0.04364287948608398, 0.043656864166259766, 0.0433606071472168, 0.04348089599609375, 0.04334012985229492, 0.04325334548950195, 0.04331315231323242, 0.04350592041015625, 0.04357120132446289, 0.04347916793823242, 0.043638656616210934, 0.043566497802734375, 0.04363734436035156, 0.04346255874633789, 0.04348118209838867, 0.043595775604248044, 0.04356025695800781, 0.04366815948486328, 0.04392499160766602, 0.04358563232421875, 0.04438457489013672, 0.0434400634765625, 0.04302403259277344, 0.04281135940551758, 0.04287238311767578, 0.042906494140625, 0.042796737670898435, 0.04268828964233398, 0.04268086242675781, 0.04282572937011719, 0.042692607879638675, 0.042780670166015625, 0.04280319976806641, 0.04269875335693359, 0.04268032073974609, 0.04272272109985351, 0.043618144989013674, 0.04297395324707031, 0.0429567985534668, 0.043216705322265625, 0.04286896133422852, 0.04317385482788086, 0.04279500961303711, 0.04287078475952148, 0.042958175659179684, 0.042937183380126955, 0.042942272186279294, 0.04305820846557617, 0.04296803283691406, 0.04289535903930664, 0.04299353790283203, 0.04295897674560547, 0.04293017578125, 0.042976734161376956, 0.04286307144165039, 0.0429788818359375, 0.04293616104125977, 0.043006622314453125, 0.043169792175292966, 0.04346265411376953, 0.04341862487792969, 0.04374591827392578, 0.043534465789794925, 0.043596031188964844, 0.04351180648803711, 0.04350080108642578, 0.0434793586730957, 0.04360768127441406, 0.043600704193115236, 0.04374118423461914, 0.04394188690185547, 0.043499519348144534, 0.04333059310913086, 0.04335030364990235, 0.043527072906494144, 0.043568511962890626, 0.04360768127441406, 0.04363241577148438, 0.04357017517089844, 0.04393926239013672, 0.04435811233520508, 0.04372079849243164, 0.04365641784667969, 0.04374195098876953, 0.04372480010986328, 0.04337254333496094, 0.042802433013916015, 0.042893505096435545, 0.042897022247314454, 0.042959392547607424, 0.04281180953979492, 0.04286873626708984, 0.04279296112060547, 0.042755073547363284, 0.042733791351318356, 0.04282857513427735, 0.04297865676879883, 0.042900127410888673, 0.04295503997802735, 0.04300156784057617, 0.0429752311706543, 0.04315135955810547, 0.04296499252319336, 0.04272127914428711, 0.04278019332885742, 0.04293795013427734, 0.043039169311523434, 0.04310675048828125, 0.0429486083984375, 0.04297727966308594, 0.043018047332763674, 0.042866878509521485, 0.04282284927368164, 0.042826560974121096, 0.043001697540283206, 0.042977439880371095, 0.04295993423461914, 0.04301900863647461, 0.04299795150756836, 0.04308153533935547, 0.043063392639160154, 0.043012001037597655, 0.043753246307373046, 0.0432880973815918, 0.04354342269897461, 0.0436121597290039, 0.0437125129699707, 0.0436420783996582, 0.04352617645263672, 0.04347110366821289, 0.04354508972167969, 0.04364492797851562, 0.04367708969116211, 0.04364064025878906, 0.04353513717651367, 0.04346060943603516, 0.04364239883422852, 0.043638336181640626, 0.04368681716918945, 0.04362035369873047, 0.04375475311279297, 0.04371737670898437, 0.04384143829345703, 0.043843616485595704, 0.043763137817382815, 0.043866142272949216, 0.04393020629882813, 0.04337667083740234, 0.04297318267822266, 0.04278598403930664, 0.04279142379760742, 0.04281171035766602, 0.04288710403442383, 0.04270700836181641, 0.04290566253662109, 0.04292396926879883, 0.04293974304199219, 0.042969024658203125, 0.04317843246459961, 0.04302876663208008, 0.0429854736328125, 0.04279500961303711, 0.04274393463134766, 0.042785888671875, 0.042848865509033204, 0.042850494384765625, 0.04291788864135742, 0.043245567321777346, 0.0431280632019043, 0.043014881134033206, 0.04306742477416992, 0.04308287811279297, 0.04304780960083008, 0.043096065521240234, 0.04297929763793945, 0.043071521759033206, 0.04295232009887695, 0.04297881698608398, 0.043101375579833984, 0.04331692886352539, 0.043278335571289066, 0.04344627380371094, 0.04338220977783203, 0.04336934280395508, 0.043306686401367187, 0.04329062271118164, 0.0433704948425293, 0.043412734985351566, 0.04351052856445312, 0.04354281616210937, 0.043568031311035156, 0.0435513916015625, 0.04344380950927734, 0.043501758575439455, 0.04361644744873047, 0.04358777618408203, 0.04364278411865234, 0.0433337287902832, 0.04330223846435547, 0.0432831039428711, 0.04330617523193359, 0.04338771057128906, 0.04368368148803711, 0.04362460708618164, 0.04395622253417969, 0.04508262252807617, 0.043671550750732424, 0.043589630126953126, 0.0437037124633789, 0.04382735824584961, 0.0434411506652832, 0.043061855316162106, 0.0432371826171875, 0.043262081146240236, 0.04292214584350586, 0.04306361770629883, 0.042711231231689455, 0.04272297668457031, 0.04269641494750977, 0.042654144287109376, 0.0426618881225586, 0.04265331268310547, 0.042936447143554685, 0.042895614624023436, 0.04315280151367187, 0.04305107116699219, 0.04309660720825195, 0.04348652648925781, 0.04330060958862305, 0.04322604751586914, 0.04332672119140625, 0.04326633453369141, 0.043377120971679686, 0.043259902954101564, 0.043151039123535156, 0.043180351257324216, 0.0432803840637207, 0.04316159820556641, 0.04325145721435547, 0.04331545639038086, 0.04329471969604492, 0.04463411331176758, 0.04345446395874023, 0.04345427322387695, 0.04338297653198242, 0.04348880004882812, 0.04342012786865234, 0.04355871963500976, 0.043509952545166014, 0.04357510375976562, 0.043535839080810546, 0.04369891357421875, 0.043617313385009765, 0.04362908935546875, 0.04385836791992188, 0.043703968048095704, 0.043563358306884764, 0.043622398376464845, 0.043619552612304685, 0.04364287948608398, 0.04361091232299805, 0.04351571273803711, 0.043788734436035155, 0.04378319931030274, 0.04368252944946289, 0.04387593460083008, 0.04371830368041992, 0.04358000183105469, 0.04358979034423828, 0.043606048583984376, 0.043748737335205075, 0.04362284851074219, 0.043652320861816404, 0.043533790588378904, 0.044195999145507814, 0.0440280647277832, 0.04262928009033203, 0.04283193588256836, 0.04286089706420899, 0.04354601669311523, 0.04268809509277344, 0.042877342224121096, 0.04268672180175781, 0.04270284652709961, 0.042987518310546875, 0.04306905746459961, 0.04310784149169922, 0.043103073120117186, 0.043154464721679685, 0.04303936004638672, 0.04306777572631836, 0.04338399887084961, 0.042932064056396484, 0.04306182479858398, 0.04311286544799805, 0.04320390319824219, 0.04305990219116211, 0.042977310180664065, 0.04290505599975586, 0.043020416259765625, 0.042996097564697265, 0.04331670379638672, 0.043221534729003905, 0.04317184066772461, 0.043024574279785156, 0.042960704803466795, 0.04293632125854492, 0.04305500793457031, 0.043144641876220705, 0.04311312103271484, 0.043075649261474606, 0.04313081741333008, 0.0433493766784668, 0.043168384552001955, 0.04335177612304687, 0.04317734527587891, 0.043283073425292966, 0.04338307189941406, 0.043340991973876954, 0.043348094940185544, 0.04329337692260742, 0.04325785446166992, 0.0434524154663086, 0.04361830520629883, 0.043548671722412106, 0.044611583709716796, 0.04356915283203125, 0.04355184173583984, 0.0434984016418457, 0.043705665588378906, 0.04373984146118164, 0.043888641357421876, 0.043869857788085935, 0.04391302490234375, 0.04386051177978516, 0.04400672149658203, 0.04355939102172852, 0.042865951538085936, 0.04258614349365234, 0.04248569488525391, 0.04257244873046875, 0.04274720001220703, 0.04273231887817383, 0.04284928131103516, 0.04286566543579102, 0.04287257766723633, 0.04290790557861328, 0.04290969467163086, 0.04290560150146484, 0.042997760772705076, 0.04299273681640625, 0.0431748161315918, 0.04299980926513672, 0.04316524887084961, 0.04323116683959961, 0.043059711456298826, 0.04292403030395508, 0.042923263549804684, 0.043018142700195314, 0.04315024185180664, 0.04326918411254883, 0.0434837760925293, 0.0433073616027832, 0.043552574157714845, 0.0432720947265625, 0.04322732925415039, 0.043202560424804685, 0.043128353118896484, 0.04307535934448242, 0.043167774200439456, 0.043184383392333985, 0.04308022308349609, 0.04314825439453125, 0.04313999938964844, 0.04315545654296875, 0.043457855224609376, 0.04343836975097656, 0.04341292953491211, 0.043330303192138674, 0.04330928039550781, 0.043218944549560545, 0.04320367813110351, 0.04324457550048828, 0.04321225738525391, 0.04325827026367188, 0.04326604843139648, 0.04341705703735352, 0.04450358581542969, 0.0433540153503418, 0.04321699142456055, 0.0433438720703125, 0.043319297790527345, 0.04336435317993164, 0.04348518371582031, 0.04327980804443359, 0.043399646759033204, 0.04363888168334961, 0.04339712142944336, 0.04362035369873047, 0.04355619049072266, 0.043439071655273435, 0.042616832733154295, 0.04238240051269531, 0.04244985580444336, 0.042743743896484374, 0.04248735809326172, 0.042514240264892575, 0.04259910583496094, 0.042555679321289064, 0.0427168960571289, 0.04266099166870117, 0.04264230346679688, 0.042724895477294925, 0.04263958358764648, 0.04267852783203125, 0.04281958389282227, 0.043433822631835935, 0.04359164810180664, 0.04291193771362305, 0.04339731216430664, 0.04283577728271484, 0.04294451141357422, 0.042891456604003904, 0.042954879760742186, 0.042968769073486325, 0.04306534576416016, 0.042807296752929686, 0.042815486907958986, 0.04288713455200195, 0.04287491226196289, 0.04294041442871094, 0.04296467208862305, 0.04306771087646484, 0.043153408050537106, 0.0430582389831543, 0.04305811309814453, 0.04301619338989258, 0.04301824188232422, 0.04299980926513672, 0.04319811248779297, 0.043259326934814456, 0.04318048095703125, 0.04327267074584961, 0.04334796905517578, 0.04329062271118164, 0.04336435317993164, 0.044539905548095705, 0.043695903778076174, 0.04337276840209961, 0.04334796905517578, 0.0432988166809082, 0.043369823455810544, 0.04329439926147461, 0.04327932739257812, 0.04338687896728516, 0.04451523208618164, 0.05200291061401367, 0.043546302795410156, 0.04367148971557617, 0.04372848129272461, 0.04356380844116211, 0.043681793212890625]",tokens/s,23.112968796850307,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,8205.709312,12277.710848,0.0,11882.463232,11315.947008,s,1,17.264017578125,17.264017578125,0.0,17.264017578125,17.264017578125,17.264017578125,17.264017578125,[17.264017578125],,kWh,0.0002955891702458151,3.259817562144908e-05,0.0001235167654800029,0.00045170411134726706,,MB,4005.691392,12294.488064,0.0,11884.560384,11070.308352,s,10,4.10347378540039,0.410347378540039,0.0023673147913827546,0.41062973022460936,0.41252118530273435,0.4131544372558594,0.4136610388183594,"[0.40438449096679685, 0.40964364624023436, 0.409791259765625, 0.4118450622558594, 0.40939697265625, 0.41238046264648437, 0.4106231994628906, 0.4137876892089844, 0.4106362609863281, 0.4109847412109375]",tokens/s,623.8616679136923,kWh,1.1888004602831642e-05,1.311049485645935e-06,7.907128547919307e-06,2.1106182636396885e-05,tokens/kWh,12129147.388241434,MB,4010.016768,12296.585216,0.0,11886.657536,11070.310912,s,10,26.82674755859375,2.6826747558593746,0.010927280065343218,2.6856224365234374,2.692194482421875,2.69243134765625,2.69262083984375,"[2.686230224609375, 2.684601318359375, 2.692668212890625, 2.692141845703125, 2.676589111328125, 2.67872265625, 2.689929443359375, 2.687415283203125, 2.6850146484375, 2.653434814453125]",tokens/s,23.484024614761182,kWh,7.839160972883474e-05,8.645532575390907e-06,5.198583603308493e-05,0.0001390229783373106,tokens/kWh,453162.49697329523,,s,630,26.82373398590088,0.042577355533176,0.0006445835464012684,0.04254539108276367,0.04294107666015625,0.043384997367858884,0.04486556758880615,"[0.043501632690429684, 0.04275251388549805, 0.04272435379028321, 0.04483391952514648, 0.04284735870361328, 0.04276211166381836, 0.042699840545654295, 0.04291078567504883, 0.042758495330810546, 0.04269456100463867, 0.04260704040527344, 0.04274591827392578, 0.04262297439575195, 0.04490230560302735, 0.04297292709350586, 0.04265609741210938, 0.04277372741699219, 0.04230633544921875, 0.042858592987060545, 0.04202249526977539, 0.042059391021728516, 0.04201542282104492, 0.04254924774169922, 0.04251456069946289, 0.04257984161376953, 0.042471038818359376, 0.04250252914428711, 0.04264495849609375, 0.04286313629150391, 0.04249107360839844, 0.04237750244140625, 0.04495772933959961, 0.04283801651000976, 0.042490367889404294, 0.042102783203125, 0.04202278518676758, 0.04231900787353516, 0.04197036743164063, 0.042092384338378905, 0.042482112884521486, 0.04273097610473633, 0.04228137588500976, 0.04199436950683594, 0.04249379348754883, 0.04229062271118164, 0.04207904052734375, 0.0421313591003418, 0.042012481689453124, 0.04207952117919922, 0.042036128997802735, 0.04192387390136719, 0.04188835144042969, 0.04202713775634766, 0.04229439926147461, 0.04497087860107422, 0.042567455291748046, 0.04249212646484375, 0.042893310546875, 0.0426618881225586, 0.042807296752929686, 0.042485759735107424, 0.04276736068725586, 0.04273180770874024, 0.043911167144775394, 0.043012096405029294, 0.04281756973266602, 0.04317910385131836, 0.04297817611694336, 0.04275609588623047, 0.04257791900634766, 0.04256300735473633, 0.04261891174316406, 0.04271971130371094, 0.04254111862182617, 0.04229251098632812, 0.04233903884887695, 0.042670177459716796, 0.043022174835205075, 0.043068702697753904, 0.042918689727783205, 0.042966846466064454, 0.04287875366210937, 0.042690464019775394, 0.04270111846923828, 0.042686527252197265, 0.04294614410400391, 0.04288092803955078, 0.042809982299804684, 0.04272332763671875, 0.04267753601074219, 0.04266672134399414, 0.04242777633666992, 0.0423166389465332, 0.04217977523803711, 0.042594913482666016, 0.042491584777832034, 0.04250041580200195, 0.042536224365234375, 0.042465568542480465, 0.04303436660766601, 0.042547393798828125, 0.04246796798706055, 0.0426453742980957, 0.042573631286621096, 0.04243001556396484, 0.042346942901611326, 0.042324161529541014, 0.042223617553710936, 0.04207779312133789, 0.04221993637084961, 0.042143489837646486, 0.042047744750976564, 0.0421847038269043, 0.042477569580078124, 0.042575870513916016, 0.04244275283813476, 0.04239772796630859, 0.04232815933227539, 0.04229107284545899, 0.04245094299316406, 0.04273356628417969, 0.04266377639770508, 0.04258832168579101, 0.042815486907958986, 0.04264243316650391, 0.04250457763671875, 0.044290046691894534, 0.042940513610839844, 0.04259939193725586, 0.04258297729492187, 0.042437793731689454, 0.04241107177734375, 0.042444576263427736, 0.04249190521240234, 0.04271712112426758, 0.04315347290039063, 0.042297344207763675, 0.0425984001159668, 0.042412033081054686, 0.04280902481079102, 0.04364729690551758, 0.04273766326904297, 0.0427061767578125, 0.04295657730102539, 0.042742591857910156, 0.04269686508178711, 0.04276211166381836, 0.042649726867675784, 0.04252467346191406, 0.04232774353027344, 0.042232128143310545, 0.04251446533203125, 0.04235424041748047, 0.04212524795532226, 0.0421278076171875, 0.0424400634765625, 0.044878494262695315, 0.04339616012573242, 0.043242431640625, 0.04276838302612305, 0.04268646240234375, 0.042651519775390626, 0.04252419281005859, 0.042977886199951174, 0.04281913757324219, 0.042654144287109376, 0.04276838302612305, 0.04264316940307617, 0.042540863037109376, 0.04277478408813477, 0.04276860809326172, 0.04254515075683594, 0.04261273574829102, 0.04272332763671875, 0.04270467376708984, 0.04286076736450195, 0.04259193420410156, 0.042860702514648436, 0.04254521560668945, 0.04269884872436523, 0.0424238395690918, 0.042394081115722654, 0.04249391937255859, 0.043276321411132815, 0.04261177444458008, 0.04250236892700195, 0.042844894409179685, 0.04321206283569336, 0.04261552047729492, 0.0442716178894043, 0.042753631591796876, 0.042307998657226564, 0.04288723373413086, 0.042493343353271484, 0.042057857513427735, 0.042111392974853515, 0.041924606323242186, 0.041944671630859375, 0.04224665451049805, 0.042170272827148435, 0.04275404739379883, 0.042671104431152344, 0.04272854232788086, 0.042444225311279296, 0.0423419189453125, 0.04285488128662109, 0.04240547180175781, 0.04243775939941406, 0.04383513641357422, 0.04365311813354492, 0.04306707382202148, 0.04256784057617188, 0.042587871551513674, 0.04268076705932617, 0.042261566162109375, 0.042098945617675784, 0.042123966217041016, 0.042487808227539066, 0.04315107345581055, 0.04422275161743164, 0.04238131332397461, 0.04246323013305664, 0.04245276641845703, 0.04237305450439453, 0.04270713424682617, 0.05246985626220703, 0.042667072296142576, 0.04254816055297852, 0.04258531188964844, 0.042728225708007814, 0.04274176025390625, 0.04255539321899414, 0.042385406494140625, 0.042288448333740236, 0.04254175949096679, 0.0425634880065918, 0.04234454345703125, 0.04244070434570312, 0.04224975967407227, 0.042141185760498044, 0.04230038452148437, 0.04200803375244141, 0.04290614318847656, 0.04225820922851563, 0.042126911163330075, 0.042500926971435544, 0.042528606414794924, 0.042842113494873046, 0.04270489501953125, 0.04256358337402344, 0.04256972885131836, 0.04237516784667969, 0.0435464973449707, 0.042721473693847656, 0.04249564743041992, 0.04232556915283203, 0.04222032165527344, 0.04193231964111328, 0.04200899124145508, 0.04176796722412109, 0.04195836639404297, 0.04318560028076172, 0.042439231872558596, 0.04254719924926758, 0.041895553588867186, 0.04224192047119141, 0.04200089645385742, 0.041780353546142575, 0.04175347137451172, 0.04171571350097656, 0.041662464141845705, 0.041769088745117186, 0.041713409423828125, 0.041941120147705076, 0.042000385284423826, 0.041924606323242186, 0.04181180953979492, 0.04174454498291016, 0.04216012954711914, 0.042708992004394535, 0.04246246337890625, 0.04246966552734375, 0.042621086120605466, 0.04356537628173828, 0.04259561538696289, 0.04235951995849609, 0.04242419052124023, 0.04246745681762695, 0.04254828643798828, 0.042372032165527346, 0.04234444808959961, 0.0422553596496582, 0.042719905853271484, 0.04274825668334961, 0.042805248260498044, 0.04289311981201172, 0.042778785705566404, 0.042829631805419925, 0.04307785415649414, 0.04285747146606445, 0.042843135833740234, 0.04312684631347656, 0.04303209686279297, 0.04341801452636719, 0.04295862579345703, 0.04287919998168945, 0.04274723052978516, 0.0424925765991211, 0.04352934265136719, 0.04282457733154297, 0.04281958389282227, 0.04254272079467773, 0.04259811019897461, 0.04266870498657226, 0.042627071380615236, 0.044226974487304685, 0.04269049453735352, 0.04276633453369141, 0.04273366546630859, 0.043165889739990235, 0.04268431854248047, 0.04268246459960937, 0.042544990539550784, 0.04251772689819336, 0.04237561416625977, 0.0419920654296875, 0.04256422424316406, 0.04281753540039063, 0.04244831848144531, 0.04254556655883789, 0.042393470764160154, 0.042436126708984376, 0.04258278274536133, 0.042401695251464845, 0.04241046524047851, 0.04235385513305664, 0.042400032043457034, 0.042436126708984376, 0.04209651184082031, 0.04199068832397461, 0.04229347229003906, 0.04203519821166992, 0.042266334533691406, 0.04194537734985351, 0.04257344055175781, 0.04290524673461914, 0.042672863006591795, 0.04278681564331055, 0.04276614379882813, 0.04267436981201172, 0.04274995040893555, 0.04251238250732422, 0.042362144470214844, 0.04242095947265625, 0.04254515075683594, 0.04252262496948242, 0.04220284652709961, 0.04222800064086914, 0.04226047897338867, 0.042315616607666015, 0.04255964660644531, 0.042469375610351565, 0.04426137542724609, 0.042641407012939454, 0.042412033081054686, 0.042007839202880856, 0.041931488037109374, 0.04187564849853516, 0.041993633270263675, 0.04226063919067383, 0.04234415817260742, 0.04256335830688476, 0.04277119827270508, 0.04261177444458008, 0.04246163177490234, 0.04250864028930664, 0.04255132675170899, 0.04290342330932617, 0.04420198440551758, 0.04258127975463867, 0.042500831604003905, 0.04232912063598633, 0.042627777099609375, 0.0428372802734375, 0.042369281768798825, 0.042479679107666014, 0.042355392456054686, 0.04247951889038086, 0.042446945190429686, 0.04247251129150391, 0.042750911712646486, 0.04312659072875977, 0.04262844848632812, 0.04272828674316406, 0.04259183883666992, 0.042918304443359374, 0.04272335815429688, 0.042614368438720705, 0.042961246490478514, 0.042584095001220706, 0.04245913696289062, 0.04275209426879883, 0.04256377410888672, 0.04277148818969727, 0.04248441696166992, 0.04228083038330078, 0.04243878555297852, 0.042280960083007815, 0.04203891372680664, 0.042596416473388674, 0.04266630554199219, 0.04279001617431641, 0.04248681640625, 0.04258390426635742, 0.043319297790527345, 0.043439422607421875, 0.04283875274658203, 0.04262470245361328, 0.04286102294921875, 0.04260230255126953, 0.04247347259521484, 0.04258518218994141, 0.0425337905883789, 0.042581600189208986, 0.0430450553894043, 0.0464304313659668, 0.04254924774169922, 0.04254832077026367, 0.04247644805908203, 0.04269875335693359, 0.04263270568847656, 0.0423818244934082, 0.042524574279785156, 0.042584224700927736, 0.04246931076049805, 0.042657791137695314, 0.042313728332519535, 0.0421899528503418, 0.0424986572265625, 0.04264966583251953, 0.04258806228637695, 0.04410713577270508, 0.0425109748840332, 0.04253462219238281, 0.0425126724243164, 0.04274576187133789, 0.042840160369873044, 0.042813438415527344, 0.042622081756591795, 0.042621822357177735, 0.04263935852050781, 0.042657791137695314, 0.042659103393554686, 0.04278550338745117, 0.04265545654296875, 0.0426396484375, 0.042602497100830077, 0.04580966567993164, 0.04293545532226563, 0.04277948760986328, 0.042863807678222655, 0.042477569580078124, 0.04238828659057617, 0.04226662445068359, 0.04220060729980469, 0.04205615997314453, 0.04234588623046875, 0.04337315368652344, 0.042651134490966795, 0.04261119842529297, 0.04315545654296875, 0.042759166717529294, 0.04267708969116211, 0.04240809631347656, 0.04250419235229492, 0.04277174377441406, 0.04267273712158203, 0.04265382385253906, 0.04268387222290039, 0.04290204620361328, 0.04235887908935547, 0.04243638229370117, 0.04219302368164062, 0.04245836639404297, 0.04221807861328125, 0.04219510269165039, 0.0421847038269043, 0.042106399536132814, 0.042080734252929686, 0.0424648323059082, 0.04264595031738281, 0.04260182571411133, 0.04274607849121094, 0.04264595031738281, 0.042851329803466794, 0.04237318420410156, 0.042099391937255856, 0.04235494232177734, 0.042616832733154295, 0.042683712005615236, 0.042764991760253904, 0.042757793426513674, 0.04279331207275391, 0.042584064483642575, 0.04441497421264649, 0.042592254638671875, 0.042450241088867184, 0.04257452774047851, 0.04339468765258789, 0.04274367904663086, 0.04264988708496094, 0.04322531127929687, 0.04286777496337891, 0.04266870498657226, 0.04265158462524414, 0.04270700836181641, 0.04265577697753906, 0.04251382446289063, 0.04184764862060547, 0.04190003204345703, 0.04205491256713867, 0.04184262466430664, 0.042110912322998045, 0.042580863952636716, 0.04251561737060547, 0.042326881408691404, 0.04223081588745117, 0.042120159149169924, 0.04215193557739258, 0.042162174224853514, 0.04456447982788086, 0.0425799674987793, 0.042732864379882815, 0.04285305786132813, 0.04272681427001953, 0.042637855529785156, 0.04255340957641601, 0.042559486389160156, 0.04275814437866211, 0.042461185455322265, 0.042309406280517575, 0.042109153747558595, 0.042102497100830076, 0.042514144897460936, 0.04229561614990234, 0.0421844482421875, 0.042738174438476564, 0.042602497100830077, 0.04327443313598633, 0.042786624908447264, 0.04288857650756836, 0.04346944046020508, 0.04284963226318359, 0.042824512481689454, 0.04277027130126953, 0.04271059036254883, 0.04259657669067383, 0.04275001525878906, 0.04272348785400391, 0.042520416259765624, 0.04263033676147461, 0.04240083312988281, 0.042542144775390624, 0.04218352127075195, 0.042196990966796875, 0.042539199829101565, 0.04279072189331055, 0.04378009414672852, 0.042456321716308594, 0.04241648101806641, 0.04246774291992188, 0.04230144119262695, 0.04222524642944336, 0.042685951232910156, 0.04234902572631836, 0.042090110778808594, 0.042259265899658206, 0.04195145416259766, 0.04196697616577148, 0.0415830078125, 0.04163174438476563, 0.04175667190551758, 0.04165631866455078, 0.041629344940185546, 0.04199590301513672, 0.042908000946044925, 0.04212569427490234, 0.04189593505859375, 0.04201251220703125, 0.042135711669921874, 0.04204262542724609, 0.041855743408203125, 0.041891136169433595, 0.04198393630981445, 0.04217948913574219, 0.042016254425048825, 0.04184099197387695, 0.04194713592529297, 0.04203699111938477, 0.04195353698730469, 0.042229888916015625, 0.04293824005126953, 0.042169921875, 0.04175676727294922, 0.04191635131835938, 0.04170163345336914, 0.041727649688720704, 0.04182419204711914, 0.041877567291259764, 0.042229984283447264, 0.042162464141845706, 0.04213145446777344, 0.042159393310546876, 0.04204822540283203, 0.041998336791992184, 0.04198604965209961, 0.041989311218261716, 0.041960254669189456, 0.04201062393188477, 0.04195062255859375, 0.043493824005126955, 0.042765857696533204, 0.042019454956054685, 0.041957023620605466, 0.04223420715332031, 0.04198809432983398, 0.04198809432983398, 0.041908222198486327, 0.04201027297973633, 0.042020671844482424]",tokens/s,23.48666298029727,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,848.891904,558.825472,0.0,163.577856,154.631168,s,1,7.92300146484375,7.92300146484375,0.0,7.92300146484375,7.92300146484375,7.92300146484375,7.92300146484375,[7.92300146484375],,kWh,2.2599866625023424e-05,2.4855643478874524e-06,7.5397282540135e-06,3.262515922692438e-05,,MB,1229.242368,609.15712,0.0,199.22944,187.147776,s,25,0.19424972772598267,0.007769989109039306,0.0002163029875023168,0.007699391841888428,0.007972064208984376,0.008142291069030761,0.008529240531921386,"[0.008082976341247558, 0.007766304016113282, 0.007660768032073975, 0.007634943962097168, 0.00768995189666748, 0.008157119750976563, 0.007699391841888428, 0.007702784061431885, 0.007675039768218994, 0.0078056960105895995, 0.007701280117034912, 0.007663839817047119, 0.007749184131622315, 0.007765471935272217, 0.007688096046447754, 0.007707424163818359, 0.007665760040283203, 0.00864675235748291, 0.007653791904449463, 0.007707615852355957, 0.007694464206695556, 0.007737055778503418, 0.0076406078338623045, 0.007692543983459472, 0.007660863876342773]",tokens/s,32947.279128381204,kWh,2.25515588333984e-07,2.4870321807205812e-08,1.0638918866615677e-07,3.567750988073466e-07,tokens/kWh,717538866.5178012,MB,1265.037312,611.254272,0.0,201.326592,187.150336,s,25,9.943050720214842,0.3977220288085938,0.0026772775040471266,0.3964454040527344,0.40168916625976564,0.40314699096679685,0.4038981884765625,"[0.39874029541015626, 0.396388671875, 0.39532431030273435, 0.39672189331054686, 0.3955487976074219, 0.4033482971191406, 0.39606280517578124, 0.396182861328125, 0.40071026611328125, 0.39798458862304686, 0.39671002197265626, 0.4002525329589844, 0.3995301513671875, 0.40234176635742186, 0.3990752563476562, 0.3964454040527344, 0.39581195068359376, 0.39555010986328126, 0.39614224243164065, 0.394832275390625, 0.3995881652832031, 0.40407183837890626, 0.395975341796875, 0.3956080627441406, 0.3941028137207031]",tokens/s,158.40208848557177,kWh,1.1427905487627088e-05,1.2603048490791262e-06,4.2150779437173385e-06,1.690328828042355e-05,tokens/kWh,3727085.461410671,,s,1575,9.931675614833836,0.006305825787196084,0.00013328746779218132,0.006275167942047119,0.006393958282470704,0.00645662407875061,0.006788728313446045,"[0.00632425594329834, 0.006336639881134034, 0.006358399868011475, 0.006400479793548584, 0.006379551887512207, 0.006355231761932373, 0.006375135898590088, 0.006479743957519531, 0.006611072063446045, 0.006379263877868652, 0.006393983840942383, 0.006370463848114014, 0.006386655807495117, 0.006393919944763184, 0.006348608016967773, 0.006373600006103515, 0.006319712162017823, 0.006422719955444336, 0.006295711994171143, 0.006266848087310791, 0.006301695823669433, 0.006322336196899414, 0.0063047041893005374, 0.0062432317733764646, 0.006288896083831787, 0.006272736072540283, 0.00624124813079834, 0.006289472103118896, 0.006262527942657471, 0.006469632148742676, 0.006276288032531738, 0.0063187842369079586, 0.0062772479057312015, 0.006254591941833496, 0.006276447772979737, 0.00623683214187622, 0.006290976047515869, 0.0062243518829345704, 0.006272895812988281, 0.006290719985961914, 0.006237184047698975, 0.006279104232788086, 0.006223775863647461, 0.006323904037475586, 0.0064085121154785155, 0.006354944229125976, 0.006266880035400391, 0.006272928237915039, 0.006350272178649903, 0.006375967979431152, 0.006404287815093994, 0.0063056640625, 0.006303616046905517, 0.0064347519874572755, 0.006286816120147705, 0.0062882242202758785, 0.006277376174926757, 0.006291071891784668, 0.0062873601913452145, 0.0062583680152893065, 0.006248832225799561, 0.006264512062072754, 0.006249855995178222, 0.006164480209350586, 0.00655072021484375, 0.006279232025146485, 0.00624886417388916, 0.006273248195648193, 0.006254720211029053, 0.006307839870452881, 0.0062341117858886715, 0.006309887886047363, 0.006237279891967773, 0.00656060791015625, 0.006297440052032471, 0.006347040176391602, 0.006348832130432129, 0.006268703937530517, 0.006287487983703613, 0.006260831832885742, 0.006250400066375733, 0.006262784004211426, 0.006247424125671387, 0.006267199993133545, 0.006670559883117676, 0.006274687767028809, 0.006243167877197265, 0.006375423908233643, 0.006297535896301269, 0.006223936080932618, 0.006307839870452881, 0.006260960102081299, 0.00632374382019043, 0.006303999900817871, 0.006266880035400391, 0.006305791854858398, 0.006315455913543701, 0.006288032054901123, 0.006385471820831299, 0.006283360004425049, 0.006331552028656006, 0.006260608196258545, 0.006252960205078125, 0.006264736175537109, 0.006234399795532227, 0.006238463878631592, 0.006240384101867676, 0.006254687786102295, 0.0062585282325744625, 0.006261856079101563, 0.006203904151916504, 0.006265312194824219, 0.0062392959594726564, 0.006230815887451172, 0.006254752159118652, 0.006246399879455566, 0.006259007930755615, 0.006232992172241211, 0.006255231857299805, 0.006213151931762696, 0.006255231857299805, 0.006246399879455566, 0.006313983917236328, 0.006225344181060791, 0.006244448184967041, 0.006275263786315918, 0.006176224231719971, 0.006279359817504883, 0.006270624160766601, 0.006258592128753662, 0.006256896018981934, 0.006257279872894287, 0.006246335983276367, 0.006243807792663574, 0.006271168231964112, 0.006259168148040771, 0.006256480216979981, 0.006260735988616943, 0.006270976066589356, 0.006250495910644531, 0.006233312129974365, 0.006301631927490234, 0.006218592166900635, 0.006266240119934082, 0.00625113582611084, 0.006225599765777588, 0.006259263992309571, 0.006233376026153564, 0.006246880054473877, 0.006316031932830811, 0.006276256084442139, 0.006233280181884766, 0.006305439949035645, 0.006236447811126709, 0.0062297282218933105, 0.006248672008514405, 0.006231840133666992, 0.00625875186920166, 0.0062687678337097165, 0.006275167942047119, 0.006266880035400391, 0.006264832019805908, 0.006252543926239014, 0.006266975879669189, 0.0062626562118530274, 0.0062259521484375, 0.006542816162109375, 0.006330111980438233, 0.00651958417892456, 0.006299647808074951, 0.006224095821380615, 0.006341792106628418, 0.0062798080444335935, 0.006247424125671387, 0.006288064002990723, 0.006248767852783203, 0.00626854419708252, 0.006244927883148193, 0.006358079910278321, 0.006265120029449463, 0.006246880054473877, 0.006242303848266601, 0.006231520175933838, 0.006254655838012695, 0.006242784023284912, 0.006279232025146485, 0.006221663951873779, 0.00624070405960083, 0.00625216007232666, 0.006130335807800293, 0.006248703956604004, 0.006416384220123291, 0.006322495937347412, 0.006252416133880615, 0.006225152015686035, 0.006239840030670166, 0.006224575996398926, 0.0062507839202880855, 0.006217728137969971, 0.006745888233184815, 0.006349023818969727, 0.006329535961151123, 0.006263519763946533, 0.006234208106994629, 0.0062791681289672855, 0.0063647680282592775, 0.006246816158294677, 0.006238399982452392, 0.006223552227020264, 0.006260863780975342, 0.006278783798217774, 0.006269567966461181, 0.0062362561225891115, 0.006280064105987549, 0.006247200012207031, 0.006243743896484375, 0.006263391971588135, 0.006245471954345703, 0.006249375820159912, 0.0062864961624145505, 0.006345759868621826, 0.006485568046569824, 0.0062650880813598635, 0.0062997121810913085, 0.0062709121704101565, 0.006297567844390869, 0.006250847816467285, 0.006289087772369385, 0.006276224136352539, 0.006245376110076905, 0.006358687877655029, 0.00623583984375, 0.006246943950653076, 0.0062624640464782715, 0.006252128124237061, 0.006352640151977539, 0.006327616214752197, 0.006256288051605225, 0.006252543926239014, 0.006299903869628906, 0.006231808185577393, 0.0066557440757751465, 0.006322303771972656, 0.006248640060424805, 0.006321216106414795, 0.006267583847045898, 0.006265024185180664, 0.006244351863861084, 0.0062887039184570314, 0.006308544158935547, 0.006233407974243164, 0.006361023902893066, 0.006199711799621582, 0.006338560104370118, 0.006198463916778564, 0.0062576642036437985, 0.0062575039863586425, 0.006229184150695801, 0.00625055980682373, 0.0063318080902099606, 0.006246719837188721, 0.006238207817077636, 0.006250016212463379, 0.006258495807647705, 0.00624502420425415, 0.006221824169158936, 0.0062912960052490235, 0.006246560096740723, 0.006223423957824707, 0.00627785587310791, 0.006225535869598389, 0.0062403521537780764, 0.006242303848266601, 0.006223104000091553, 0.00622873592376709, 0.006243360042572021, 0.006232704162597656, 0.006225344181060791, 0.0063283839225769046, 0.006265696048736572, 0.0062566399574279785, 0.006235616207122803, 0.006335008144378662, 0.006242144107818604, 0.006236288070678711, 0.006267168045043945, 0.006235712051391602, 0.006287136077880859, 0.006274911880493164, 0.0062911357879638675, 0.006257887840270996, 0.006253600120544434, 0.006275008201599121, 0.006243008136749268, 0.006293504238128662, 0.0062873601913452145, 0.006377632141113281, 0.006265791893005371, 0.006282144069671631, 0.006260479927062988, 0.006238592147827148, 0.006284480094909668, 0.006279871940612793, 0.006266975879669189, 0.006275167942047119, 0.0062707839012146, 0.0064280638694763186, 0.0062715840339660645, 0.006373472213745117, 0.006281407833099365, 0.006274208068847657, 0.006323775768280029, 0.006359263896942139, 0.0063366079330444336, 0.006359744071960449, 0.006301055908203125, 0.006356095790863037, 0.006304448127746582, 0.006328991889953613, 0.006317823886871338, 0.00630844783782959, 0.006391551971435547, 0.006316287994384766, 0.006325888156890869, 0.006321728229522705, 0.006282080173492432, 0.00641542387008667, 0.006314911842346191, 0.0063448319435119626, 0.006287231922149659, 0.006339615821838379, 0.006298592090606689, 0.006273024082183838, 0.006355103969573975, 0.00626694393157959, 0.006270815849304199, 0.0062557759284973145, 0.006306367874145508, 0.006287327766418457, 0.006254591941833496, 0.006276832103729248, 0.00624294376373291, 0.006369184017181397, 0.006205440044403076, 0.006284639835357666, 0.006269696235656738, 0.006244256019592285, 0.006270175933837891, 0.006236735820770264, 0.00629315185546875, 0.006281792163848877, 0.00628710412979126, 0.006305056095123291, 0.0062659201622009275, 0.008454015731811523, 0.007951648235321044, 0.008065024375915527, 0.006589119911193848, 0.006395359992980957, 0.0063060479164123535, 0.006271327972412109, 0.006377503871917724, 0.006299647808074951, 0.006254591941833496, 0.0062945599555969236, 0.006308095932006836, 0.006292479991912842, 0.006317759990692139, 0.006281216144561768, 0.006293504238128662, 0.006262784004211426, 0.006288735866546631, 0.006302015781402588, 0.006268671989440918, 0.006264575958251953, 0.006251359939575195, 0.006540544033050537, 0.006279935836791992, 0.006230016231536865, 0.006299647808074951, 0.006289408206939697, 0.006318079948425293, 0.006252543926239014, 0.006307136058807373, 0.0063487358093261715, 0.006302464008331299, 0.006268479824066162, 0.006267327785491944, 0.006260447978973389, 0.006269216060638428, 0.006350751876831055, 0.006238304138183594, 0.006271071910858154, 0.006249792098999023, 0.006255199909210205, 0.006254496097564698, 0.006228223800659179, 0.006246560096740723, 0.00640172815322876, 0.006290880203247071, 0.006277696132659912, 0.006280511856079102, 0.006306496143341065, 0.006252352237701416, 0.006337056159973145, 0.0063056960105896, 0.006321983814239502, 0.006362559795379638, 0.006279679775238037, 0.006338560104370118, 0.006280511856079102, 0.006258912086486816, 0.006222303867340088, 0.006268928050994873, 0.006238207817077636, 0.006246399879455566, 0.006225056171417236, 0.0062258877754211425, 0.006245247840881348, 0.006251935958862305, 0.00628111982345581, 0.006244160175323486, 0.00625548791885376, 0.006274144172668457, 0.006263967990875244, 0.006248479843139649, 0.006255616188049316, 0.006298336029052734, 0.0062770237922668455, 0.0062763838768005375, 0.006269248008728028, 0.006275839805603027, 0.006303487777709961, 0.006368576049804687, 0.0062921600341796875, 0.006301152229309082, 0.0062960958480834965, 0.006281216144561768, 0.0062873601913452145, 0.006274432182312012, 0.006257279872894287, 0.006252416133880615, 0.0063155522346496585, 0.006242847919464111, 0.006289472103118896, 0.00630076789855957, 0.006275296211242676, 0.006277440071105957, 0.006345088005065918, 0.006296671867370605, 0.006236576080322266, 0.00625113582611084, 0.006260128021240235, 0.006271455764770508, 0.006264832019805908, 0.006281216144561768, 0.006276576042175293, 0.0062811517715454105, 0.006320064067840576, 0.006250815868377686, 0.006255008220672607, 0.006264768123626709, 0.006460608005523682, 0.006278240203857422, 0.006270495891571045, 0.006329792022705078, 0.006284095764160156, 0.0062626562118530274, 0.006258431911468506, 0.006240575790405273, 0.006241312026977539, 0.0062509760856628415, 0.006294015884399414, 0.006264927864074707, 0.006258656024932861, 0.006291391849517822, 0.006289087772369385, 0.006288000106811523, 0.006418047904968261, 0.006561855792999268, 0.006266047954559326, 0.006316864013671875, 0.006317696094512939, 0.0063564801216125484, 0.006299967765808105, 0.006314559936523438, 0.006270624160766601, 0.006259039878845215, 0.006367231845855713, 0.006303743839263916, 0.006240287780761719, 0.006257855892181396, 0.006267360210418701, 0.00625273609161377, 0.006236447811126709, 0.006238048076629639, 0.0062156801223754886, 0.0062475199699401854, 0.006228608131408691, 0.006213503837585449, 0.006215744018554688, 0.006242144107818604, 0.0062510080337524416, 0.006208896160125732, 0.006189824104309082, 0.006248447895050049, 0.006220863819122314, 0.006255551815032959, 0.0062259521484375, 0.006207647800445556, 0.006321792125701904, 0.00624454402923584, 0.006227968215942382, 0.006244383811950684, 0.006303711891174317, 0.006207712173461914, 0.006317952156066895, 0.006251488208770752, 0.0062486081123352055, 0.00623199987411499, 0.006267744064331055, 0.006269152164459228, 0.0062902398109436035, 0.006291615962982178, 0.006279232025146485, 0.006379776000976563, 0.006286848068237305, 0.006292448043823242, 0.006273087978363037, 0.006295519828796387, 0.006340608119964599, 0.006309023857116699, 0.006693439960479736, 0.0064085121154785155, 0.006391776084899902, 0.0064264960289001465, 0.006381120204925537, 0.006402624130249023, 0.006449183940887451, 0.00639792013168335, 0.00643891191482544, 0.006432767868041992, 0.006354144096374512, 0.0063695359230041505, 0.0062911038398742675, 0.006320223808288574, 0.0063762240409851074, 0.00649945592880249, 0.006423423767089844, 0.006432672023773193, 0.006397151947021484, 0.006376319885253906, 0.006450943946838379, 0.0064330239295959475, 0.00665177583694458, 0.006407392024993896, 0.006359039783477783, 0.006445919990539551, 0.006416160106658936, 0.0063753600120544434, 0.006433343887329102, 0.0063810238838195804, 0.006379839897155762, 0.006497536182403564, 0.006398719787597656, 0.006404096126556396, 0.006377471923828125, 0.006307487964630127, 0.006374015808105469, 0.006306591987609863, 0.006316991806030273, 0.0062724161148071285, 0.006341504096984863, 0.006419456005096436, 0.00633519983291626, 0.0063448319435119626, 0.006334335803985596, 0.0063506560325622555, 0.006322112083435059, 0.006353151798248291, 0.006311295986175537, 0.006304384231567383, 0.006322495937347412, 0.006391456127166748, 0.006346784114837647, 0.006322175979614258, 0.0063089919090271, 0.0062960958480834965, 0.006547808170318603, 0.006307839870452881, 0.006264512062072754, 0.006261055946350098, 0.006291456222534179, 0.006262784004211426, 0.006296832084655762, 0.006230144023895263, 0.0063023362159729006, 0.006274720191955566, 0.006260704040527343, 0.006273248195648193, 0.006238368034362793, 0.006301695823669433, 0.00628227186203003, 0.006292416095733643, 0.006277503967285156, 0.006240992069244384, 0.006273983955383301, 0.006253888130187988, 0.006259359836578369, 0.006230048179626465, 0.0062947521209716795, 0.006226336002349854, 0.0062527041435241695, 0.006242400169372559, 0.006208767890930176, 0.006452095985412598, 0.006266880035400391, 0.006481696128845215, 0.006406688213348389, 0.006383296012878418, 0.006326272010803223, 0.006285312175750732, 0.00630790376663208, 0.006295263767242432, 0.0063021121025085445, 0.006345823764801025, 0.00628604793548584, 0.006320032119750976, 0.006305888175964356, 0.006338560104370118, 0.006201727867126465, 0.006266848087310791, 0.006258336067199707, 0.006269504070281982, 0.006284639835357666, 0.00627561616897583, 0.006352799892425537, 0.006219808101654053, 0.006291359901428223, 0.006285215854644775, 0.006238304138183594, 0.006237504005432129, 0.006244351863861084, 0.006366112232208252, 0.006246367931365967, 0.006301504135131836, 0.006260735988616943, 0.006248447895050049, 0.006242303848266601, 0.006228032112121582, 0.006262239933013916, 0.006261216163635254, 0.006346848011016846, 0.0063610877990722655, 0.006315872192382812, 0.0063134398460388185, 0.006328927993774414, 0.006370719909667969, 0.006318848133087158, 0.00630460786819458, 0.00628223991394043, 0.006262207984924316, 0.006305823802947998, 0.006288127899169922, 0.006323999881744384, 0.006288671970367431, 0.006337247848510742, 0.006264832019805908, 0.0062494721412658695, 0.006247456073760986, 0.006245696067810059, 0.006503071784973145, 0.006307680130004883, 0.006264832019805908, 0.006268576145172119, 0.006253056049346924, 0.006260064125061035, 0.006251167774200439, 0.006290976047515869, 0.006300127983093261, 0.006275072097778321, 0.0062871999740600586, 0.006264256000518799, 0.006273439884185791, 0.0062709121704101565, 0.006283648014068603, 0.006273024082183838, 0.0064445118904113766, 0.006412831783294678, 0.006270271778106689, 0.006389696121215821, 0.006247168064117432, 0.006272192001342773, 0.006156383991241455, 0.006267104148864746, 0.006404128074645996, 0.00634444808959961, 0.0063201279640197755, 0.006359039783477783, 0.006348639965057373, 0.006332575798034668, 0.0063079681396484375, 0.00633132791519165, 0.006339200019836426, 0.006425151824951172, 0.006409984111785888, 0.0063396477699279785, 0.006316991806030273, 0.006336703777313233, 0.006516191959381103, 0.006395840167999268, 0.006369696140289306, 0.0063569917678833006, 0.00632422399520874, 0.006453248023986816, 0.006362720012664795, 0.006946879863739014, 0.0063487358093261715, 0.006372096061706543, 0.006344351768493652, 0.006385824203491211, 0.006303872108459473, 0.006370272159576416, 0.006394591808319092, 0.006349184036254883, 0.006352543830871582, 0.006352479934692383, 0.006359231948852539, 0.006341023921966553, 0.006371007919311523, 0.006418560028076172, 0.006389920234680176, 0.006337823867797851, 0.006339200019836426, 0.006421792030334473, 0.0063248958587646485, 0.00636736011505127, 0.006352672100067139, 0.006278783798217774, 0.00635753583908081, 0.00631388807296753, 0.006289535999298095, 0.00626691198348999, 0.006365056037902832, 0.006254591941833496, 0.00627126407623291, 0.006285024166107178, 0.006236192226409912, 0.006303711891174317, 0.0062733120918273926, 0.006229375839233399, 0.006288127899169922, 0.006276000022888184, 0.006283967971801758, 0.0062811517715454105, 0.006283328056335449, 0.006199488162994385, 0.006357728004455566, 0.0062873601913452145, 0.006280863761901855, 0.006335872173309326, 0.006331103801727295, 0.0063448319435119626, 0.006312320232391358, 0.006326015949249268, 0.006534527778625488, 0.006357759952545166, 0.006456480026245117, 0.006288127899169922, 0.00628323221206665, 0.00628329610824585, 0.006284543991088867, 0.006314720153808594, 0.006363135814666748, 0.006285056114196778, 0.006293759822845459, 0.006281216144561768, 0.0062804799079895016, 0.0062687678337097165, 0.006324831962585449, 0.00630134391784668, 0.00632806396484375, 0.0063678078651428225, 0.0063155522346496585, 0.0063414077758789066, 0.006293600082397461, 0.006311840057373047, 0.006293280124664307, 0.006534463882446289, 0.006325151920318604, 0.006338655948638916, 0.006338272094726562, 0.006335872173309326, 0.00626691198348999, 0.006290559768676758, 0.006364831924438477, 0.006379903793334961, 0.0063478717803955075, 0.006375679969787598, 0.006342944145202637, 0.0063079681396484375, 0.006341887950897217, 0.006321023941040039, 0.006342400074005127, 0.00628326416015625, 0.006283328056335449, 0.006297632217407227, 0.006313759803771973, 0.006305439949035645, 0.006285952091217041, 0.006347936153411865, 0.006408192157745361, 0.0063434882164001465, 0.006381440162658692, 0.006381472110748291, 0.006401631832122803, 0.006439424037933349, 0.006376448154449463, 0.006406816005706787, 0.006405439853668213, 0.00638431978225708, 0.006354527950286865, 0.00634716796875, 0.00638156795501709, 0.00632422399520874, 0.006409728050231934, 0.006494751930236816, 0.0065392317771911625, 0.0065270719528198245, 0.00644700813293457, 0.006414239883422852, 0.0064245758056640625, 0.006457568168640137, 0.006467455863952636, 0.006451263904571534, 0.006430655956268311, 0.006416704177856445, 0.006377151966094971, 0.006493951797485352, 0.006445312023162842, 0.0064364480972290035, 0.0066926078796386715, 0.006433440208435059, 0.006426784038543701, 0.006456960201263428, 0.006361120223999024, 0.006400191783905029, 0.006391295909881592, 0.0063861761093139645, 0.006475872039794922, 0.006407423973083496, 0.006449215888977051, 0.006502079963684082, 0.006380447864532471, 0.006391903877258301, 0.006363103866577149, 0.006337471961975098, 0.00633900785446167, 0.006267615795135498, 0.006304671764373779, 0.006321087837219238, 0.00631987190246582, 0.0065796799659729, 0.006327199935913086, 0.00630944013595581, 0.006258624076843262, 0.0062993597984313966, 0.00624502420425415, 0.0062791681289672855, 0.0062967357635498045, 0.006275936126708985, 0.006252543926239014, 0.006520671844482422, 0.0062837119102478025, 0.006249311923980713, 0.00628111982345581, 0.006298592090606689, 0.006244351863861084, 0.006291456222534179, 0.006252543926239014, 0.006277376174926757, 0.006235904216766357, 0.006196415901184082, 0.006286079883575439, 0.006248511791229248, 0.0062975997924804685, 0.006271008014678955, 0.006270944118499756, 0.006253632068634033, 0.006267519950866699, 0.0062262721061706544, 0.006284416198730469, 0.006220287799835205, 0.006232416152954102, 0.006240096092224121, 0.006668511867523193, 0.006512576103210449, 0.0073207998275756836, 0.006783103942871094, 0.007082367897033691, 0.006322751998901368, 0.006258399963378907, 0.006317088127136231, 0.00649721622467041, 0.0062724480628967285, 0.0062756800651550294, 0.00625267219543457, 0.006288544178009033, 0.006296288013458252, 0.0062873601913452145, 0.006266719818115234, 0.0062667841911315915, 0.00628111982345581, 0.006277472019195557, 0.006285312175750732, 0.006262784004211426, 0.006254591941833496, 0.006804736137390137, 0.006252543926239014, 0.006297440052032471, 0.006222752094268799, 0.006291584014892578, 0.006242176055908203, 0.006236159801483154, 0.006239456176757812, 0.006253471851348877, 0.006243936061859131, 0.006242591857910157, 0.006262911796569824, 0.006238080024719238, 0.006258272171020508, 0.00625705623626709, 0.006238207817077636, 0.006264832019805908, 0.006292992115020752, 0.006300159931182861, 0.006258240222930908, 0.006294015884399414, 0.006293439865112304, 0.006313983917236328, 0.006287295818328858, 0.006248095989227295, 0.006281023979187012, 0.006247231960296631, 0.006327648162841797, 0.006221663951873779, 0.006268928050994873, 0.006279200077056885, 0.006264800071716308, 0.006273024082183838, 0.0062770237922668455, 0.006230112075805664, 0.006246592044830322, 0.006284255981445313, 0.006269023895263672, 0.006227871894836426, 0.0062592320442199706, 0.006229599952697754, 0.006263519763946533, 0.006266880035400391, 0.006248640060424805, 0.006243775844573975, 0.006273056030273437, 0.006257279872894287, 0.0062665920257568355, 0.0062975997924804685, 0.00628329610824585, 0.006264095783233643, 0.006272960186004639, 0.006250912189483642, 0.006274496078491211, 0.006241087913513184, 0.006289343833923339, 0.006260896205902099, 0.006292704105377197, 0.006295360088348388, 0.006290207862854004, 0.006310080051422119, 0.006266079902648926, 0.00632092809677124, 0.00630790376663208, 0.006395423889160156, 0.006318272113800049, 0.006302175998687744, 0.0062605757713317875, 0.006264736175537109, 0.006280191898345947, 0.006251520156860352, 0.006247776031494141, 0.0062772479057312015, 0.006261312007904053, 0.006285280227661133, 0.006236159801483154, 0.006283455848693847, 0.0062782721519470215, 0.006895679950714112, 0.0062531838417053225, 0.006502655982971192, 0.006285151958465576, 0.006270463943481445, 0.006252960205078125, 0.006293600082397461, 0.006241663932800293, 0.0063023362159729006, 0.0062683200836181644, 0.0062848000526428225, 0.0062715840339660645, 0.0062674241065979, 0.006451456069946289, 0.006232160091400147, 0.006247712135314941, 0.006236415863037109, 0.006248640060424805, 0.006234367847442627, 0.006223680019378662, 0.006268864154815674, 0.006231552124023438, 0.006394464015960694, 0.006393439769744873, 0.006303296089172363, 0.0063348479270935055, 0.006238592147827148, 0.0062631359100341795, 0.0062442879676818846, 0.006273791790008545, 0.006232639789581299, 0.006269343852996826, 0.006243775844573975, 0.006294047832489014, 0.00625267219543457, 0.006239776134490967, 0.0062552962303161624, 0.006210432052612304, 0.0062485761642456054, 0.00621449613571167, 0.006285120010375977, 0.006254623889923096, 0.006243552207946777, 0.006236959934234619, 0.006246399879455566, 0.006273087978363037, 0.006323616027832032, 0.0062626562118530274, 0.006291584014892578, 0.00624294376373291, 0.006264736175537109, 0.006240032196044922, 0.006239808082580566, 0.006235136032104492, 0.0063004159927368165, 0.00628550386428833, 0.006257184028625488, 0.006703296184539795, 0.0062873601913452145, 0.0063240318298339844, 0.006254496097564698, 0.006289696216583252, 0.006277376174926757, 0.006264448165893555, 0.006277440071105957, 0.006242112159729004, 0.006285312175750732, 0.006232063770294189, 0.0063162240982055666, 0.006264800071716308, 0.006237343788146972, 0.006292479991912842, 0.006251967906951904, 0.006262976169586182, 0.0062564477920532225, 0.006283360004425049, 0.006361055850982666, 0.006250527858734131, 0.006270976066589356, 0.006272928237915039, 0.006226016044616699, 0.006244448184967041, 0.006238111972808838, 0.006270944118499756, 0.006274240016937256, 0.006273087978363037, 0.006253344058990478, 0.006234272003173828, 0.006255871772766113, 0.006248832225799561, 0.006252511978149414, 0.0062416958808898925, 0.006324160099029541, 0.006214560031890869, 0.006366720199584961, 0.006285823822021484, 0.006260704040527343, 0.006276800155639649, 0.006256991863250732, 0.006266880035400391, 0.0062293438911437985, 0.0064330239295959475, 0.006250815868377686, 0.006483104228973389, 0.006273600101470947, 0.006270463943481445, 0.006306943893432618, 0.006348800182342529, 0.006260640144348144, 0.006311935901641846, 0.006283167839050293, 0.0063089919090271, 0.006257472038269043, 0.006264832019805908, 0.006259744167327881, 0.006259456157684327, 0.006246592044830322, 0.006248032093048096, 0.006264383792877197, 0.00625708818435669, 0.006275519847869873, 0.006221824169158936, 0.006254591941833496, 0.006244351863861084, 0.006227680206298828, 0.006258975982666015, 0.006256735801696777, 0.006309887886047363, 0.006239456176757812, 0.00628601598739624, 0.0062425599098205565, 0.006232895851135254, 0.006237343788146972, 0.0062379841804504396, 0.006275392055511474, 0.006261760234832763, 0.006304768085479737, 0.00623583984375, 0.00625977611541748, 0.006173567771911621, 0.0062518720626831056, 0.006240672111511231, 0.0062485761642456054, 0.006281248092651367, 0.006244512081146241, 0.006244192123413086, 0.006242303848266601, 0.00633241605758667, 0.006272543907165527, 0.00624832010269165, 0.006243040084838867, 0.006241663932800293, 0.006275712013244629, 0.0062772479057312015, 0.006254432201385498, 0.006270847797393799, 0.0062639999389648435, 0.006292319774627685, 0.006475776195526123, 0.006301695823669433, 0.006267072200775147, 0.006258272171020508, 0.006263008117675781, 0.00621779203414917, 0.006234047889709472, 0.006221183776855469, 0.006277184009552002, 0.0062039680480957034, 0.006266464233398437, 0.006418848037719726, 0.006275072097778321, 0.006299808025360108, 0.0062709121704101565, 0.006291615962982178, 0.006285056114196778, 0.006262784004211426, 0.006287487983703613, 0.006285183906555176, 0.006297632217407227, 0.006461408138275147, 0.006439008235931396, 0.0063036479949951174, 0.006251935958862305, 0.006254432201385498, 0.0062594242095947265, 0.00636521577835083, 0.006262784004211426, 0.006293504238128662, 0.006352896213531494, 0.006297088146209716, 0.006301983833312989, 0.006256864070892334, 0.006305791854858398, 0.006248288154602051, 0.006270592212677002, 0.006257343769073487, 0.006254176139831543, 0.00627455997467041, 0.006310656070709229, 0.006266880035400391, 0.006272895812988281, 0.0062706880569458005, 0.006168511867523194, 0.006259488105773926, 0.006267839908599853, 0.006246816158294677, 0.0062674241065979, 0.0062486081123352055, 0.0062623038291931155, 0.0062549118995666505, 0.006237887859344483, 0.006244671821594238, 0.006264832019805908, 0.006278240203857422, 0.0062271361351013185, 0.0062707839012146, 0.0062135357856750485, 0.006249567985534668, 0.006261663913726807, 0.006238111972808838, 0.0062278079986572265, 0.006250080108642578, 0.00625929594039917, 0.006279200077056885, 0.006290976047515869, 0.0062708802223205564, 0.006257184028625488, 0.00626694393157959, 0.006254496097564698, 0.006248544216156006, 0.0062644162178039555, 0.00627785587310791, 0.006296607971191406, 0.006261248111724854, 0.0062707200050354005, 0.006236703872680664, 0.006281184196472168, 0.006264736175537109, 0.006282400131225586, 0.00627513599395752, 0.0063001918792724605, 0.006287551879882812, 0.0062644162178039555, 0.00627891206741333, 0.006243040084838867, 0.00627839994430542, 0.006238815784454346, 0.0063161921501159665, 0.006266111850738525, 0.006252768039703369, 0.006273280143737793, 0.006256927967071533, 0.006317440032958984, 0.006226560115814209, 0.006291744232177735, 0.006258399963378907, 0.006244351863861084, 0.006237215995788574, 0.00623305606842041, 0.0062566399574279785, 0.006223872184753418, 0.0062782721519470215, 0.006230656147003174, 0.006261023998260498, 0.006296800136566162, 0.006216224193572998, 0.00625171184539795, 0.006247104167938233, 0.006291584014892578, 0.006242303848266601, 0.006240479946136475, 0.006253632068634033, 0.006261792182922363, 0.006266687870025635, 0.006242527961730957, 0.006435743808746338, 0.0062369279861450196, 0.006309728145599366, 0.006254303932189942, 0.006236288070678711, 0.006258399963378907, 0.006265503883361816, 0.006269951820373535, 0.006219871997833252, 0.006244671821594238, 0.006274752140045166, 0.006218048095703125, 0.006244895935058593, 0.006241407871246338, 0.006232831954956055, 0.006230144023895263, 0.006254720211029053, 0.0062280001640319825, 0.006255807876586914, 0.0062614078521728515, 0.0062791681289672855, 0.006295008182525635, 0.006264768123626709, 0.006323935985565185, 0.0062592320442199706, 0.006313695907592774, 0.006310527801513672, 0.006293248176574707, 0.006299903869628906, 0.006294623851776123, 0.006308544158935547, 0.006260223865509033, 0.006295296192169189, 0.0062863039970397945, 0.006252128124237061, 0.00642300796508789, 0.006753888130187988, 0.006943071842193603, 0.006821887969970703, 0.006658239841461182, 0.0064141440391540525, 0.006651904106140137, 0.006684671878814697, 0.006401472091674805, 0.006343232154846191, 0.00632422399520874, 0.006473567962646485, 0.006299392223358154, 0.006473696231842041, 0.006341055870056152, 0.0062791681289672855, 0.006301695823669433, 0.006491583824157715, 0.006342751979827881, 0.0064204797744750975, 0.006262432098388672, 0.006377024173736572, 0.006424479961395264, 0.006283199787139893, 0.006775743961334228, 0.00670246410369873, 0.006959519863128662, 0.0068089919090271, 0.007252799987792969, 0.006895616054534912, 0.006481919765472412, 0.006317887783050537, 0.006246592044830322, 0.006331391811370849, 0.006251296043395996, 0.006375775814056396, 0.0062317438125610355, 0.006334047794342041, 0.006496863842010498, 0.006286431789398193, 0.0063108158111572265, 0.006391808032989502, 0.006318079948425293, 0.006287295818328858, 0.006444799900054932, 0.006317567825317383, 0.0062533760070800784, 0.006293280124664307, 0.006252960205078125, 0.006401855945587159, 0.006316031932830811, 0.006393856048583985, 0.006300864219665528, 0.006323008060455322, 0.0062975997924804685, 0.006252031803131103, 0.006243008136749268, 0.0062252159118652345, 0.006219615936279297, 0.006240287780761719, 0.006293791770935059, 0.006281568050384522, 0.006218976020812989, 0.006217599868774414, 0.006232287883758545, 0.006275072097778321, 0.006242208003997803, 0.0062791681289672855, 0.006472479820251465, 0.006246304035186768, 0.006688064098358155, 0.007977759838104248, 0.0068579201698303225, 0.006299935817718506, 0.006314527988433838, 0.006479743957519531, 0.006246528148651123, 0.006273024082183838, 0.006256095886230469, 0.006265376091003418, 0.006238207817077636, 0.006192704200744629, 0.006247168064117432, 0.006213632106781006, 0.006246399879455566, 0.006217728137969971, 0.006225279808044434, 0.00625113582611084, 0.006287040233612061, 0.006231904029846191, 0.006289055824279785, 0.006222655773162842, 0.006229568004608155, 0.0062408318519592285, 0.006275167942047119, 0.006278528213500977, 0.006250912189483642, 0.006295551776885986, 0.006320032119750976, 0.006289120197296143, 0.006264319896697998, 0.006271232128143311, 0.0062689599990844724, 0.006253151893615722, 0.0065699520111083985, 0.00628329610824585, 0.006309887886047363, 0.006336095809936523, 0.006281407833099365, 0.006318304061889648, 0.006274687767028809, 0.006252927780151367, 0.006271232128143311, 0.006301375865936279, 0.0062997121810913085, 0.0062854719161987305, 0.006301536083221436, 0.0062791681289672855, 0.006267039775848388, 0.0062605757713317875, 0.006238207817077636, 0.006261856079101563, 0.006257887840270996, 0.006311552047729492, 0.006278592109680176, 0.00625932788848877, 0.006268928050994873, 0.006244703769683838, 0.006276768207550049, 0.006246399879455566, 0.006301504135131836, 0.00623635196685791, 0.006328320026397705, 0.006318016052246094, 0.006276768207550049, 0.006310304164886474, 0.006305791854858398, 0.0063055682182312014, 0.00625654411315918, 0.006313663959503174, 0.006288000106811523, 0.006291456222534179, 0.006323775768280029, 0.006269375801086426, 0.006173855781555175, 0.006361440181732177, 0.0062551040649414065, 0.006252543926239014, 0.006240255832672119, 0.006291615962982178, 0.006243807792663574, 0.006300064086914062, 0.006241919994354248, 0.006277184009552002, 0.006230303764343262, 0.006213280200958252, 0.006240511894226074, 0.006237887859344483, 0.0062708802223205564, 0.006257152080535889, 0.006257984161376953, 0.006230720043182373, 0.006249983787536621, 0.006265120029449463, 0.006246335983276367, 0.006269216060638428, 0.006256383895874023, 0.006301951885223389, 0.006223296165466309, 0.006487711906433106, 0.00624351978302002, 0.006235328197479248, 0.006238751888275147, 0.006245471954345703, 0.0062837119102478025, 0.006265312194824219, 0.006277120113372803, 0.006254591941833496, 0.006259903907775879, 0.00626361608505249, 0.006453248023986816, 0.006285312175750732, 0.006227968215942382, 0.006282400131225586, 0.006325088024139404, 0.006240255832672119, 0.006281472206115722, 0.006238175868988037, 0.006252448081970215, 0.0061950721740722655, 0.006275072097778321, 0.006256608009338379, 0.006293536186218262, 0.006274816036224365, 0.006251071929931641, 0.006248127937316894, 0.006250048160552978, 0.006434271812438965, 0.006230591773986816, 0.006214144229888916, 0.006283167839050293, 0.006219168186187744, 0.006740831851959229, 0.006266623973846436, 0.006231904029846191, 0.006234272003173828, 0.006238207817077636, 0.006150144100189209, 0.0062585601806640625, 0.006305920124053955, 0.006299647808074951, 0.006256192207336426, 0.006259136199951172, 0.0062566399574279785, 0.006288415908813476, 0.006255424022674561, 0.0063062400817871095, 0.006241824150085449, 0.0062791681289672855, 0.006275263786315918, 0.006250495910644531, 0.006266816139221192, 0.0062321281433105466, 0.006232063770294189, 0.006231328010559082, 0.006256703853607178, 0.006214303970336914, 0.006273024082183838, 0.006266880035400391, 0.006236480236053467, 0.006244031906127929, 0.006229279994964599, 0.006238111972808838, 0.006222655773162842, 0.0062583680152893065, 0.006248767852783203, 0.006248640060424805, 0.0062724161148071285, 0.006238143920898437, 0.0062646718025207515, 0.006250495910644531, 0.006266655921936035, 0.006242335796356201, 0.006273471832275391, 0.006252287864685058, 0.006232992172241211, 0.006260735988616943, 0.006231872081756592, 0.006239488124847412, 0.006232160091400147, 0.006246304035186768, 0.006224544048309326, 0.0062259202003479, 0.006242400169372559, 0.006246304035186768, 0.006242303848266601, 0.006213183879852295, 0.006242144107818604, 0.006223936080932618, 0.006259039878845215, 0.006245855808258057, 0.0062326078414916995, 0.006244095802307129, 0.006222271919250488, 0.00628326416015625, 0.006227968215942382, 0.006256415843963623, 0.006240479946136475, 0.006266240119934082, 0.006233823776245117]",tokens/s,158.58351209614608,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,2188.3904,2549.02272,0.0,2153.775104,2041.744384,s,1,11.085216796875,11.085216796875,0.0,11.085216796875,11.085216796875,11.085216796875,11.085216796875,[11.085216796875],,kWh,8.28118539208276e-05,9.12761928141536e-06,3.244780373600625e-05,0.00012438727693824922,,MB,2259.472384,2563.702784,0.0,2153.775104,1917.689856,s,10,0.986597511291504,0.0986597511291504,0.00023554764159677342,0.09864913940429687,0.09887231750488282,0.09902524719238281,0.09914759094238282,"[0.09917817687988281, 0.09829135894775391, 0.0986562271118164, 0.09849993896484376, 0.0986915512084961, 0.09854611206054688, 0.09864205169677734, 0.09843129730224609, 0.09882246398925781, 0.09883833312988281]",tokens/s,2594.7764622361915,kWh,2.9290764352499914e-06,3.2302500115680017e-07,1.9342209918200847e-06,5.186322428226876e-06,tokens/kWh,49360602.5353735,MB,2267.312128,2563.702784,0.0,2153.775104,2001.551872,s,10,14.554385620117186,1.4554385620117187,0.008469782365896452,1.4553158569335936,1.4653969116210939,1.466353546142578,1.4671188537597657,"[1.462813720703125, 1.449466796875, 1.4407430419921874, 1.4507158203125, 1.4492359619140625, 1.465184326171875, 1.44768603515625, 1.4599158935546874, 1.4613138427734376, 1.4673101806640625]",tokens/s,43.28592195119587,kWh,4.23991138060001e-05,4.6763035155543845e-06,2.2585743068579413e-05,6.966116039013388e-05,tokens/kWh,904377.699813951,,s,630,14.551532554626458,0.02309767072162931,0.0003988010129198367,0.02298471927642822,0.023503081703186036,0.02363767557144165,0.024713653583526615,"[0.024669599533081055, 0.023513696670532228, 0.02348236846923828, 0.022975807189941407, 0.022803136825561524, 0.022976383209228516, 0.0230380802154541, 0.02292531204223633, 0.022818944931030274, 0.02282304000854492, 0.02290457534790039, 0.022890752792358398, 0.022785791397094725, 0.023414783477783203, 0.02318934440612793, 0.023064735412597657, 0.023159839630126952, 0.022975616455078125, 0.0232077751159668, 0.022969343185424804, 0.02297894477844238, 0.02293414306640625, 0.02296623992919922, 0.022913055419921877, 0.0232857608795166, 0.02297881507873535, 0.02299673652648926, 0.02306662368774414, 0.022998079299926758, 0.02303420829772949, 0.02303411293029785, 0.02328428840637207, 0.02305820846557617, 0.022988288879394532, 0.023513151168823243, 0.023439680099487305, 0.023545984268188477, 0.023390207290649414, 0.023504896163940428, 0.023502687454223632, 0.023478431701660155, 0.023475296020507814, 0.023463872909545897, 0.023399295806884764, 0.023474239349365236, 0.0235948486328125, 0.023548095703125, 0.023554271697998046, 0.023514911651611327, 0.02332057571411133, 0.02326323127746582, 0.02335536003112793, 0.023502880096435547, 0.02353561592102051, 0.02337123107910156, 0.023494495391845702, 0.023341760635375977, 0.023103456497192382, 0.023002912521362304, 0.023099647521972657, 0.02292736053466797, 0.02294169616699219, 0.022790143966674805, 0.02372403144836426, 0.023453983306884765, 0.023357152938842773, 0.0237589111328125, 0.02384889602661133, 0.02308095932006836, 0.023154687881469727, 0.022962175369262695, 0.022929216384887697, 0.022914783477783203, 0.02327731132507324, 0.022911584854125977, 0.02278812789916992, 0.02294758415222168, 0.022782304763793945, 0.02291529655456543, 0.022816543579101563, 0.022769664764404295, 0.022816768646240236, 0.022722143173217774, 0.022923871994018553, 0.022949760437011718, 0.023365568161010743, 0.02534809684753418, 0.022970367431640625, 0.023015424728393553, 0.02304819107055664, 0.0229171199798584, 0.02293270492553711, 0.023015520095825196, 0.022798559188842774, 0.022775423049926757, 0.023298912048339844, 0.02290281677246094, 0.022878175735473634, 0.022964223861694336, 0.02287753677368164, 0.02288912010192871, 0.022982656478881838, 0.02272870445251465, 0.022734848022460938, 0.022744352340698243, 0.02295078468322754, 0.02306550407409668, 0.02287868881225586, 0.02279583930969238, 0.02279884719848633, 0.0228110408782959, 0.022835071563720704, 0.022789472579956054, 0.02302345657348633, 0.022780864715576173, 0.022757375717163086, 0.023193248748779295, 0.022865503311157227, 0.022815168380737303, 0.022856000900268555, 0.0228853759765625, 0.022874975204467775, 0.0229005126953125, 0.022772096633911134, 0.022787071228027343, 0.02277897644042969, 0.02299283218383789, 0.022958751678466796, 0.022894880294799805, 0.022977888107299806, 0.022786176681518555, 0.02276937675476074, 0.022944416046142578, 0.022922752380371093, 0.02303657531738281, 0.02286422348022461, 0.022978208541870118, 0.022969823837280273, 0.022780448913574218, 0.022785120010375977, 0.022852512359619142, 0.02278972816467285, 0.02276188850402832, 0.022773727416992188, 0.022810239791870118, 0.02287161636352539, 0.022965087890625, 0.02287228775024414, 0.02281999969482422, 0.022839935302734374, 0.022759424209594727, 0.02284297561645508, 0.02277212715148926, 0.02274508857727051, 0.02273219108581543, 0.0227325439453125, 0.02297260856628418, 0.02287273597717285, 0.022812095642089844, 0.022935199737548827, 0.022836544036865233, 0.022953311920166017, 0.022786304473876952, 0.02279814338684082, 0.02275551986694336, 0.022792192459106447, 0.022768640518188478, 0.022810976028442384, 0.022802175521850584, 0.022767711639404296, 0.02283958435058594, 0.022782495498657226, 0.022820863723754883, 0.022878400802612303, 0.022896160125732423, 0.022871679306030273, 0.02288844871520996, 0.02289321517944336, 0.023443456649780273, 0.022962175369262695, 0.02293065643310547, 0.022952735900878905, 0.022816768646240236, 0.022869247436523438, 0.02275404739379883, 0.02322364807128906, 0.0228372802734375, 0.022768255233764648, 0.022773759841918945, 0.02352998352050781, 0.023880096435546876, 0.02355561637878418, 0.02357004737854004, 0.023380767822265624, 0.0233123836517334, 0.02387942314147949, 0.02342428779602051, 0.023253984451293945, 0.02332431983947754, 0.023031808853149413, 0.02291747283935547, 0.02322777557373047, 0.023298751831054686, 0.02313814353942871, 0.022982751846313477, 0.022915424346923827, 0.022918815612792968, 0.02281497573852539, 0.02270182418823242, 0.022812671661376953, 0.022779903411865234, 0.022887424468994142, 0.022940671920776368, 0.022999040603637694, 0.022969631195068358, 0.02336147117614746, 0.023334911346435547, 0.023292703628540037, 0.02306252861022949, 0.022845439910888672, 0.02305059242248535, 0.022904287338256835, 0.022884544372558595, 0.02278940773010254, 0.022865631103515624, 0.022811647415161132, 0.022708223342895507, 0.022810943603515627, 0.022867168426513672, 0.022888927459716796, 0.02308032035827637, 0.023040639877319337, 0.022874111175537108, 0.02281785583496094, 0.022815679550170897, 0.02278972816467285, 0.022862239837646483, 0.022733951568603517, 0.022834047317504883, 0.022849536895751952, 0.022837247848510742, 0.02284752082824707, 0.022984672546386718, 0.022759424209594727, 0.022818080902099608, 0.022905567169189452, 0.02286591911315918, 0.02282598304748535, 0.022897663116455077, 0.023080671310424804, 0.02312835121154785, 0.02287788772583008, 0.02314854431152344, 0.023004959106445313, 0.022857376098632812, 0.023005760192871094, 0.023277536392211914, 0.022906911849975585, 0.0227061767578125, 0.02292915153503418, 0.02283750343322754, 0.02286083221435547, 0.023014368057250975, 0.022990591049194337, 0.02291708755493164, 0.022903072357177735, 0.022914560317993164, 0.023295679092407227, 0.02326406478881836, 0.0230665283203125, 0.02306649589538574, 0.02305023956298828, 0.023007007598876954, 0.02304863929748535, 0.022884191513061522, 0.02301763153076172, 0.022921215057373046, 0.023069791793823242, 0.023159135818481447, 0.022938175201416014, 0.02290892791748047, 0.023131776809692382, 0.022929792404174806, 0.023043359756469727, 0.023114463806152345, 0.022915071487426757, 0.02292736053466797, 0.022921440124511718, 0.02293494415283203, 0.023017183303833007, 0.02284137535095215, 0.022905471801757813, 0.02290892791748047, 0.022804479598999023, 0.02291878318786621, 0.02302137565612793, 0.022964799880981445, 0.02290505599975586, 0.022961952209472655, 0.0231014404296875, 0.022910463333129884, 0.022858591079711915, 0.02292291259765625, 0.02307276725769043, 0.022898687362670898, 0.023068351745605467, 0.023470399856567382, 0.02331648063659668, 0.02308915138244629, 0.02302128028869629, 0.023048416137695312, 0.02318751907348633, 0.02309119987487793, 0.02288163185119629, 0.02288912010192871, 0.023566463470458984, 0.02326515197753906, 0.023549312591552733, 0.023114368438720702, 0.02296348762512207, 0.023113536834716796, 0.023128992080688478, 0.023033695220947267, 0.023070655822753906, 0.02309552001953125, 0.023011327743530274, 0.022916255950927736, 0.022914047241210937, 0.023004383087158204, 0.022846080780029296, 0.022779424667358397, 0.022890911102294922, 0.02291859245300293, 0.02292799949645996, 0.02314841651916504, 0.023230144500732422, 0.023210399627685546, 0.023363744735717774, 0.02344086456298828, 0.02378531265258789, 0.023405248641967774, 0.023380895614624024, 0.02344393539428711, 0.023349472045898437, 0.02335804748535156, 0.02346566390991211, 0.023498687744140625, 0.023250112533569334, 0.023190528869628906, 0.023241727828979493, 0.023474592208862305, 0.023336767196655273, 0.02328780746459961, 0.023227039337158202, 0.022994720458984375, 0.023042463302612306, 0.023238048553466797, 0.023635616302490236, 0.022948736190795897, 0.022998912811279297, 0.023203071594238282, 0.023231103897094728, 0.023058015823364256, 0.023218528747558594, 0.023288000106811525, 0.023549951553344727, 0.0235248966217041, 0.023607072830200197, 0.02343017578125, 0.024481439590454103, 0.024731647491455077, 0.02346988868713379, 0.02312620735168457, 0.02305843162536621, 0.023068672180175782, 0.022967327117919923, 0.023030752182006835, 0.022812320709228517, 0.02376192092895508, 0.02334534454345703, 0.023184064865112305, 0.023214176177978517, 0.02307072067260742, 0.023144447326660156, 0.022983808517456055, 0.02294028854370117, 0.022853631973266602, 0.022805791854858398, 0.022854272842407226, 0.022911327362060547, 0.022767328262329103, 0.022839231491088866, 0.02621670341491699, 0.025442399978637696, 0.022906879425048828, 0.023074047088623047, 0.022796831130981444, 0.022777856826782225, 0.022739168167114257, 0.02293484878540039, 0.02277609634399414, 0.02276393508911133, 0.022872064590454103, 0.022749183654785156, 0.022799808502197264, 0.02276380729675293, 0.02283344078063965, 0.022838432312011717, 0.02283942413330078, 0.02279292869567871, 0.02290278434753418, 0.022703519821166994, 0.02275299263000488, 0.022751359939575194, 0.022795007705688475, 0.022876096725463868, 0.022813791275024413, 0.022706207275390626, 0.02280463981628418, 0.022809375762939454, 0.022836896896362306, 0.023425376892089844, 0.022845439910888672, 0.022771711349487304, 0.02294169616699219, 0.022887807846069336, 0.02275529670715332, 0.022817440032958984, 0.022921152114868164, 0.022843135833740234, 0.022774080276489257, 0.023154783248901366, 0.022923168182373048, 0.022792192459106447, 0.022796287536621093, 0.022736864089965822, 0.022835231781005858, 0.022725984573364257, 0.02270070457458496, 0.022673408508300782, 0.022761280059814454, 0.02326291275024414, 0.023208288192749022, 0.023080896377563477, 0.02303388786315918, 0.022779552459716797, 0.022820383071899413, 0.022776512145996092, 0.02281484794616699, 0.022831104278564454, 0.022686784744262695, 0.022709087371826173, 0.022718559265136717, 0.022773632049560545, 0.022818944931030274, 0.02291231918334961, 0.02277609634399414, 0.022824960708618162, 0.022753087997436524, 0.022765888214111327, 0.022730783462524416, 0.022924991607666017, 0.022903104782104493, 0.02310371208190918, 0.023088384628295898, 0.023154560089111327, 0.023215007781982423, 0.023334911346435547, 0.023345151901245118, 0.023377920150756838, 0.023322399139404298, 0.023544031143188475, 0.023457311630249025, 0.02327190399169922, 0.02350284767150879, 0.023817472457885742, 0.02419728088378906, 0.023159679412841798, 0.023129823684692383, 0.023078176498413087, 0.023046239852905274, 0.023066591262817383, 0.0229403190612793, 0.022931488037109374, 0.02296009635925293, 0.023162879943847657, 0.022962175369262695, 0.023004575729370116, 0.022899423599243164, 0.022859104156494142, 0.02287606430053711, 0.02282521629333496, 0.023187040328979492, 0.023360288619995118, 0.026224639892578124, 0.023268447875976563, 0.023165056228637695, 0.023032127380371095, 0.0234256649017334, 0.023271360397338868, 0.023330720901489257, 0.023373823165893554, 0.023520288467407228, 0.02487571144104004, 0.026834943771362304, 0.023702848434448243, 0.02381279945373535, 0.02347007942199707, 0.02328166389465332, 0.023318527221679687, 0.023516672134399414, 0.023349760055541992, 0.023207935333251953, 0.023171072006225587, 0.023262399673461914, 0.023140512466430663, 0.02303027153015137, 0.02302921676635742, 0.02313852882385254, 0.023144927978515625, 0.02330611228942871, 0.0235536003112793, 0.023169567108154297, 0.023345184326171876, 0.023639360427856446, 0.02345439910888672, 0.023392255783081056, 0.023195775985717773, 0.023203712463378906, 0.02361078453063965, 0.02352547264099121, 0.023079391479492187, 0.0229847354888916, 0.02305023956298828, 0.023013023376464842, 0.022946144104003908, 0.023134208679199218, 0.022920192718505858, 0.02295091247558594, 0.023072639465332032, 0.022878335952758788, 0.023029760360717775, 0.022958080291748048, 0.022959808349609374, 0.02288876724243164, 0.022956031799316406, 0.02286796760559082, 0.022808576583862306, 0.022902528762817384, 0.022980863571166993, 0.022966079711914063, 0.022866079330444336, 0.022929439544677733, 0.023347200393676756, 0.0229171199798584, 0.02284499168395996, 0.02284339141845703, 0.023011775970458986, 0.022984703063964843, 0.022788095474243163, 0.02281062316894531, 0.022838911056518554, 0.022898624420166016, 0.023003103256225586, 0.02334681510925293, 0.023245855331420897, 0.023164735794067384, 0.023825536727905272, 0.023835647583007814, 0.023535680770874024, 0.023566047668457032, 0.02328057670593262, 0.023022687911987305, 0.022999679565429688, 0.023077280044555663, 0.023358623504638673, 0.023290592193603514, 0.023424928665161132, 0.023364864349365234, 0.023138208389282225, 0.023036319732666014, 0.023387744903564454, 0.023235136032104493, 0.023058816909790038, 0.0230645751953125, 0.023033760070800782, 0.023034303665161134, 0.022984415054321288, 0.023264799118041992, 0.02334351921081543, 0.023467840194702147, 0.02334943962097168, 0.023750143051147463, 0.023142080307006836, 0.022999488830566406, 0.023001472473144532, 0.02302355194091797, 0.02317081642150879, 0.02292972755432129, 0.023203840255737306, 0.0230882568359375, 0.022962207794189452, 0.022915903091430663, 0.022923519134521484, 0.022900768280029297, 0.022916576385498048, 0.023097631454467773, 0.02306831932067871, 0.023032159805297853, 0.023111679077148437, 0.023072256088256835, 0.023107423782348632, 0.023089088439941407, 0.023089887619018555, 0.0229783992767334, 0.023135936737060547, 0.02328851127624512, 0.023602975845336913, 0.023662591934204103, 0.023756799697875978, 0.02369536018371582, 0.02365235137939453, 0.023603200912475586, 0.023635520935058593, 0.023615936279296874, 0.023591264724731446, 0.023873279571533203, 0.023758304595947265, 0.023789440155029297, 0.02373641586303711]",tokens/s,43.29440886277644,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Deci/DeciLM-7B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Deci/DeciLM-7B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1521.627136,1569.652736,0.0,1174.40512,1147.036672,s,1,9.174837890625,9.174837890625,0.0,9.174837890625,9.174837890625,9.174837890625,9.174837890625,[9.174837890625],,kWh,5.877842605833052e-05,6.457500539367533e-06,2.1065850186002127e-05,8.630177678370019e-05,,MB,1591.238656,1640.955904,0.0,1231.028224,1064.778752,s,10,0.5417076148986816,0.05417076148986817,0.0002991612835500622,0.054034416198730464,0.054629844665527345,0.05464852142333985,0.054663462829589844,"[0.05466719818115234, 0.05395267105102539, 0.054403553009033205, 0.05388374328613281, 0.05387664031982422, 0.05386707305908203, 0.0539851188659668, 0.054625694274902346, 0.05436220932006836, 0.05408371353149414]",tokens/s,4725.7965913564085,kWh,1.6224050773842514e-06,1.7891919610032442e-07,1.0721351169666835e-06,2.8734593904512597e-06,tokens/kWh,89091219.05488169,MB,1595.375616,1643.053056,0.0,1233.125376,1119.936512,s,10,11.516790893554688,1.1516790893554687,0.014175968572519957,1.157119384765625,1.1659436767578124,1.1674348510742187,1.168627790527344,"[1.156865966796875, 1.1615103759765626, 1.1206495361328126, 1.168926025390625, 1.13904345703125, 1.1395523681640625, 1.157372802734375, 1.147115234375, 1.160142822265625, 1.1656123046875]",tokens/s,54.70273844709435,kWh,3.264305194969919e-05,3.599656571623945e-06,1.5546405955633424e-05,5.1789114476956555e-05,tokens/kWh,1216471.8519763781,,s,630,11.513171941757191,0.0182748760980273,0.0003702732853264695,0.018321136474609374,0.018600220489501954,0.018759527015686034,0.019429015655517582,"[0.018496543884277343, 0.018222143173217773, 0.018337696075439454, 0.018274303436279296, 0.018227231979370116, 0.01834185600280762, 0.01832931137084961, 0.018333984375, 0.018300640106201173, 0.01833919906616211, 0.0183855037689209, 0.01843846321105957, 0.018487167358398438, 0.018321792602539064, 0.018392383575439455, 0.018314720153808594, 0.018430368423461914, 0.018307647705078124, 0.018386943817138672, 0.01954636764526367, 0.018539424896240234, 0.018700159072875977, 0.01852720069885254, 0.018660928726196287, 0.018301376342773436, 0.018692096710205077, 0.018381023406982423, 0.01868281555175781, 0.018606815338134765, 0.018315168380737306, 0.018128351211547853, 0.01819753646850586, 0.018078752517700195, 0.01788163185119629, 0.017928159713745118, 0.01817807960510254, 0.019560543060302735, 0.017892608642578123, 0.018021120071411132, 0.017933727264404297, 0.017898143768310545, 0.01781711959838867, 0.01778086471557617, 0.017873247146606444, 0.018137407302856446, 0.018095808029174806, 0.01839286422729492, 0.01839344024658203, 0.018333568572998046, 0.018400543212890624, 0.018452287673950196, 0.01827097511291504, 0.018306432723999025, 0.018232128143310548, 0.018562047958374024, 0.018413728713989257, 0.01830790328979492, 0.01802649688720703, 0.018167808532714845, 0.019142112731933593, 0.019446304321289062, 0.018234367370605468, 0.01837059211730957, 0.01831907272338867, 0.01845155143737793, 0.018242752075195313, 0.0184006404876709, 0.0183110408782959, 0.018409631729125978, 0.018238752365112305, 0.018168832778930662, 0.018165504455566406, 0.018143455505371095, 0.018368383407592774, 0.018282655715942384, 0.01828000068664551, 0.018264480590820312, 0.018645023345947264, 0.018462047576904297, 0.01852073669433594, 0.018412864685058594, 0.018557632446289062, 0.01840742492675781, 0.01846272087097168, 0.01846272087097168, 0.01858470344543457, 0.018535295486450196, 0.018395135879516602, 0.018266111373901366, 0.018489343643188477, 0.019142656326293944, 0.018550783157348632, 0.01850534439086914, 0.01858188819885254, 0.01838243293762207, 0.018948511123657228, 0.018390975952148437, 0.018322591781616212, 0.01827699279785156, 0.018274816513061523, 0.018249216079711913, 0.0182410888671875, 0.01827449607849121, 0.018307392120361327, 0.018492704391479493, 0.01855788803100586, 0.018755584716796874, 0.018665504455566407, 0.018591487884521484, 0.018516191482543944, 0.01858870315551758, 0.018467807769775392, 0.01842585563659668, 0.018520063400268554, 0.018509824752807616, 0.018337791442871093, 0.018472032546997072, 0.01823766326904297, 0.01841632080078125, 0.01852579116821289, 0.018423999786376953, 0.018179359436035155, 0.018650047302246092, 0.01828643226623535, 0.018331327438354493, 0.01842598342895508, 0.018340639114379883, 0.01820035171508789, 0.018094175338745116, 0.018360063552856444, 0.018367647171020508, 0.018084863662719726, 0.017983423233032227, 0.017958080291748047, 0.017991743087768554, 0.017875232696533204, 0.017993631362915038, 0.017659839630126954, 0.017674943923950196, 0.017479871749877928, 0.017536832809448243, 0.017511808395385742, 0.017820287704467773, 0.017499263763427735, 0.01753379249572754, 0.01755753517150879, 0.017525760650634766, 0.017783775329589843, 0.017527040481567384, 0.01758595275878906, 0.01762303924560547, 0.01767737579345703, 0.017633472442626953, 0.01770572853088379, 0.01773583984375, 0.017748096466064452, 0.017774303436279296, 0.017727487564086913, 0.01803209686279297, 0.017801759719848632, 0.017958976745605468, 0.017848608016967773, 0.017886432647705078, 0.0178090877532959, 0.017840351104736328, 0.017855007171630858, 0.018051071166992186, 0.017905183792114258, 0.017913503646850584, 0.017873727798461914, 0.017768447875976562, 0.017580032348632812, 0.01759548759460449, 0.01757702445983887, 0.01755945587158203, 0.01758201599121094, 0.017979040145874023, 0.017885536193847657, 0.017957887649536132, 0.017752128601074217, 0.017714111328125, 0.01759833526611328, 0.01760063934326172, 0.017564672470092774, 0.017586624145507813, 0.017543743133544922, 0.01784832000732422, 0.01761075210571289, 0.017762304306030274, 0.018194175720214843, 0.018288799285888672, 0.01822719955444336, 0.018292255401611328, 0.018169376373291017, 0.01840336036682129, 0.019065759658813478, 0.02022777557373047, 0.019386688232421876, 0.01820579147338867, 0.018299007415771486, 0.018180896759033203, 0.01822096061706543, 0.018063711166381835, 0.01829043197631836, 0.01984716796875, 0.019367231369018554, 0.018485248565673826, 0.01829782485961914, 0.018226015090942384, 0.018361215591430665, 0.018366464614868162, 0.018692096710205077, 0.018499584197998048, 0.01843404769897461, 0.01836851119995117, 0.018388992309570314, 0.01842585563659668, 0.01869808006286621, 0.01913667106628418, 0.01914291191101074, 0.018812671661376953, 0.01844223976135254, 0.01843600082397461, 0.018398847579956055, 0.01847542381286621, 0.018434112548828124, 0.018868223190307617, 0.01836851119995117, 0.018440191268920898, 0.018437183380126954, 0.01856812858581543, 0.01888876724243164, 0.01857734489440918, 0.01845622444152832, 0.018479455947875978, 0.018548063278198242, 0.018498207092285158, 0.018484800338745118, 0.01905504035949707, 0.01847091293334961, 0.01862575912475586, 0.018534496307373048, 0.018560800552368164, 0.018583456039428712, 0.01860009574890137, 0.018366592407226563, 0.018410207748413086, 0.01831430435180664, 0.01832032012939453, 0.01836444854736328, 0.018283679962158204, 0.018248512268066407, 0.01836412811279297, 0.01833366394042969, 0.018542112350463866, 0.01830771255493164, 0.018352703094482423, 0.01860748863220215, 0.0184400634765625, 0.018434560775756836, 0.018614143371582032, 0.018347808837890625, 0.01881532859802246, 0.018511871337890624, 0.018388992309570314, 0.018315263748168945, 0.01838489532470703, 0.018350080490112306, 0.01844633674621582, 0.018254880905151368, 0.018329856872558593, 0.018445024490356444, 0.018431232452392577, 0.01827619171142578, 0.01832048034667969, 0.01823315238952637, 0.018420896530151366, 0.018191200256347656, 0.018241695404052734, 0.018271839141845703, 0.01820025634765625, 0.018149471282958983, 0.018106847763061523, 0.01803468894958496, 0.018061664581298827, 0.017942176818847657, 0.017963008880615236, 0.017881088256835938, 0.01787494468688965, 0.01779097557067871, 0.01846643257141113, 0.01795110321044922, 0.01797887992858887, 0.0179368953704834, 0.017784927368164064, 0.017620416641235353, 0.01780169677734375, 0.017593696594238283, 0.017632095336914063, 0.01760598373413086, 0.017749568939208985, 0.017718175888061523, 0.017852224349975587, 0.017643007278442382, 0.01769696044921875, 0.01757222366333008, 0.017655935287475586, 0.01763942337036133, 0.017719295501708983, 0.017936384201049805, 0.017688575744628905, 0.01762918472290039, 0.017688575744628905, 0.017636800765991213, 0.017633855819702147, 0.01811075210571289, 0.018186336517333986, 0.018018463134765624, 0.017983808517456054, 0.017743871688842772, 0.017648960113525392, 0.017669855117797853, 0.01750239944458008, 0.01754185676574707, 0.017592031478881835, 0.01757798385620117, 0.01762665557861328, 0.017916736602783204, 0.017788511276245117, 0.017915712356567384, 0.017659584045410157, 0.017725343704223632, 0.01780633544921875, 0.017948448181152345, 0.018820671081542967, 0.01916729545593262, 0.0181824951171875, 0.018285919189453125, 0.01803766441345215, 0.017979391098022462, 0.017935712814331053, 0.01807744026184082, 0.017956832885742187, 0.018332128524780274, 0.018667999267578124, 0.018171295166015625, 0.018061279296875, 0.018168447494506836, 0.018218048095703127, 0.01843827247619629, 0.018371391296386718, 0.018145280838012694, 0.01815337562561035, 0.018138879776000975, 0.018137439727783203, 0.01820876884460449, 0.018237056732177733, 0.018016895294189452, 0.017993471145629884, 0.018100223541259765, 0.018019872665405272, 0.01795145606994629, 0.018171648025512695, 0.01839030456542969, 0.018147199630737305, 0.01806764793395996, 0.017955167770385742, 0.01793846321105957, 0.01792848014831543, 0.01789132881164551, 0.017957952499389647, 0.018260927200317384, 0.018457632064819336, 0.01839833641052246, 0.01852400016784668, 0.018380800247192384, 0.018348031997680665, 0.018398656845092773, 0.018251039505004882, 0.01826479911804199, 0.018163551330566408, 0.018497695922851564, 0.018251775741577148, 0.018294111251831054, 0.018309791564941405, 0.018241535186767577, 0.018274303436279296, 0.018292800903320312, 0.01829190444946289, 0.01842252731323242, 0.018343967437744142, 0.018341119766235352, 0.018442975997924806, 0.018359584808349608, 0.018364160537719727, 0.018234336853027344, 0.018904895782470704, 0.018556447982788087, 0.018555551528930663, 0.018493280410766602, 0.01854275131225586, 0.01843836784362793, 0.018530080795288086, 0.018397184371948243, 0.018565120697021483, 0.01863884735107422, 0.01815337562561035, 0.018669664382934572, 0.018147104263305663, 0.01802262306213379, 0.018108415603637695, 0.018050464630126953, 0.018236000061035155, 0.018042367935180666, 0.018299392700195313, 0.018175039291381836, 0.01839529609680176, 0.01826211166381836, 0.01841632080078125, 0.018418720245361328, 0.018336864471435548, 0.018269887924194338, 0.018486719131469726, 0.018410240173339844, 0.018297119140625, 0.01836345672607422, 0.018343839645385742, 0.018323583602905272, 0.018411840438842773, 0.018329919815063475, 0.018305023193359374, 0.01840278434753418, 0.018389535903930665, 0.018337791442871093, 0.01839308738708496, 0.01874502372741699, 0.01837116813659668, 0.01843312072753906, 0.018551679611206056, 0.018544384002685547, 0.01828175926208496, 0.018601343154907225, 0.018679712295532228, 0.01848940849304199, 0.01838287925720215, 0.01839366340637207, 0.018312799453735353, 0.01843596839904785, 0.01926633644104004, 0.01834828758239746, 0.018445152282714844, 0.018559904098510743, 0.018511871337890624, 0.018502880096435546, 0.018432416915893556, 0.018174400329589845, 0.018030336380004883, 0.018011455535888673, 0.01810051155090332, 0.017983808517456054, 0.01877020835876465, 0.017947967529296876, 0.018074304580688476, 0.018083839416503905, 0.018044927597045898, 0.017809343338012696, 0.017829824447631835, 0.017886911392211914, 0.017740320205688477, 0.017788671493530275, 0.01773379135131836, 0.01764352035522461, 0.01761484718322754, 0.018059263229370116, 0.017761407852172853, 0.017931135177612304, 0.01799273681640625, 0.017998815536499024, 0.017924095153808595, 0.01909350395202637, 0.018093088150024413, 0.01804115104675293, 0.01800668716430664, 0.0179814395904541, 0.01803868865966797, 0.018141151428222656, 0.018350208282470703, 0.01818828773498535, 0.018349088668823243, 0.018006879806518553, 0.018018367767333985, 0.017964544296264647, 0.018038911819458006, 0.017934783935546875, 0.018345983505249023, 0.019271263122558592, 0.02040227127075195, 0.01827027130126953, 0.018086143493652344, 0.01784809684753418, 0.017931711196899413, 0.017865631103515626, 0.01794758415222168, 0.0181214714050293, 0.018433792114257812, 0.018362367630004883, 0.018323455810546875, 0.018280448913574218, 0.018315391540527345, 0.018317184448242187, 0.018227264404296874, 0.018206783294677734, 0.0182205753326416, 0.01874185562133789, 0.0183191032409668, 0.018155519485473632, 0.01821286392211914, 0.018266111373901366, 0.018272096633911133, 0.018477439880371094, 0.018488351821899413, 0.018399999618530272, 0.018319360733032225, 0.01826153564453125, 0.018176576614379884, 0.018466720581054686, 0.018340160369873047, 0.018284351348876952, 0.01839414405822754, 0.018350208282470703, 0.018494207382202147, 0.01834185600280762, 0.018422847747802736, 0.01844758415222168, 0.018483936309814455, 0.018498559951782227, 0.018534400939941405, 0.018433088302612304, 0.018717567443847657, 0.018488927841186522, 0.018506336212158202, 0.01843596839904785, 0.01853660774230957, 0.018480480194091795, 0.018864639282226564, 0.018728384017944334, 0.018506303787231445, 0.018597888946533202, 0.01847500801086426, 0.018354047775268556, 0.01840959930419922, 0.018343936920166014, 0.018406911849975584, 0.01826255989074707, 0.018429920196533202, 0.01836025619506836, 0.0187761287689209, 0.018327552795410155, 0.018390399932861328, 0.01831999969482422, 0.019230335235595704, 0.01838323211669922, 0.01835215950012207, 0.018264192581176758, 0.018208223342895506, 0.018151039123535155, 0.018210624694824217, 0.01848179244995117, 0.018748735427856444, 0.018473344802856444, 0.018477376937866212, 0.018558879852294922, 0.018548831939697266, 0.018446271896362304, 0.018501632690429686, 0.01839017677307129, 0.018666400909423828, 0.01861631965637207, 0.018472959518432617, 0.018502784729003907, 0.018977535247802733, 0.019945247650146485, 0.018634592056274414, 0.018618560791015624, 0.018594432830810546, 0.01856070327758789, 0.018503040313720704, 0.01868579292297363, 0.01852672004699707, 0.018519872665405272, 0.018384767532348634, 0.01849318313598633, 0.01833996772766113, 0.018376735687255858, 0.018328224182128906, 0.018434080123901367, 0.018482208251953125, 0.01847747230529785, 0.018309247970581054, 0.01846931266784668, 0.018431999206542968, 0.018394975662231444, 0.018317567825317384, 0.018255231857299804, 0.018248191833496095, 0.018351648330688478, 0.018377216339111328, 0.018762752532958983, 0.01860416030883789, 0.018494592666625977, 0.018346944808959962, 0.01830940818786621, 0.018315391540527345, 0.018483776092529297, 0.018567007064819337, 0.0184967041015625, 0.018387775421142578, 0.018423807144165038, 0.018386943817138672, 0.01843152046203613, 0.01866908836364746, 0.01838073539733887, 0.018529279708862305, 0.01833193588256836, 0.018412544250488282, 0.01840787124633789, 0.018359615325927735, 0.01839571189880371, 0.018364160537719727, 0.01842857551574707]",tokens/s,54.719933237082024,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-4B,Qwen/Qwen1.5-4B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3145.99424,4639.883264,0.0,4244.635648,4125.520384,s,1,11.7520712890625,11.7520712890625,0.0,11.7520712890625,11.7520712890625,11.7520712890625,11.7520712890625,[11.7520712890625],,kWh,0.00014004919984996983,1.5440516300272363e-05,5.297976460599818e-05,0.00020846948075624037,,MB,3185.455104,4813.94688,0.0,4404.0192,4310.797312,s,10,1.1998705825805664,0.11998705825805664,0.00018602052537982502,0.11992507171630859,0.12016988677978516,0.1202754379272461,0.12035987884521485,"[0.12038098907470703, 0.12014105224609375, 0.11976278686523438, 0.12014643096923829, 0.11981970977783203, 0.11997030639648437, 0.11983433532714843, 0.12008499145507813, 0.11987983703613281, 0.11985014343261718]",tokens/s,2133.5634335614745,kWh,3.577076461839147e-06,3.944667214168399e-07,2.3632674732686106e-06,6.3348106565245984e-06,tokens/kWh,40411626.15276123,MB,3193.921536,4816.044032,0.0,4406.116352,4310.799872,s,10,24.742931640625002,2.4742931640625003,0.013718663129249119,2.4744366455078124,2.4924291015625,2.495043408203125,2.4971348535156253,"[2.49765771484375, 2.463353515625, 2.475403564453125, 2.451627685546875, 2.459360595703125, 2.4734697265625, 2.4754619140625, 2.49184814453125, 2.48601416015625, 2.468734619140625]",tokens/s,25.46181710196433,kWh,7.21368456039991e-05,7.956612219909271e-06,3.629331900752945e-05,0.00011638677683143779,tokens/kWh,541298.605521506,,s,630,24.74024644851682,0.039270232457963246,0.0006882776196660147,0.03920785522460937,0.039810585021972655,0.04008425540924072,0.041340187988281256,"[0.04134908676147461, 0.03932627105712891, 0.03967132949829102, 0.03985990524291992, 0.03962326431274414, 0.04112403106689453, 0.039345344543457034, 0.039279296875, 0.039335742950439456, 0.03915945434570312, 0.039241886138916014, 0.03938787078857422, 0.0393639030456543, 0.039313888549804686, 0.039327392578125, 0.03936399841308594, 0.03978335952758789, 0.039684097290039064, 0.03917824172973633, 0.03903219223022461, 0.03914355087280273, 0.03904790496826172, 0.03911420822143555, 0.03987420654296875, 0.039467681884765626, 0.03952025604248047, 0.03935577774047851, 0.03941849517822266, 0.03946086502075195, 0.039239200592041015, 0.03966204833984375, 0.03923484802246094, 0.03908809661865234, 0.03959414291381836, 0.047214942932128905, 0.03951795196533203, 0.039358272552490234, 0.03986092758178711, 0.03931260681152344, 0.03900495910644531, 0.039089344024658204, 0.03871334457397461, 0.03917907333374023, 0.04056268692016601, 0.039919200897216796, 0.041062976837158205, 0.039481407165527345, 0.03950774383544922, 0.04131840133666992, 0.039462913513183595, 0.04014688110351562, 0.03930527877807617, 0.03909632110595703, 0.03945676803588867, 0.03942195129394531, 0.03930492782592773, 0.0393935661315918, 0.03907920074462891, 0.03916419219970703, 0.03928067016601562, 0.039219615936279296, 0.03927859115600586, 0.039699615478515624, 0.03983657455444336, 0.03904307174682617, 0.03890995025634766, 0.03902668762207031, 0.03947865676879883, 0.040522144317626956, 0.03919279861450195, 0.03918588638305664, 0.0391923828125, 0.03957219314575195, 0.039288833618164064, 0.03920598220825195, 0.039025566101074216, 0.03917571258544922, 0.03917801666259765, 0.039072288513183596, 0.039220897674560544, 0.039010177612304686, 0.03896384048461914, 0.03891145706176758, 0.039043487548828124, 0.039026176452636716, 0.039434623718261716, 0.03989120101928711, 0.03944249725341797, 0.03893241500854492, 0.03923763275146484, 0.03898172760009765, 0.03903395080566406, 0.039093246459960936, 0.03908572769165039, 0.03881318283081055, 0.03879183959960938, 0.03866592025756836, 0.03891846466064453, 0.03999887847900391, 0.03904556655883789, 0.038551712036132814, 0.03845119857788086, 0.03920268630981445, 0.03899955368041992, 0.03890240097045899, 0.03877872085571289, 0.03876620864868164, 0.038455776214599606, 0.03851398468017578, 0.03846329498291016, 0.03863462448120117, 0.03848172760009766, 0.0390041618347168, 0.03900636672973633, 0.039198558807373045, 0.039131038665771486, 0.03933318328857422, 0.03920975875854492, 0.039433696746826175, 0.039249984741210935, 0.03890595245361328, 0.038965118408203124, 0.039112735748291015, 0.039526878356933595, 0.039245376586914064, 0.039131168365478516, 0.039691104888916015, 0.039272449493408204, 0.04541030502319336, 0.03937484741210937, 0.03950950241088867, 0.03945471954345703, 0.0392784309387207, 0.0396038703918457, 0.039397567749023435, 0.03933241653442383, 0.039495552062988284, 0.039475582122802735, 0.039360511779785154, 0.039282302856445316, 0.03952595138549805, 0.039140159606933594, 0.03994374465942383, 0.039206687927246096, 0.039209022521972656, 0.039041633605957034, 0.04006707382202149, 0.03893209457397461, 0.03925030517578125, 0.03924374389648438, 0.03879446411132813, 0.038881473541259766, 0.03858457565307617, 0.03849679946899414, 0.03861078262329101, 0.03842867279052734, 0.0385167350769043, 0.03870457458496094, 0.03838009643554687, 0.03862262344360352, 0.03850915145874023, 0.03853311920166016, 0.03851366424560547, 0.03840703964233398, 0.03878022384643555, 0.03849708938598633, 0.03909427261352539, 0.039261760711669924, 0.03910700988769531, 0.03938304138183594, 0.03920399856567383, 0.03914223861694336, 0.039505470275878904, 0.039774654388427734, 0.0393175048828125, 0.03963289642333984, 0.03934822463989258, 0.039293087005615235, 0.03917302322387695, 0.03941676712036133, 0.03948339080810547, 0.039419902801513675, 0.039693729400634765, 0.0394134407043457, 0.03949456024169922, 0.039645313262939456, 0.03939862442016601, 0.0395940170288086, 0.039546592712402344, 0.04001500701904297, 0.039836414337158205, 0.039124385833740234, 0.038442943572998045, 0.03866396713256836, 0.040225761413574215, 0.03870105743408203, 0.03854655838012695, 0.038839038848876954, 0.03893670272827148, 0.03851011276245117, 0.03849795150756836, 0.0384101448059082, 0.03845177459716797, 0.038441150665283204, 0.03855542373657227, 0.03845158386230469, 0.03829708862304688, 0.03836703872680664, 0.03844476699829102, 0.038487007141113284, 0.03859436798095703, 0.03867052841186523, 0.04043299102783203, 0.039614879608154296, 0.039549182891845704, 0.03916755294799805, 0.03917830276489258, 0.03935475158691406, 0.039179840087890626, 0.039077632904052736, 0.03913132858276367, 0.0391992301940918, 0.03897958374023437, 0.03949897766113281, 0.03916851043701172, 0.03915804672241211, 0.03900435256958008, 0.0387786865234375, 0.03868876647949219, 0.03868876647949219, 0.038966911315917965, 0.038663646697998044, 0.03876752090454102, 0.03939110565185547, 0.03914275360107422, 0.03888185501098633, 0.03885395050048828, 0.039059776306152344, 0.03883411026000977, 0.03850102233886719, 0.03887859344482422, 0.038666622161865234, 0.03851440048217773, 0.03874256134033203, 0.038490207672119144, 0.03881564712524414, 0.03877062225341797, 0.03852000045776367, 0.038699905395507814, 0.03887923049926758, 0.0385794563293457, 0.03940972900390625, 0.03936665725708008, 0.03906710433959961, 0.03904975891113281, 0.03959807968139648, 0.03955260848999023, 0.039350688934326174, 0.039213054656982424, 0.039820735931396484, 0.03925049591064453, 0.039317054748535155, 0.039442558288574216, 0.03907110214233398, 0.0391212158203125, 0.03901033782958984, 0.03903753662109375, 0.03883212661743164, 0.03873779296875, 0.03899532699584961, 0.038834304809570314, 0.03863843154907227, 0.03903776168823242, 0.03879811096191406, 0.038998046875, 0.039227550506591796, 0.03921654510498047, 0.04144803237915039, 0.038956737518310545, 0.03899628829956055, 0.03876012802124024, 0.03850223922729492, 0.03845987319946289, 0.03855721664428711, 0.03838111877441406, 0.03839888000488281, 0.038508544921875, 0.03845667266845703, 0.041138206481933594, 0.03907244873046875, 0.038708927154541016, 0.03881600189208984, 0.03864985656738281, 0.03890745544433594, 0.03896771240234375, 0.03858243179321289, 0.03849407958984375, 0.04115817642211914, 0.03872169494628906, 0.03843683242797852, 0.03849660873413086, 0.038368896484375, 0.0386187858581543, 0.03859667205810547, 0.03860448074340821, 0.03848223876953125, 0.03895391845703125, 0.039704288482666016, 0.03907174301147461, 0.03928793716430664, 0.03901887893676758, 0.03924224090576172, 0.03911065673828125, 0.03908758544921875, 0.03880809783935547, 0.03966595077514649, 0.03893494415283203, 0.03852492904663086, 0.03840358352661133, 0.03842508697509766, 0.03868467330932617, 0.03884848022460938, 0.039012382507324216, 0.03910246276855469, 0.039041023254394534, 0.03906268692016602, 0.03902483367919922, 0.03907180786132813, 0.039387744903564455, 0.03939731216430664, 0.03899808120727539, 0.03890585708618164, 0.03919820785522461, 0.0391992301940918, 0.038843711853027346, 0.038892223358154294, 0.038749313354492186, 0.03870809555053711, 0.038661312103271485, 0.038812000274658205, 0.038793472290039065, 0.039629024505615236, 0.0400849609375, 0.03909891128540039, 0.039032833099365234, 0.03902668762207031, 0.03973251342773437, 0.039277278900146484, 0.039198078155517575, 0.039293567657470704, 0.03915516662597656, 0.039332382202148436, 0.03959014511108398, 0.0396753921508789, 0.03968150329589844, 0.039446975708007814, 0.039083999633789064, 0.03919091033935547, 0.03918412780761719, 0.03893471908569336, 0.03896531295776367, 0.039049217224121094, 0.039016448974609375, 0.039190528869628906, 0.03946012878417969, 0.03944694519042969, 0.039629119873046875, 0.039394847869873045, 0.03953446578979492, 0.03957411193847656, 0.03942569732666015, 0.03968649673461914, 0.03944831848144531, 0.03938943862915039, 0.03933958435058594, 0.04362284851074219, 0.03954483032226563, 0.03944243240356445, 0.04016880035400391, 0.039348670959472656, 0.03923353576660156, 0.039300926208496095, 0.0392852783203125, 0.03922521591186524, 0.03951628875732422, 0.03936665725708008, 0.03937833786010742, 0.0392341423034668, 0.039394943237304685, 0.03970086288452149, 0.03967180633544922, 0.03934828948974609, 0.039312576293945314, 0.03913600158691406, 0.03929827117919922, 0.03882620620727539, 0.038986305236816406, 0.038803455352783206, 0.03857926559448242, 0.03850495910644531, 0.03877318572998047, 0.03856291198730469, 0.03852790451049805, 0.038573345184326174, 0.03888816070556641, 0.03899932861328125, 0.03910284805297851, 0.03941388702392578, 0.03931363296508789, 0.03980287933349609, 0.03911043167114258, 0.03920105743408203, 0.03930828857421875, 0.039113662719726563, 0.03892428970336914, 0.03893840026855469, 0.03915987014770508, 0.03896271896362305, 0.03897817611694336, 0.039230751037597655, 0.03930550384521484, 0.039449024200439456, 0.03946416091918945, 0.03971049499511719, 0.03945369720458984, 0.03952006530761719, 0.03932153701782227, 0.039551231384277345, 0.03940556716918946, 0.039387134552001955, 0.03965478515625, 0.03946355056762695, 0.03949977493286133, 0.03943388748168945, 0.039325504302978515, 0.039680545806884765, 0.03949929428100586, 0.03961471939086914, 0.0395, 0.03959603118896484, 0.04082876968383789, 0.03980204772949219, 0.03959894561767578, 0.03932156753540039, 0.03921920013427734, 0.03950592041015625, 0.039376895904541014, 0.03960319900512695, 0.039430721282958985, 0.0396923828125, 0.04012633514404297, 0.0394818229675293, 0.0395976333618164, 0.03931289672851562, 0.039010848999023434, 0.039489952087402344, 0.03988668823242188, 0.03945872116088867, 0.038946495056152344, 0.0390948486328125, 0.03925142288208008, 0.03919926452636719, 0.03928585433959961, 0.04003113555908203, 0.03934316635131836, 0.03969734573364258, 0.039773632049560546, 0.039634849548339846, 0.03948419189453125, 0.040287967681884765, 0.03998438262939453, 0.040465023040771486, 0.03992995071411133, 0.04022700881958008, 0.039895103454589846, 0.040083393096923825, 0.03973734283447266, 0.03988275146484375, 0.03961999893188477, 0.03933222579956055, 0.03924604797363281, 0.039198719024658206, 0.03909427261352539, 0.039065601348876954, 0.040382240295410154, 0.03973756790161133, 0.03940358352661133, 0.03982262420654297, 0.04093964767456055, 0.03904771041870117, 0.03912704086303711, 0.03878911972045898, 0.03920099258422852, 0.040629150390625, 0.03907900619506836, 0.03906092834472656, 0.03892185592651367, 0.03923952102661133, 0.03951475143432617, 0.040087329864501954, 0.03931321716308594, 0.039315166473388674, 0.03920582580566406, 0.03905023956298828, 0.03998118209838867, 0.039470272064208986, 0.0390904655456543, 0.03932419204711914, 0.03901567840576172, 0.039160575866699215, 0.039534591674804685, 0.04044800186157226, 0.039487232208251954, 0.03938751983642578, 0.039882625579833984, 0.03957110214233398, 0.039412063598632814, 0.03908153533935547, 0.039124736785888674, 0.042922687530517575, 0.039774208068847655, 0.039569408416748046, 0.039518207550048826, 0.04056883239746094, 0.039567039489746096, 0.03917446517944336, 0.03937484741210937, 0.039397087097167965, 0.039772449493408205, 0.039425857543945314, 0.04220332717895508, 0.039395328521728515, 0.03929849624633789, 0.03913580703735352, 0.03916185760498047, 0.03991551971435547, 0.03898489761352539, 0.03862611389160156, 0.03849625778198242, 0.03853311920166016, 0.03874390411376953, 0.038395488739013675, 0.03866476821899414, 0.038799072265625, 0.03881331253051758, 0.03893683242797852, 0.03870352172851563, 0.03901232147216797, 0.03957148742675781, 0.03980297470092774, 0.03981916809082031, 0.03969555282592773, 0.03970038223266602, 0.039682304382324216, 0.03976668930053711, 0.039774112701416016, 0.03982121658325195, 0.03970307159423828, 0.03955062484741211, 0.03924991989135742, 0.039279678344726565, 0.039574462890625, 0.039019775390625, 0.03919948959350586, 0.03893388748168945, 0.0389536018371582, 0.039806976318359374, 0.039007614135742184, 0.03877260971069336, 0.03833318328857422, 0.03852188873291015, 0.03854025650024414, 0.03827206420898437, 0.03819974517822266, 0.03820499038696289, 0.0382020149230957, 0.03854159927368164, 0.038606464385986326, 0.0386297607421875, 0.03903263854980469, 0.038846656799316405, 0.0390467529296875, 0.03876214218139649, 0.03980963134765625, 0.039780513763427734, 0.03992575836181641, 0.03991551971435547, 0.03995238494873047, 0.039823265075683595, 0.039612510681152346, 0.03942195129394531, 0.03949337768554687, 0.03951241683959961, 0.03984502410888672, 0.03977471923828125, 0.039524608612060544, 0.03958972930908203, 0.03986652755737305, 0.039172096252441405, 0.03922739028930664, 0.03906569671630859, 0.039118465423583985, 0.039151073455810544, 0.03900262451171875, 0.03886111831665039, 0.038782527923583984, 0.03903084945678711, 0.03913683319091797, 0.03912073516845703, 0.03952297592163086, 0.03900406265258789, 0.038582687377929685, 0.03848102569580078, 0.03857497787475586, 0.03879116821289062, 0.039272449493408204, 0.039371807098388674, 0.040106655120849606, 0.041148414611816404, 0.039405887603759765, 0.03928604888916016, 0.03935836791992187, 0.039594432830810544, 0.039250049591064456, 0.039145729064941404, 0.03913264083862305, 0.039066303253173826, 0.039540576934814456, 0.03934822463989258, 0.03945033645629883]",tokens/s,25.464580610019265,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1048.969216,965.67296,0.0,570.425344,536.326656,s,1,8.42441015625,8.42441015625,0.0,8.42441015625,8.42441015625,8.42441015625,8.42441015625,[8.42441015625],,kWh,3.914825401667864e-05,4.310742464555275e-06,1.281612136400101e-05,5.6275117845234926e-05,,MB,1284.702208,1034.878976,0.0,624.951296,594.377728,s,10,0.26368310546875,0.026368310546875003,0.0002944813357621705,0.026420432090759276,0.026643379402160643,0.026729434108734133,0.026798277873992922,"[0.026592159271240236, 0.026053504943847658, 0.026624256134033204, 0.02681548881530762, 0.026454591751098634, 0.02652511978149414, 0.02638627243041992, 0.025766719818115236, 0.026129184722900392, 0.02633580780029297]",tokens/s,9708.623521590745,kWh,7.829544049486437e-07,8.630932426148413e-08,5.197487029598338e-07,1.3890124321699618e-06,tokens/kWh,184303605.98001865,MB,1320.767488,1049.55904,0.0,639.63136,607.71072,s,10,14.155046142578126,1.4155046142578125,0.020044638068717466,1.4226021728515623,1.4353281616210938,1.439636065673828,1.4430823889160156,"[1.4244794921875, 1.4072734375, 1.4439439697265626, 1.4047330322265625, 1.428071533203125, 1.4275911865234374, 1.3831318359375, 1.3807259521484374, 1.434370849609375, 1.420724853515625]",tokens/s,44.50709617293096,kWh,4.129228820421656e-05,4.554160190913781e-06,1.5726347627237077e-05,6.157279602236742e-05,tokens/kWh,1023179.1321789923,,s,630,14.148588958740227,0.022458077712286084,0.0006456810369411583,0.022481712341308593,0.022870140266418457,0.023023983955383298,0.024744324131011963,"[0.022230239868164064, 0.022661535263061524, 0.022646303176879882, 0.022819456100463868, 0.022765184402465822, 0.022737503051757812, 0.022618112564086915, 0.022514879226684572, 0.02265990447998047, 0.02262015914916992, 0.022624256134033204, 0.02262835121154785, 0.02246451187133789, 0.02240121650695801, 0.022331104278564454, 0.02237654495239258, 0.022538368225097655, 0.02249235153198242, 0.022801088333129882, 0.02239244842529297, 0.022370687484741213, 0.022194271087646485, 0.02226371192932129, 0.022259616851806642, 0.02264179229736328, 0.02233238410949707, 0.022403072357177735, 0.022494400024414062, 0.02242639923095703, 0.02242767906188965, 0.022456352233886718, 0.022386655807495118, 0.022379871368408202, 0.022436511993408202, 0.022386528015136718, 0.02252012825012207, 0.02275926399230957, 0.022765663146972655, 0.022750463485717774, 0.02276790428161621, 0.02277619171142578, 0.022676992416381835, 0.02260799980163574, 0.023404928207397462, 0.02253647994995117, 0.02274070358276367, 0.02272870445251465, 0.022656351089477538, 0.02271913528442383, 0.025694400787353515, 0.023480127334594727, 0.02301923179626465, 0.02263478469848633, 0.022355583190917967, 0.02245804786682129, 0.02247478485107422, 0.022327455520629882, 0.022335487365722655, 0.02236262321472168, 0.02229471969604492, 0.022413408279418946, 0.02223865509033203, 0.022210847854614257, 0.021541919708251953, 0.021699552536010743, 0.02169593620300293, 0.021744480133056642, 0.02181292724609375, 0.021600288391113283, 0.02167398452758789, 0.02247270393371582, 0.022636640548706056, 0.021931936264038086, 0.02292870330810547, 0.022119007110595702, 0.021884416580200194, 0.021832544326782225, 0.021957536697387696, 0.022623071670532225, 0.02195193672180176, 0.021917247772216798, 0.02222591972351074, 0.022218751907348632, 0.022341184616088867, 0.022106559753417968, 0.022046720504760742, 0.022157087326049804, 0.02312390327453613, 0.02233100891113281, 0.02221232032775879, 0.021914560317993163, 0.022068384170532227, 0.022149759292602537, 0.02252207946777344, 0.02199955177307129, 0.022005823135375975, 0.02215936088562012, 0.02231881523132324, 0.02212278366088867, 0.022176767349243166, 0.02241958427429199, 0.02247283172607422, 0.02234601593017578, 0.02240764808654785, 0.02254643249511719, 0.02271843147277832, 0.022761856079101563, 0.022601343154907225, 0.022374591827392577, 0.02256675148010254, 0.02273855972290039, 0.02259721565246582, 0.022784799575805665, 0.022550527572631835, 0.022378496170043945, 0.02249113655090332, 0.022565088272094726, 0.02289641571044922, 0.022597984313964845, 0.022608575820922853, 0.022694879531860352, 0.022681503295898436, 0.022941919326782228, 0.022915103912353515, 0.022800479888916016, 0.022905664443969728, 0.022567487716674803, 0.023592384338378906, 0.02298486328125, 0.022836864471435545, 0.022844383239746094, 0.022733055114746093, 0.02266048049926758, 0.022923616409301757, 0.02274723243713379, 0.022918815612792968, 0.022958335876464845, 0.02289664077758789, 0.022748479843139647, 0.022710975646972657, 0.02279814338684082, 0.022712511062622072, 0.02285468864440918, 0.022933887481689452, 0.022849952697753906, 0.022847679138183592, 0.022763263702392577, 0.022997247695922853, 0.02292051124572754, 0.022944448471069336, 0.02304739189147949, 0.02282192039489746, 0.022801216125488282, 0.022708480834960937, 0.022758079528808595, 0.0226910400390625, 0.02277238464355469, 0.02323244857788086, 0.02267568016052246, 0.022734207153320314, 0.022651487350463868, 0.02277299118041992, 0.022581792831420897, 0.025107807159423828, 0.023477119445800783, 0.02338323211669922, 0.0227061767578125, 0.022720447540283205, 0.0229303035736084, 0.022675455093383787, 0.022935232162475585, 0.022612415313720702, 0.02292076873779297, 0.022589759826660158, 0.022910688400268556, 0.022802719116210936, 0.022703712463378906, 0.022647392272949218, 0.02275641632080078, 0.022675743103027345, 0.02266364860534668, 0.02272991943359375, 0.023200639724731444, 0.02551171112060547, 0.02286150360107422, 0.022741472244262696, 0.02254204750061035, 0.02268307113647461, 0.02280022430419922, 0.022249343872070313, 0.022563072204589845, 0.023061952590942382, 0.022554527282714842, 0.022545120239257813, 0.02247091293334961, 0.022449024200439455, 0.02235251235961914, 0.022627904891967775, 0.02249951934814453, 0.022523967742919922, 0.02254800033569336, 0.022458368301391602, 0.02219059181213379, 0.022142751693725586, 0.02208620834350586, 0.022031423568725585, 0.022264448165893555, 0.022249792098999025, 0.02225315284729004, 0.022200735092163085, 0.0219931526184082, 0.02188038444519043, 0.021945119857788086, 0.021933759689331055, 0.021831008911132814, 0.0241591682434082, 0.022920831680297852, 0.022227071762084962, 0.022195903778076172, 0.022137407302856446, 0.02189926338195801, 0.022084703445434572, 0.022037408828735353, 0.022169151306152345, 0.021963327407836915, 0.022085184097290038, 0.022168960571289063, 0.022094112396240234, 0.022012256622314454, 0.02210032081604004, 0.022177183151245117, 0.02250809669494629, 0.022284000396728516, 0.022129056930541992, 0.022218656539916993, 0.022446048736572265, 0.022401023864746093, 0.022374176025390626, 0.022182111740112306, 0.022257728576660155, 0.022124479293823242, 0.022063104629516602, 0.022116352081298828, 0.022044607162475586, 0.021835199356079103, 0.021891712188720703, 0.022025760650634767, 0.02226019287109375, 0.022475839614868164, 0.022703039169311524, 0.02260361671447754, 0.02277187156677246, 0.022630048751831056, 0.02281827163696289, 0.022805120468139647, 0.022776063919067384, 0.02276259231567383, 0.022795167922973633, 0.02275958442687988, 0.022668256759643554, 0.022815616607666015, 0.0228701114654541, 0.022869951248168947, 0.022789823532104493, 0.022687999725341797, 0.022605087280273436, 0.022616832733154298, 0.022611520767211915, 0.022647232055664063, 0.02268956756591797, 0.022579519271850586, 0.022534048080444336, 0.022419679641723634, 0.022355743408203125, 0.022331424713134766, 0.022402624130249023, 0.022618528366088866, 0.02252150344848633, 0.022464063644409178, 0.023083808898925782, 0.02253113555908203, 0.022402015686035157, 0.022338752746582032, 0.02244588851928711, 0.022416351318359375, 0.02265910339355469, 0.022857887268066406, 0.022740800857543944, 0.022626304626464845, 0.022585567474365235, 0.022735967636108398, 0.02264950370788574, 0.022586784362792968, 0.022610048294067382, 0.022679391860961913, 0.022612640380859375, 0.022902431488037108, 0.02264713668823242, 0.02268707275390625, 0.022573728561401368, 0.02277484893798828, 0.02296703910827637, 0.023137887954711913, 0.022805248260498047, 0.022820512771606447, 0.022620351791381835, 0.0225316162109375, 0.022723039627075194, 0.02264678382873535, 0.02272774314880371, 0.022524864196777343, 0.022482751846313476, 0.02257734489440918, 0.022591487884521484, 0.02263382339477539, 0.022296575546264647, 0.02247270393371582, 0.022562816619873048, 0.023068672180175782, 0.02245193672180176, 0.02238902473449707, 0.02245449638366699, 0.022599584579467775, 0.0226507511138916, 0.022616064071655274, 0.022597696304321287, 0.022534080505371094, 0.022494815826416017, 0.02284160041809082, 0.022648000717163087, 0.022466527938842774, 0.022537023544311523, 0.022449695587158203, 0.022506143569946287, 0.02247478485107422, 0.022665088653564452, 0.022691904067993166, 0.022765727996826173, 0.02272233581542969, 0.022800479888916016, 0.022888511657714845, 0.022847423553466795, 0.022665023803710938, 0.022679391860961913, 0.022782304763793945, 0.022734943389892577, 0.022713504791259765, 0.02268227195739746, 0.022478527069091796, 0.02260419273376465, 0.022769664764404295, 0.02276483154296875, 0.022657024383544923, 0.022596384048461916, 0.02263238334655762, 0.022760831832885742, 0.02263104057312012, 0.02273689651489258, 0.022572927474975586, 0.022982847213745116, 0.02248067283630371, 0.022706239700317384, 0.022696287155151366, 0.022655040740966796, 0.022545568466186522, 0.022585887908935547, 0.02415407943725586, 0.022650943756103516, 0.022558496475219725, 0.02252783966064453, 0.0222291202545166, 0.02223459243774414, 0.022532863616943358, 0.022775423049926757, 0.022685279846191408, 0.022871936798095703, 0.022520223617553712, 0.022563488006591796, 0.02228000068664551, 0.022599231719970702, 0.02259868812561035, 0.02249510383605957, 0.022646432876586915, 0.02288627243041992, 0.022534624099731445, 0.022411264419555665, 0.022458240509033202, 0.022691999435424805, 0.022903871536254884, 0.02255459213256836, 0.022387744903564454, 0.022339487075805665, 0.022369440078735352, 0.02219424057006836, 0.02207414436340332, 0.02189926338195801, 0.021794591903686523, 0.021588191986083985, 0.02152448081970215, 0.021470399856567384, 0.021520511627197265, 0.02165158462524414, 0.021717567443847657, 0.021770240783691407, 0.021890432357788085, 0.02168230438232422, 0.02177689552307129, 0.021614912033081055, 0.02165318489074707, 0.021642303466796874, 0.02165171241760254, 0.021748416900634764, 0.02195590400695801, 0.0218221435546875, 0.0216760311126709, 0.021702783584594727, 0.02170252799987793, 0.02203251266479492, 0.021796735763549804, 0.02178483200073242, 0.021927679061889648, 0.021696512222290038, 0.02175699234008789, 0.02161961555480957, 0.021741600036621095, 0.021705823898315428, 0.021734304428100586, 0.02166169548034668, 0.0216712646484375, 0.02199344062805176, 0.02184671974182129, 0.021614591598510743, 0.021962751388549806, 0.021805055618286134, 0.021679519653320312, 0.021831392288208008, 0.021601152420043946, 0.021540864944458008, 0.0220897274017334, 0.021940223693847655, 0.021694784164428712, 0.021395456314086913, 0.021712896347045898, 0.021675775527954102, 0.02169481658935547, 0.02166543960571289, 0.021649663925170898, 0.021673887252807618, 0.021743520736694336, 0.021651168823242188, 0.021694015502929688, 0.02165443229675293, 0.02172857666015625, 0.021656255722045898, 0.02170572853088379, 0.022193248748779298, 0.021867488861083983, 0.02181999969482422, 0.021706111907958986, 0.02170537567138672, 0.021672256469726564, 0.022026336669921875, 0.021776287078857422, 0.021911808013916016, 0.021849279403686524, 0.021915712356567384, 0.021823999404907226, 0.021964736938476562, 0.02200966453552246, 0.021903167724609374, 0.02172915267944336, 0.021859039306640626, 0.021864288330078124, 0.021890239715576174, 0.022182016372680663, 0.0219965763092041, 0.021898080825805664, 0.022161439895629884, 0.021939231872558595, 0.021934112548828124, 0.02176518440246582, 0.021797536849975586, 0.021700607299804688, 0.021745664596557617, 0.02184774398803711, 0.021850431442260742, 0.021765920639038087, 0.024104991912841798, 0.022753408432006836, 0.02234579277038574, 0.02225948715209961, 0.02225161552429199, 0.021964927673339844, 0.0218787841796875, 0.021749759674072267, 0.02194780731201172, 0.022061471939086915, 0.022114463806152344, 0.02198121643066406, 0.021916799545288086, 0.021775232315063477, 0.021831615447998047, 0.021843103408813475, 0.021963552474975587, 0.02179478454589844, 0.021985631942749023, 0.022177791595458983, 0.02188083267211914, 0.022029920578002928, 0.023253408432006836, 0.029998912811279296, 0.02225606346130371, 0.022163200378417968, 0.022011808395385742, 0.021987743377685547, 0.022122175216674804, 0.022115520477294922, 0.02216659164428711, 0.02209078407287598, 0.022051551818847655, 0.021993471145629884, 0.022378528594970703, 0.02201913642883301, 0.02197555160522461, 0.022140735626220702, 0.02336409568786621, 0.022612064361572266, 0.02259744071960449, 0.02255891227722168, 0.022618080139160158, 0.022609376907348634, 0.022784576416015626, 0.022644287109375, 0.02262063980102539, 0.022716384887695312, 0.022800479888916016, 0.022617279052734376, 0.022631135940551758, 0.022548479080200197, 0.022757343292236328, 0.02272591972351074, 0.02285385513305664, 0.02288092803955078, 0.02329737663269043, 0.022691743850708008, 0.02257369613647461, 0.022509567260742186, 0.02253363227844238, 0.022486623764038087, 0.0224736328125, 0.02253385543823242, 0.022399168014526367, 0.02242799949645996, 0.022594879150390625, 0.022448575973510743, 0.022367904663085938, 0.022295167922973633, 0.02226736068725586, 0.022261247634887696, 0.022811391830444335, 0.023198783874511717, 0.02328876876831055, 0.023117824554443358, 0.0228818244934082, 0.029682144165039063, 0.024685888290405272, 0.02232569694519043, 0.022047103881835936, 0.022380640029907226, 0.022466560363769532, 0.022769535064697265, 0.022400447845458984, 0.02231158447265625, 0.022340991973876952, 0.022120672225952147, 0.022204927444458008, 0.02229987144470215, 0.02232579231262207, 0.0221976318359375, 0.022252351760864257, 0.02232729530334473, 0.022468671798706055, 0.022394304275512696, 0.02230678367614746, 0.022261920928955077, 0.02233283233642578, 0.022385215759277342, 0.022448543548583985, 0.022361728668212892, 0.022400800704956054, 0.022436447143554687, 0.02249728012084961, 0.022548511505126954, 0.022783327102661132, 0.022585760116577147, 0.022823135375976564, 0.022595199584960937, 0.02256934356689453, 0.02257030487060547, 0.022589792251586915, 0.02250489616394043, 0.024867744445800782, 0.02332876777648926, 0.02274518394470215, 0.022632064819335936, 0.022769567489624023, 0.022870399475097655, 0.022716415405273437, 0.023000511169433593, 0.02267350387573242, 0.022567392349243164, 0.022333120346069334, 0.02231715202331543, 0.02249545669555664, 0.022681312561035158, 0.022849279403686522, 0.022771360397338868, 0.022710752487182618, 0.022309024810791014, 0.02228268814086914, 0.023027872085571287, 0.024768192291259764, 0.022228511810302734, 0.022177600860595705, 0.02204640007019043, 0.021866592407226562, 0.021965919494628908, 0.022081567764282228, 0.022271711349487303, 0.022380544662475587]",tokens/s,44.52740848131149,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1490.198528,1861.156864,0.0,1465.909248,1358.169088,s,1,8.92091796875,8.92091796875,0.0,8.92091796875,8.92091796875,8.92091796875,8.92091796875,[8.92091796875],,kWh,5.367274289999386e-05,5.892446202189604e-06,1.9234459832012663e-05,7.879964893419612e-05,,MB,1562.64448,1882.128384,0.0,1472.200704,1356.544512,s,10,0.5081317710876465,0.05081317710876465,0.00015486380894732922,0.05077542304992676,0.05091472702026367,0.05105760421752929,0.0511719059753418,"[0.05120048141479492, 0.050786014556884765, 0.05086569595336914, 0.050716190338134765, 0.050882335662841796, 0.05088297653198242, 0.05076483154296875, 0.05066867065429687, 0.05075222396850586, 0.05061235046386719]",tokens/s,5038.063245918215,kWh,1.5367383819298004e-06,1.693980679567003e-07,1.0171937962105986e-06,2.7233302460970993e-06,tokens/kWh,94002554.5439752,MB,1566.969856,1882.128384,0.0,1472.200704,1409.728,s,10,13.755901000976564,1.3755901000976565,0.010149455197494195,1.3747086181640626,1.3863067016601562,1.3911563415527344,1.395036053466797,"[1.3666591796875, 1.3815029296875, 1.37841259765625, 1.3679105224609376, 1.38522900390625, 1.3802357177734375, 1.371004638671875, 1.369410888671875, 1.3960059814453125, 1.359529541015625]",tokens/s,45.79852675264781,kWh,3.9929619992651026e-05,4.403737437878737e-06,1.818709934498727e-05,6.252045677551703e-05,tokens/kWh,1007670.181076968,,s,630,13.75327059555053,0.021830588246905612,0.0005318301540078125,0.021719439506530762,0.02215329875946045,0.02244603204727173,0.023784326286315927,"[0.021613216400146483, 0.022085823059082032, 0.021815040588378905, 0.0216680965423584, 0.021667808532714845, 0.021643295288085937, 0.021800960540771484, 0.02176527976989746, 0.021879648208618162, 0.02168422317504883, 0.021696512222290038, 0.021541919708251953, 0.02150704002380371, 0.0214835205078125, 0.021432319641113282, 0.021542495727539062, 0.02154742431640625, 0.02147532844543457, 0.021661632537841796, 0.021585248947143556, 0.021725471496582032, 0.02163142395019531, 0.021708927154541015, 0.022439136505126953, 0.02157548713684082, 0.021566303253173828, 0.021622783660888673, 0.021726463317871095, 0.021621631622314452, 0.021550079345703126, 0.02178963279724121, 0.02153785514831543, 0.0215230712890625, 0.021641183853149414, 0.021709312438964845, 0.02167990493774414, 0.021585439682006834, 0.02156742477416992, 0.02157417678833008, 0.021571264266967774, 0.021595584869384767, 0.021650304794311525, 0.021582048416137697, 0.021617504119873048, 0.02160508728027344, 0.021489952087402345, 0.021641151428222656, 0.021610496520996093, 0.022153215408325197, 0.021655904769897462, 0.02179987144470215, 0.02160326385498047, 0.021557024002075195, 0.02169599914550781, 0.021551296234130858, 0.021716543197631836, 0.02183776092529297, 0.021844224929809572, 0.021924415588378907, 0.021893440246582033, 0.0219586238861084, 0.021964351654052736, 0.021996864318847655, 0.022323711395263672, 0.02274070358276367, 0.022243200302124025, 0.02209164810180664, 0.02187494468688965, 0.021698175430297853, 0.021667871475219726, 0.021665952682495118, 0.02226425552368164, 0.02184921646118164, 0.02217024040222168, 0.021674016952514648, 0.021690336227416993, 0.021659231185913085, 0.021671615600585937, 0.021672672271728515, 0.02161017608642578, 0.021604671478271484, 0.021631263732910157, 0.021666847229003906, 0.02176041603088379, 0.02166934394836426, 0.022010431289672852, 0.021982879638671876, 0.02301011276245117, 0.022120223999023438, 0.022070943832397463, 0.021960351943969728, 0.022176095962524414, 0.022585920333862305, 0.02421084785461426, 0.02257139205932617, 0.022220256805419922, 0.022251840591430663, 0.022133983612060548, 0.022149663925170898, 0.021976959228515624, 0.021926496505737306, 0.021749151229858397, 0.021627487182617186, 0.02160256004333496, 0.021640159606933593, 0.021815616607666014, 0.02167228889465332, 0.02174086380004883, 0.02167046356201172, 0.02160870361328125, 0.021593856811523437, 0.02177168083190918, 0.021775232315063477, 0.02168396759033203, 0.021597984313964844, 0.02154364776611328, 0.021581279754638673, 0.021588512420654297, 0.021515039443969725, 0.02184009552001953, 0.02181769561767578, 0.02197542381286621, 0.021849376678466797, 0.02182406425476074, 0.021991519927978515, 0.02190332794189453, 0.022125984191894533, 0.021993728637695314, 0.02211199951171875, 0.021815200805664063, 0.021950368881225587, 0.021970752716064454, 0.022051807403564452, 0.022048191070556642, 0.026231039047241212, 0.022007904052734374, 0.02209814453125, 0.021773984909057617, 0.0218702392578125, 0.021920448303222657, 0.02166988754272461, 0.021588031768798827, 0.021597503662109375, 0.021589855194091796, 0.02164179229736328, 0.021725120544433593, 0.021469247817993163, 0.021561567306518554, 0.02152448081970215, 0.021526239395141603, 0.021641504287719725, 0.0216944637298584, 0.021901216506958008, 0.021958751678466795, 0.021798912048339843, 0.02168550491333008, 0.021549503326416017, 0.021582143783569336, 0.02153654479980469, 0.02173734474182129, 0.022026592254638672, 0.02167398452758789, 0.02195631980895996, 0.021469472885131836, 0.021575040817260742, 0.02172492790222168, 0.021758848190307618, 0.022183135986328126, 0.021911712646484376, 0.02179340744018555, 0.021844255447387696, 0.021860063552856444, 0.02195408058166504, 0.021957088470458984, 0.02191564750671387, 0.0219238395690918, 0.02270534324645996, 0.02182364845275879, 0.021823423385620117, 0.021830368041992187, 0.021620704650878907, 0.021544544219970704, 0.02228860855102539, 0.021835647583007812, 0.021688703536987305, 0.02172265625, 0.021676383972167968, 0.021616735458374024, 0.02152774429321289, 0.021600000381469725, 0.024475872039794924, 0.022559295654296874, 0.021643264770507813, 0.021704959869384765, 0.021891008377075194, 0.02201603126525879, 0.02172431945800781, 0.021658239364624025, 0.02158582305908203, 0.021520351409912108, 0.021495935440063476, 0.021544960021972655, 0.02152396774291992, 0.021488096237182615, 0.021628704071044922, 0.021528831481933595, 0.021700607299804688, 0.02187868881225586, 0.022249631881713867, 0.021604095458984375, 0.021485408782958983, 0.021625471115112305, 0.021522144317626953, 0.021549280166625977, 0.021425952911376955, 0.02150998306274414, 0.021403936386108397, 0.021442304611206053, 0.021451040267944334, 0.021468191146850585, 0.02156368064880371, 0.02147292709350586, 0.021468032836914064, 0.021497856140136717, 0.02162403106689453, 0.021523487091064452, 0.021996992111206055, 0.021713184356689452, 0.021624767303466796, 0.021512287139892578, 0.02149171257019043, 0.02143846321105957, 0.021521856307983398, 0.021496383666992188, 0.021536672592163086, 0.021415903091430665, 0.021479808807373046, 0.021474720001220703, 0.02148137664794922, 0.02147782325744629, 0.021579776763916016, 0.021741855621337892, 0.02163030433654785, 0.021844575881958008, 0.02188230323791504, 0.021813600540161134, 0.021933664321899415, 0.02193459129333496, 0.021980831146240234, 0.02209612846374512, 0.022370304107666016, 0.022097408294677736, 0.02198303985595703, 0.021686975479125976, 0.021526527404785157, 0.021564640045166016, 0.021691167831420898, 0.021637439727783203, 0.02178220748901367, 0.021549055099487305, 0.021557247161865235, 0.021493631362915037, 0.02146112060546875, 0.021557247161865235, 0.02159401512145996, 0.02155897521972656, 0.02161324882507324, 0.02147260856628418, 0.021651840209960936, 0.022077856063842775, 0.022388320922851562, 0.022321151733398437, 0.022180192947387694, 0.022123680114746094, 0.022041088104248048, 0.02200713539123535, 0.022143808364868164, 0.022259552001953124, 0.021983392715454103, 0.02194144058227539, 0.02196950340270996, 0.021921855926513672, 0.022035743713378905, 0.021886848449707032, 0.021995552062988283, 0.021977920532226563, 0.02230409622192383, 0.02203865623474121, 0.021826271057128907, 0.02187241554260254, 0.021813056945800782, 0.021766368865966796, 0.02189913558959961, 0.02186662483215332, 0.021764095306396485, 0.021732864379882814, 0.02185206413269043, 0.02160700798034668, 0.021536031723022462, 0.02144112014770508, 0.021545087814331055, 0.02161664009094238, 0.02190336036682129, 0.021511743545532227, 0.02165315246582031, 0.02143017578125, 0.021822336196899415, 0.02184543991088867, 0.022589439392089843, 0.028974687576293946, 0.02359600067138672, 0.022169055938720702, 0.022321952819824218, 0.021992992401123047, 0.0220263671875, 0.022234304428100586, 0.022154048919677736, 0.02194950485229492, 0.021844928741455077, 0.021921760559082033, 0.02173161506652832, 0.02174950408935547, 0.02182310485839844, 0.021796512603759765, 0.02177507209777832, 0.02244963264465332, 0.023236352920532225, 0.02193014335632324, 0.021885568618774415, 0.021719039916992186, 0.021937631607055665, 0.021923519134521483, 0.02164956855773926, 0.021628671646118165, 0.021658048629760743, 0.021921440124511717, 0.022119264602661132, 0.021914848327636717, 0.02387228775024414, 0.023461984634399413, 0.021772192001342772, 0.021587968826293946, 0.021569696426391602, 0.021558143615722655, 0.021758752822875975, 0.021954559326171876, 0.021967039108276368, 0.02191702461242676, 0.021924287796020507, 0.021805280685424804, 0.021815296173095702, 0.02195894432067871, 0.02194985580444336, 0.021995872497558595, 0.021880800247192383, 0.02173686408996582, 0.021723743438720702, 0.021816448211669923, 0.021572479248046873, 0.021735328674316406, 0.021979232788085938, 0.021763519287109376, 0.02163564872741699, 0.021769792556762695, 0.021602272033691406, 0.021588447570800782, 0.021646751403808593, 0.021518943786621093, 0.02144211196899414, 0.021768640518188477, 0.021445632934570313, 0.02162816047668457, 0.022441631317138673, 0.021637216567993164, 0.021765600204467772, 0.02197590446472168, 0.022206047058105468, 0.021868640899658204, 0.02228022384643555, 0.022071264266967774, 0.022108160018920898, 0.02201190376281738, 0.022044927597045898, 0.022058752059936522, 0.021909343719482423, 0.021891231536865233, 0.021907360076904296, 0.02175766372680664, 0.021636831283569337, 0.021740192413330077, 0.021609920501708985, 0.021600927352905273, 0.02168003273010254, 0.02167807960510254, 0.021542335510253908, 0.021496192932128906, 0.021588415145874024, 0.021453632354736327, 0.02151315116882324, 0.02150102424621582, 0.02150217628479004, 0.02165020751953125, 0.021780384063720702, 0.021513599395751953, 0.02207551956176758, 0.021658111572265625, 0.021851648330688478, 0.02175859260559082, 0.02168614387512207, 0.021638431549072266, 0.021597055435180663, 0.021728864669799806, 0.02159663963317871, 0.0216114559173584, 0.02158064079284668, 0.021569856643676756, 0.021524160385131837, 0.021587488174438476, 0.021520864486694335, 0.021665504455566406, 0.021573247909545897, 0.02161497688293457, 0.021612800598144532, 0.021686304092407228, 0.021808671951293945, 0.021967327117919922, 0.02243084716796875, 0.021975936889648436, 0.021985279083251954, 0.021966848373413086, 0.021966848373413086, 0.02192207908630371, 0.02213430404663086, 0.02191744041442871, 0.021874303817749022, 0.02187884712219238, 0.021584320068359374, 0.02156166458129883, 0.021610496520996093, 0.021774368286132814, 0.02172675132751465, 0.021722496032714842, 0.02150092887878418, 0.0216494083404541, 0.021575679779052736, 0.021706592559814452, 0.021603647232055663, 0.02152534484863281, 0.022009632110595704, 0.021544160842895507, 0.022028671264648438, 0.022018688201904297, 0.02199087905883789, 0.021679647445678712, 0.021588991165161133, 0.02162483215332031, 0.0215285758972168, 0.021408063888549805, 0.02150092887878418, 0.02159891128540039, 0.02167318344116211, 0.021501056671142577, 0.02150774383544922, 0.021736608505249024, 0.021848192214965822, 0.021660383224487306, 0.02177039909362793, 0.021797056198120116, 0.021664863586425782, 0.022116191864013673, 0.02167411231994629, 0.02159062385559082, 0.021942335128784178, 0.02220025634765625, 0.021827583312988282, 0.021858112335205078, 0.02177043151855469, 0.021740928649902343, 0.02170694351196289, 0.021629375457763673, 0.02145894432067871, 0.021498144149780272, 0.0215852165222168, 0.021821887969970703, 0.021650400161743164, 0.021556224822998047, 0.02163539123535156, 0.021694496154785158, 0.021444255828857423, 0.021553152084350585, 0.021626880645751953, 0.02152988815307617, 0.02194095993041992, 0.02212819290161133, 0.0219051513671875, 0.021817312240600586, 0.021780448913574217, 0.021777408599853516, 0.022009599685668946, 0.021946367263793946, 0.02184716796875, 0.02206399917602539, 0.02191564750671387, 0.021919904708862306, 0.022091487884521484, 0.022042911529541017, 0.022021184921264647, 0.022110687255859376, 0.022075231552124024, 0.022236831665039064, 0.02181999969482422, 0.021727615356445313, 0.021744895935058593, 0.022137088775634764, 0.021722848892211915, 0.021719839096069334, 0.02201919937133789, 0.021950464248657226, 0.021842655181884767, 0.02181052780151367, 0.021689439773559572, 0.021653215408325197, 0.02183782386779785, 0.021826719284057616, 0.021619903564453126, 0.021661088943481444, 0.02185775947570801, 0.021805791854858397, 0.022106176376342775, 0.022526208877563476, 0.02248678398132324, 0.022536191940307617, 0.022476383209228516, 0.022396575927734374, 0.022626047134399415, 0.0227476806640625, 0.027306463241577147, 0.02386124801635742, 0.022891712188720704, 0.022737024307250976, 0.02265590476989746, 0.022454048156738283, 0.022736799240112304, 0.02254857635498047, 0.02272051239013672, 0.02231705665588379, 0.02209369659423828, 0.02194633674621582, 0.022667423248291015, 0.02208483123779297, 0.022051071166992186, 0.02211689567565918, 0.022370464324951173, 0.021884607315063476, 0.021692928314208985, 0.021583360671997072, 0.02165155220031738, 0.021517440795898436, 0.021602399826049806, 0.02144883155822754, 0.02154159927368164, 0.021571584701538086, 0.021536415100097656, 0.021631519317626954, 0.02189228820800781, 0.02197104072570801, 0.021752351760864257, 0.02164169692993164, 0.021532032012939455, 0.021899904251098633, 0.021468896865844727, 0.021408159255981444, 0.021547008514404296, 0.02143164825439453, 0.021460927963256837, 0.0215150089263916, 0.02161033630371094, 0.021573631286621094, 0.021640832901000977, 0.02158131217956543, 0.02145564842224121, 0.021573728561401367, 0.021874176025390626, 0.021486080169677735, 0.021531871795654297, 0.021501983642578125, 0.021409631729125977, 0.02139753532409668, 0.021481632232666015, 0.021409568786621095, 0.021506528854370117, 0.021636703491210937, 0.021386112213134766, 0.021499391555786132, 0.021395263671875, 0.021631359100341797, 0.021420095443725588, 0.021516544342041016, 0.021360639572143555, 0.021444095611572265, 0.021538528442382812, 0.02166793632507324, 0.022164159774780274, 0.02159395217895508, 0.021512351989746093, 0.021675167083740236, 0.02156220817565918, 0.021464191436767577, 0.021497856140136717, 0.0214168643951416, 0.021491680145263672, 0.02143436813354492, 0.02146303939819336, 0.02157535934448242, 0.021425920486450194, 0.021492128372192384, 0.021706880569458006, 0.021501983642578125, 0.02149184036254883, 0.021491584777832032, 0.02147123146057129, 0.02176153564453125, 0.021498176574707033, 0.02165551948547363, 0.021991584777832033, 0.021856319427490233, 0.02164531135559082, 0.021760255813598632, 0.022137920379638673, 0.02208633613586426]",tokens/s,45.80728602866419,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,6655.623168,9387.835392,0.0,8992.587776,8404.320768,s,1,14.1039892578125,14.1039892578125,0.0,14.1039892578125,14.1039892578125,14.1039892578125,14.1039892578125,[14.1039892578125],,kWh,0.00020940675901246473,2.3088364685130984e-05,9.230840717999733e-05,0.00032480353087759305,,MB,1649.729536,9402.515456,0.0,8992.587776,7880.40704,s,10,3.451050415039062,0.3451050415039062,0.0007356105670003265,0.34543060302734374,0.34567911071777346,0.34580835113525393,0.3459117434692383,"[0.34342340087890627, 0.34508074951171874, 0.3455641784667969, 0.34434658813476565, 0.3445519104003906, 0.3459375915527344, 0.345650390625, 0.3454937133789063, 0.34536749267578126, 0.3456343994140625]",tokens/s,741.8031300974268,kWh,1.0099989562068839e-05,1.113836428620258e-06,6.726509212620348e-06,1.7940335203309443e-05,tokens/kWh,14269521.561268032,MB,1654.054912,9402.515456,0.0,8992.587776,8123.606528,s,10,25.017800048828125,2.5017800048828125,0.004788870374114529,2.5015008544921873,2.5062755859375,2.50874619140625,2.51072267578125,"[2.491120361328125, 2.499575439453125, 2.50334716796875, 2.501220703125, 2.500439453125, 2.501781005859375, 2.50078271484375, 2.50258984375, 2.5057265625, 2.511216796875]",tokens/s,25.182070316750742,kWh,7.290349514543089e-05,8.04132675155834e-06,4.832889594278162e-05,0.00012927371783977084,tokens/kWh,487338.0378684998,,s,630,25.014439548492433,0.03970545960078164,0.0003530860509353926,0.03967923164367676,0.04011471900939941,0.04025527458190918,0.04061889789581299,"[0.03986508941650391, 0.039329792022705076, 0.03907360076904297, 0.039225536346435545, 0.03890790557861328, 0.038934528350830076, 0.038956737518310545, 0.03900447845458985, 0.03896096038818359, 0.03898387145996094, 0.03897244644165039, 0.039081024169921874, 0.03904425430297852, 0.039357120513916016, 0.03925145721435547, 0.0391317138671875, 0.03911219024658203, 0.039131649017333986, 0.039114494323730466, 0.03923820877075195, 0.03917078399658203, 0.03921123123168945, 0.03916572952270508, 0.03923616027832031, 0.03933020782470703, 0.039411102294921875, 0.03943280029296875, 0.03953641510009766, 0.03940169525146484, 0.03929087829589844, 0.03934105682373047, 0.039314430236816404, 0.039375873565673826, 0.03936153411865234, 0.039409217834472654, 0.03942854309082031, 0.03946393585205078, 0.03972512054443359, 0.039560127258300784, 0.039617599487304686, 0.0397116813659668, 0.0399637451171875, 0.0400513916015625, 0.03991475296020508, 0.03992435073852539, 0.04002646255493164, 0.04001792144775391, 0.03998720169067383, 0.040048641204833986, 0.04004438400268555, 0.04009308624267578, 0.03997158432006836, 0.0401797103881836, 0.03999916839599609, 0.03994246292114258, 0.03991097640991211, 0.03992620849609375, 0.03992918395996094, 0.03995305633544922, 0.040033760070800783, 0.040174110412597656, 0.04012646484375, 0.04041030502319336, 0.03999948883056641, 0.03946614456176758, 0.03911151885986328, 0.039007553100585936, 0.03913593673706055, 0.03915273666381836, 0.03919760131835937, 0.03927859115600586, 0.03928255844116211, 0.03907740783691406, 0.03919113540649414, 0.039152767181396486, 0.03932659149169922, 0.039378944396972655, 0.039304351806640624, 0.03935062408447266, 0.03933849716186524, 0.039193599700927735, 0.03920588684082031, 0.039213054656982424, 0.039403518676757815, 0.03924991989135742, 0.03933747100830078, 0.03956911849975586, 0.03968428802490234, 0.03956329727172852, 0.0393647346496582, 0.03940396881103515, 0.03948748779296875, 0.03952844619750977, 0.03940383911132812, 0.03947017669677735, 0.03940428924560547, 0.03947708892822266, 0.03938508987426758, 0.03945993423461914, 0.03944300842285156, 0.03948281478881836, 0.039545761108398435, 0.039610366821289066, 0.039628288269042966, 0.03976787185668945, 0.03985670471191406, 0.04022284698486328, 0.03989299011230469, 0.0400211181640625, 0.04009471893310547, 0.04008127975463867, 0.040005630493164065, 0.0400076789855957, 0.04006707382202149, 0.04023910522460938, 0.04012851333618164, 0.04016537475585937, 0.04018758392333984, 0.0406572151184082, 0.040321025848388675, 0.04031462478637695, 0.04034326553344727, 0.040298080444335936, 0.040319934844970706, 0.04151091384887695, 0.04040281677246094, 0.040008705139160154, 0.03945369720458984, 0.03962252807617188, 0.03921728134155274, 0.03914547348022461, 0.039194561004638674, 0.03925158309936523, 0.039430591583251955, 0.03948339080810547, 0.03997081756591797, 0.03926630401611328, 0.03928268814086914, 0.03920832061767578, 0.039216991424560546, 0.03911964797973633, 0.03916185760498047, 0.039289886474609376, 0.03941270446777344, 0.039360511779785154, 0.03953452682495117, 0.03937411117553711, 0.039478240966796876, 0.039813983917236326, 0.039555614471435546, 0.03965792083740234, 0.039651199340820314, 0.04068979263305664, 0.0402940788269043, 0.039524673461914066, 0.03947660827636719, 0.039524993896484374, 0.03966476821899414, 0.03958668899536133, 0.03966566467285156, 0.03989913558959961, 0.03978192138671875, 0.03968252944946289, 0.03963228988647461, 0.039583518981933595, 0.03966649627685547, 0.03982329559326172, 0.03976947021484375, 0.0396561279296875, 0.03966156768798828, 0.039640384674072264, 0.039736129760742187, 0.03984515380859375, 0.039922271728515625, 0.03989913558959961, 0.039888320922851564, 0.03985260772705078, 0.03988275146484375, 0.03995423889160156, 0.039876705169677736, 0.03974563217163086, 0.04027350234985352, 0.04225606536865235, 0.040241310119628906, 0.04014956665039063, 0.04039254379272461, 0.04028176116943359, 0.04017407989501953, 0.0402303352355957, 0.03999337768554687, 0.04015718460083008, 0.04039907073974609, 0.039617984771728516, 0.03936492919921875, 0.03950595092773437, 0.03957964706420898, 0.039485439300537106, 0.03954687881469727, 0.03948748779296875, 0.03951001739501953, 0.039425121307373044, 0.039430335998535154, 0.03964387130737305, 0.0395489273071289, 0.03950592041015625, 0.039347969055175784, 0.03934028625488281, 0.03957724761962891, 0.03954723358154297, 0.039483200073242186, 0.03957779312133789, 0.039585792541503906, 0.03964313507080078, 0.0396943359375, 0.03971481704711914, 0.039534591674804685, 0.03969036865234375, 0.03959590530395508, 0.039556320190429685, 0.03969331359863281, 0.03965315246582031, 0.039671134948730466, 0.039596702575683596, 0.039570430755615234, 0.0395777587890625, 0.039535457611083985, 0.03953049468994141, 0.039657470703125, 0.03998857498168945, 0.03979727935791016, 0.03983526229858399, 0.039819774627685545, 0.0398131217956543, 0.039972862243652346, 0.03994214248657227, 0.039731201171875, 0.03969023895263672, 0.039739391326904294, 0.03987865447998047, 0.03978649520874023, 0.040016990661621094, 0.03978742218017578, 0.03976396942138672, 0.03977190399169922, 0.039973121643066406, 0.03981926345825195, 0.039981056213378906, 0.03995024108886719, 0.039933055877685544, 0.039803871154785155, 0.03981244659423828, 0.03975987243652344, 0.040004543304443356, 0.03956265640258789, 0.039293567657470704, 0.03926371383666992, 0.03937331390380859, 0.039624126434326175, 0.03932396697998047, 0.039336193084716795, 0.039346176147460936, 0.0393955192565918, 0.03952211380004883, 0.03941580963134766, 0.03946905517578125, 0.03950371170043945, 0.03941584014892578, 0.039462879180908204, 0.03943171310424805, 0.03990591812133789, 0.03975328063964844, 0.03955551910400391, 0.039562305450439456, 0.03946499252319336, 0.0394986572265625, 0.03958169555664062, 0.039684097290039064, 0.039724224090576174, 0.039570240020751955, 0.03962220764160156, 0.03970502471923828, 0.039696449279785155, 0.03988054275512695, 0.0395838394165039, 0.039595169067382814, 0.0396030387878418, 0.03963289642333984, 0.0395489273071289, 0.03952640151977539, 0.03953609466552734, 0.039515968322753905, 0.03966230392456055, 0.03967180633544922, 0.03978630447387695, 0.039665088653564454, 0.03971750259399414, 0.03967808151245117, 0.03976192092895508, 0.03975167846679688, 0.03984998321533203, 0.039825408935546876, 0.03999641418457031, 0.04005376052856445, 0.040080638885498045, 0.04003097534179687, 0.0400074577331543, 0.03992803192138672, 0.03997919845581055, 0.04007097625732422, 0.04006054306030273, 0.03992816162109375, 0.04003334426879883, 0.04001276779174805, 0.040032257080078126, 0.04005068969726563, 0.03992374420166016, 0.039288833618164064, 0.038968353271484374, 0.03905209732055664, 0.03937260818481445, 0.039351806640625, 0.03947004699707031, 0.039435550689697264, 0.0400263671875, 0.03942009735107422, 0.039358047485351565, 0.03929964828491211, 0.039428096771240234, 0.03948716735839844, 0.03940998458862305, 0.03934975814819336, 0.03935078430175781, 0.03931955337524414, 0.03933776092529297, 0.03936892700195312, 0.039415870666503906, 0.03961587142944336, 0.039694526672363284, 0.039734817504882815, 0.03976688003540039, 0.039790592193603515, 0.04046438217163086, 0.03976521682739258, 0.039747425079345707, 0.03976691055297851, 0.039737407684326174, 0.040170913696289064, 0.03981372833251953, 0.03970457458496094, 0.039585758209228515, 0.03966336059570313, 0.03959017562866211, 0.03971379089355469, 0.039508255004882815, 0.04063100814819336, 0.039839744567871094, 0.039897087097167966, 0.03998025512695313, 0.039852832794189455, 0.039812671661376954, 0.039850433349609374, 0.03992079925537109, 0.04005974578857422, 0.03995663833618164, 0.03988243103027344, 0.03986140823364258, 0.03980972671508789, 0.03991782379150391, 0.03982076644897461, 0.03975347137451172, 0.03975183868408203, 0.03978332901000976, 0.03989868927001953, 0.04011439895629883, 0.040048641204833986, 0.03988275146484375, 0.03993929672241211, 0.03996137619018555, 0.03987308883666992, 0.03935641479492188, 0.039239070892333985, 0.03930976104736328, 0.03931468963623047, 0.0392911376953125, 0.03920742416381836, 0.039094432830810544, 0.039188480377197264, 0.03963667297363281, 0.03966793441772461, 0.03946096038818359, 0.03939152145385742, 0.03944028854370117, 0.03941151809692383, 0.03934988784790039, 0.03936016082763672, 0.03945750427246094, 0.039567230224609375, 0.03967795181274414, 0.04005440139770508, 0.04031673431396485, 0.03964995193481445, 0.039862079620361326, 0.03958560180664063, 0.039667713165283204, 0.039726848602294924, 0.03967628860473633, 0.04004278564453125, 0.03970572662353516, 0.03962543869018555, 0.03985635375976562, 0.04058924865722656, 0.039946239471435545, 0.03979241561889649, 0.039799007415771484, 0.03963644790649414, 0.039803424835205076, 0.0401324462890625, 0.03968406295776367, 0.03969247817993164, 0.039671390533447266, 0.03955257415771484, 0.039648353576660154, 0.039651073455810544, 0.03965542221069336, 0.03963875198364258, 0.039680286407470705, 0.03972710418701172, 0.0398131217956543, 0.03981951904296875, 0.03994188690185547, 0.03978649520874023, 0.039874561309814455, 0.0399013442993164, 0.039868255615234376, 0.03986636734008789, 0.03983052825927735, 0.03976704025268555, 0.03993936157226562, 0.039813854217529296, 0.03995852661132813, 0.03995564651489258, 0.03999055862426758, 0.03950460815429688, 0.03931497573852539, 0.03923401641845703, 0.0394354248046875, 0.03926704025268555, 0.03933196640014648, 0.039706623077392575, 0.039425281524658205, 0.03917900848388672, 0.03934003067016602, 0.03925753784179688, 0.0393447036743164, 0.03954073715209961, 0.039462913513183595, 0.03946227264404297, 0.0396622085571289, 0.03968000030517578, 0.03936665725708008, 0.03940678405761719, 0.03952288055419922, 0.03945830535888672, 0.03951827239990234, 0.03958147048950195, 0.039478240966796876, 0.03966918563842774, 0.039682559967041016, 0.03963904190063477, 0.03960374450683594, 0.0399150390625, 0.039529407501220706, 0.039515617370605466, 0.03950428771972656, 0.03976764678955078, 0.0396824951171875, 0.039816608428955076, 0.039662273406982425, 0.039593505859375, 0.03971120071411133, 0.039702529907226565, 0.03975964736938477, 0.0400118408203125, 0.0399013442993164, 0.03975289535522461, 0.03981600189208984, 0.039882110595703124, 0.03987519836425781, 0.03993190383911133, 0.03993740844726563, 0.040471168518066404, 0.039981056213378906, 0.040065025329589846, 0.03993360137939453, 0.040122207641601564, 0.04008396911621094, 0.040035774230957034, 0.04005535888671875, 0.04014080047607422, 0.0401715202331543, 0.040269054412841794, 0.0401761589050293, 0.040261856079101564, 0.04012406539916992, 0.04000156784057617, 0.03958227157592773, 0.039352127075195316, 0.03924009704589844, 0.03929679870605469, 0.03936870574951172, 0.03947897720336914, 0.03928649520874023, 0.03946966552734375, 0.03940095901489258, 0.0395863037109375, 0.039428096771240234, 0.03960422515869141, 0.03948748779296875, 0.03954483032226563, 0.03945033645629883, 0.039372768402099606, 0.03957382583618164, 0.03960627365112305, 0.03939123153686523, 0.03946700668334961, 0.04038860702514648, 0.03962060928344727, 0.039671710968017575, 0.03985795211791992, 0.039841697692871096, 0.03980534362792969, 0.03986636734008789, 0.03967705535888672, 0.039719806671142575, 0.0396124153137207, 0.039613792419433594, 0.03966124725341797, 0.03967667388916016, 0.0396638412475586, 0.0397209587097168, 0.039798782348632815, 0.03983103942871094, 0.03982704162597656, 0.039785377502441405, 0.039686145782470705, 0.039779937744140625, 0.03986588668823242, 0.039869312286376954, 0.03978361511230469, 0.04031366348266602, 0.0401715202331543, 0.03995782470703125, 0.04002067184448242, 0.040030113220214845, 0.03997670364379883, 0.040013599395751956, 0.04009632110595703, 0.0401899528503418, 0.0400912971496582, 0.0399958381652832, 0.04003801727294922, 0.04013808059692383, 0.04002912139892578, 0.04016361618041992, 0.04007814407348633, 0.04007414245605469, 0.04034764862060547, 0.0401448974609375, 0.03967795181274414, 0.039392833709716794, 0.039356864929199216, 0.03964518356323242, 0.03927654266357422, 0.039357696533203125, 0.03928128051757813, 0.039416961669921875, 0.03934105682373047, 0.03942195129394531, 0.03950796890258789, 0.03969164657592773, 0.03957004928588867, 0.039495681762695314, 0.03948300933837891, 0.03961183929443359, 0.03947171020507813, 0.03948704147338867, 0.03945462417602539, 0.03955785751342773, 0.03950403213500977, 0.03957964706420898, 0.0399441909790039, 0.04011164855957031, 0.04000201416015625, 0.03974870300292969, 0.03992623901367188, 0.03970256042480469, 0.039781822204589846, 0.04150780868530273, 0.039656768798828124, 0.03967846298217773, 0.039740959167480466, 0.03973782348632812, 0.03990518569946289, 0.0398359375, 0.0397209587097168, 0.03986953735351562, 0.039900062561035156, 0.03996057510375976, 0.039923713684082034, 0.03993945693969726, 0.040032001495361326, 0.039911712646484375, 0.040083168029785156, 0.0400495376586914, 0.04008265686035156, 0.040264480590820315, 0.040117599487304687, 0.040042686462402347, 0.04002825546264648, 0.04014688110351562, 0.040247230529785155, 0.04018227386474609, 0.04013875198364258, 0.04009574508666992, 0.040725887298583986, 0.040456832885742186, 0.040269824981689455, 0.040114177703857425, 0.04031206512451172, 0.040207103729248045]",tokens/s,25.185453337009456,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Qwen/Qwen-72B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Qwen/Qwen-72B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,11038.990336,14897.053696,0.0,14501.80608,13634.065408,s,1,18.719181640625,18.719181640625,0.0,18.719181640625,18.719181640625,18.719181640625,18.719181640625,[18.719181640625],,kWh,0.00034198201161249244,3.7715848971386735e-05,0.00015437345683199516,0.0005340713174158743,,MB,2041.290752,14911.73376,0.0,14501.80608,12898.96192,s,10,6.8069766235351565,0.6806976623535157,0.00043245650314753796,0.6807008361816407,0.6812607788085937,0.6812755004882812,0.6812872778320312,"[0.679794677734375, 0.6805288696289062, 0.6803720703125, 0.6804213256835937, 0.6808062133789062, 0.680595458984375, 0.6810659790039062, 0.6808442993164062, 0.6812575073242187, 0.6812902221679688]",tokens/s,376.08473505679257,kWh,1.9883856241666405e-05,2.1928306243289064e-06,1.3160492009867131e-05,3.523717887586244e-05,tokens/kWh,7265053.791674584,MB,2047.729664,14911.73376,0.0,14501.80608,13241.259008,s,10,39.032643798828126,3.9032643798828124,0.004029642901741416,3.9040800781250002,3.9065517822265625,3.9080163696289065,3.909188039550781,"[3.9038291015625, 3.90353759765625, 3.9043310546875, 3.90595458984375, 3.897573974609375, 3.89501171875, 3.9050361328125, 3.90948095703125, 3.906226318359375, 3.901662353515625]",tokens/s,16.140336361712563,kWh,0.00011425802097208299,1.2603256586144216e-05,7.603860712713239e-05,0.00020289988468535962,tokens/kWh,310497.9586247434,,s,630,39.02842597198486,0.061949882495214076,0.0004613444707387764,0.06196040153503418,0.06255189361572266,0.06267083473205566,0.06304915794372558,"[0.06164121627807617, 0.061228641510009764, 0.06112857437133789, 0.061268447875976566, 0.06133059310913086, 0.061350814819335936, 0.06136422348022461, 0.06142566299438477, 0.06140927886962891, 0.06138265609741211, 0.06134783935546875, 0.061345470428466796, 0.0614749755859375, 0.061418750762939456, 0.06154089736938476, 0.06231897735595703, 0.06151168060302734, 0.06147686386108398, 0.06159564971923828, 0.06175948715209961, 0.06155878448486328, 0.06156492614746094, 0.06227763366699219, 0.06169116973876953, 0.061587936401367185, 0.062097663879394534, 0.062137535095214844, 0.061911712646484374, 0.06161219024658203, 0.0615997428894043, 0.06200320053100586, 0.06176563262939453, 0.061615966796875, 0.06203612899780273, 0.06177724838256836, 0.06220383834838867, 0.06191286468505859, 0.062277793884277344, 0.061907070159912106, 0.06220457458496094, 0.061908447265625, 0.062230751037597655, 0.062292289733886716, 0.062271488189697265, 0.06204415893554688, 0.06237900924682617, 0.062051326751708984, 0.06346137619018555, 0.06247407913208008, 0.06242319869995117, 0.06216268920898437, 0.0625624008178711, 0.06241043090820313, 0.06240668869018555, 0.06226374435424804, 0.06232876968383789, 0.06249440002441406, 0.06299481582641601, 0.06236774444580078, 0.06259846496582032, 0.06272684860229492, 0.06275481414794921, 0.06271590423583985, 0.06171187210083008, 0.061659168243408204, 0.06137699127197266, 0.06139875030517578, 0.061394527435302736, 0.0614714241027832, 0.06140240097045899, 0.06153472137451172, 0.061458656311035156, 0.06145228958129883, 0.06138665771484375, 0.061962337493896485, 0.06156902313232422, 0.06266239929199219, 0.06156723022460937, 0.06157497787475586, 0.06143561553955078, 0.061438430786132814, 0.06146047973632812, 0.061676673889160157, 0.06139788818359375, 0.061484416961669924, 0.06150377655029297, 0.061473121643066404, 0.06148201751708984, 0.06150179290771484, 0.061684223175048826, 0.061843582153320316, 0.06161203384399414, 0.062125343322753906, 0.06183737564086914, 0.06204796981811524, 0.06195619201660156, 0.06192623901367188, 0.06198476791381836, 0.0620206413269043, 0.061856735229492185, 0.061882015228271484, 0.06214009475708008, 0.062115966796875, 0.062113311767578124, 0.06231552124023437, 0.062033920288085936, 0.062273536682128906, 0.06227558517456055, 0.06211993789672852, 0.06234726333618164, 0.06253104019165039, 0.06233961486816406, 0.062058494567871096, 0.06237593460083008, 0.062498622894287106, 0.06241660690307617, 0.062351295471191406, 0.06242899322509766, 0.06247625732421875, 0.06254463958740235, 0.06306611251831054, 0.06240777587890625, 0.06300764846801758, 0.06256553649902344, 0.06254678344726562, 0.06252339172363282, 0.06167955017089844, 0.0612589111328125, 0.06110275268554687, 0.0612022705078125, 0.06126217651367188, 0.06138019180297852, 0.06142323303222656, 0.061394878387451175, 0.06137737655639648, 0.06136217498779297, 0.06141484832763672, 0.06135804748535156, 0.061373023986816405, 0.06151987075805664, 0.061965503692626954, 0.061700927734375, 0.06151903915405273, 0.06150044631958008, 0.06198659133911133, 0.06177587127685547, 0.06135193634033203, 0.06161612701416016, 0.06193561553955078, 0.06171990585327149, 0.06161270523071289, 0.06199705505371094, 0.061712383270263675, 0.06179008102416992, 0.06375164794921875, 0.06231241607666015, 0.06183391952514648, 0.0625541114807129, 0.06218902587890625, 0.06183785629272461, 0.062252159118652346, 0.06281619262695312, 0.0618004150390625, 0.06212499237060547, 0.06205379104614258, 0.06216767883300781, 0.061768768310546875, 0.06212704086303711, 0.061975582122802735, 0.06207961654663086, 0.06178441619873047, 0.06226124954223633, 0.061871841430664064, 0.0621632308959961, 0.062250015258789065, 0.062072864532470705, 0.06227840042114258, 0.06210579299926758, 0.0620052490234375, 0.06248992156982422, 0.062427745819091794, 0.06239446258544922, 0.06258659362792969, 0.06242947387695313, 0.06259276962280273, 0.06262400054931641, 0.06254182434082031, 0.06344704055786132, 0.06260940933227539, 0.062093246459960935, 0.06120249557495117, 0.061050880432128904, 0.06111804962158203, 0.061499393463134766, 0.061515296936035156, 0.06122304153442383, 0.06132524871826172, 0.0614337272644043, 0.06138291168212891, 0.06142022323608398, 0.061513729095458984, 0.06145833587646484, 0.06189628982543945, 0.061727104187011717, 0.061610111236572264, 0.06158950424194336, 0.06209651184082031, 0.06155763244628906, 0.061663230895996096, 0.06200870513916015, 0.061704639434814454, 0.06162451171875, 0.06205644989013672, 0.061652992248535154, 0.06168166351318359, 0.062048030853271485, 0.061763809204101565, 0.06174310302734375, 0.062183006286621094, 0.06175507354736328, 0.06209798431396484, 0.06192697525024414, 0.062032127380371095, 0.061799999237060546, 0.062151199340820314, 0.06181033706665039, 0.06224137496948242, 0.0619417610168457, 0.06217046356201172, 0.061884769439697264, 0.062417217254638675, 0.06234931182861328, 0.062149887084960935, 0.062128894805908205, 0.06247174453735352, 0.06288550567626953, 0.06251334381103515, 0.06245235061645508, 0.06238611221313477, 0.06260079956054687, 0.0622305908203125, 0.06246441650390625, 0.06286540985107422, 0.06234112167358399, 0.062459903717041014, 0.062394367218017575, 0.06234726333618164, 0.06237593460083008, 0.06260531234741211, 0.06268518447875976, 0.06313907241821289, 0.0625792007446289, 0.062005760192871094, 0.06126998519897461, 0.06124137496948242, 0.061241344451904295, 0.06124534225463867, 0.06129878234863281, 0.06138470458984375, 0.06138675308227539, 0.06136217498779297, 0.06146630477905273, 0.0613768310546875, 0.06132326507568359, 0.061341697692871094, 0.06136972808837891, 0.06123788833618164, 0.061439998626708986, 0.06135603332519531, 0.06132227325439453, 0.061328254699707034, 0.06140528106689453, 0.06166470336914062, 0.06177644729614258, 0.061628414154052735, 0.06161993789672852, 0.061534496307373045, 0.06219161605834961, 0.06172671890258789, 0.06164889526367188, 0.06191439819335937, 0.061700607299804686, 0.06169007873535156, 0.06221414566040039, 0.06174515151977539, 0.06221619033813477, 0.061722496032714846, 0.062093280792236326, 0.06171836853027344, 0.06187654495239258, 0.061900798797607424, 0.061810688018798826, 0.062058494567871096, 0.062189567565917966, 0.06227872085571289, 0.062316574096679685, 0.06190377426147461, 0.06230400085449219, 0.06201536178588867, 0.06210188674926758, 0.06249625778198242, 0.06273593521118163, 0.06235232162475586, 0.06217728042602539, 0.06221372985839844, 0.06230873489379883, 0.062133663177490236, 0.06238393783569336, 0.06233580780029297, 0.06259430313110352, 0.06213094329833985, 0.06245171356201172, 0.06241401672363281, 0.06291881561279297, 0.0626080322265625, 0.061744640350341794, 0.06199347305297852, 0.061109729766845707, 0.061319713592529294, 0.061337535858154296, 0.0614769287109375, 0.06126979064941406, 0.061319198608398434, 0.06138054275512695, 0.06140047836303711, 0.061502304077148434, 0.06133555221557617, 0.0613449592590332, 0.06130352020263672, 0.06129056167602539, 0.06148508834838867, 0.061396991729736325, 0.06141299057006836, 0.06132364654541016, 0.061247390747070314, 0.061306304931640625, 0.061319839477539065, 0.06142771148681641, 0.061402206420898435, 0.061393825531005856, 0.06164070510864258, 0.06184755325317383, 0.061841407775878904, 0.06160351943969727, 0.061937984466552735, 0.061712383270263675, 0.06173081588745117, 0.06210275268554687, 0.06199990463256836, 0.06192287826538086, 0.06177632141113281, 0.061834911346435546, 0.06197078323364258, 0.06187744140625, 0.061917057037353514, 0.061877185821533204, 0.062091262817382815, 0.061709758758544925, 0.06214713668823242, 0.06202163314819336, 0.062183425903320315, 0.06200668716430664, 0.06216960144042969, 0.06216508865356445, 0.06214441680908203, 0.062206047058105465, 0.06204156875610352, 0.06217782211303711, 0.06227558517456055, 0.062061920166015624, 0.06228854370117187, 0.06247219085693359, 0.06276710510253906, 0.06257664108276367, 0.06266675186157226, 0.06260326385498047, 0.06267494583129883, 0.06271385574340821, 0.06182291030883789, 0.06166944122314453, 0.06127542495727539, 0.06126793670654297, 0.06142950439453125, 0.06137942504882812, 0.061372577667236326, 0.061338817596435544, 0.06134460830688476, 0.06133308792114258, 0.06160588836669922, 0.061386783599853514, 0.06151129531860351, 0.0620079345703125, 0.06158755111694336, 0.06152185440063476, 0.0614986572265625, 0.0614879035949707, 0.06137187194824219, 0.061313568115234376, 0.06127123260498047, 0.06143033599853515, 0.06147097778320312, 0.06183321762084961, 0.061685760498046874, 0.061863296508789065, 0.06191904067993164, 0.06177260971069336, 0.061685760498046874, 0.06208230209350586, 0.06198300933837891, 0.062134143829345706, 0.06179900741577148, 0.062311710357666014, 0.06231523132324219, 0.06195404815673828, 0.062123809814453125, 0.061956161499023436, 0.06201769638061523, 0.06203596878051758, 0.062189567565917966, 0.062494720458984375, 0.062066078186035156, 0.062247520446777345, 0.06227763366699219, 0.062067966461181644, 0.06225382232666016, 0.06248038482666016, 0.06237519836425781, 0.062232929229736327, 0.06232307052612305, 0.06251059341430663, 0.062400894165039064, 0.06280121612548828, 0.06253855895996094, 0.06307430267333984, 0.0626954231262207, 0.06264012908935547, 0.06260476684570312, 0.06265705490112304, 0.06273433685302734, 0.0628545265197754, 0.06284966278076172, 0.061902687072753905, 0.06136182403564453, 0.06126847839355469, 0.061245281219482424, 0.06136848068237305, 0.06156492614746094, 0.061358081817626954, 0.06145843124389649, 0.06131625747680664, 0.06138700866699219, 0.06147747039794922, 0.06194790267944336, 0.06158659362792969, 0.061528926849365236, 0.06151331329345703, 0.06317612838745117, 0.06178649520874024, 0.061612545013427736, 0.061575233459472654, 0.06204383850097656, 0.06169356918334961, 0.061973217010498044, 0.061832416534423826, 0.061819679260253904, 0.06217036819458008, 0.06205107116699219, 0.06205641555786133, 0.06187187194824219, 0.061873695373535154, 0.0618073616027832, 0.06218105697631836, 0.06195846557617188, 0.06227763366699219, 0.06221004867553711, 0.06188598251342774, 0.06194019317626953, 0.06204006576538086, 0.06224895858764649, 0.06193766403198242, 0.06234112167358399, 0.06207692718505859, 0.062332862854003905, 0.062324798583984375, 0.062132225036621094, 0.06211993789672852, 0.062443519592285154, 0.06242892837524414, 0.062207744598388674, 0.06227814483642578, 0.06260531234741211, 0.06248038482666016, 0.06268928146362304, 0.06250291061401367, 0.06233260726928711, 0.06222060775756836, 0.0625165138244629, 0.06264460754394531, 0.06264771270751954, 0.06268934249877929, 0.06267744064331054, 0.06265622329711915, 0.06269001770019532, 0.06264575958251953, 0.062007038116455075, 0.06132592010498047, 0.061243392944335937, 0.06121001434326172, 0.06126038360595703, 0.06147481536865235, 0.06141952133178711, 0.06134566497802734, 0.061429088592529296, 0.06138140869140625, 0.061400672912597654, 0.061426078796386716, 0.06160179138183594, 0.061781600952148436, 0.06168617630004883, 0.06151168060302734, 0.06164070510864258, 0.06188851165771484, 0.06171852874755859, 0.06154966354370117, 0.061658016204833986, 0.06186371231079101, 0.061591777801513675, 0.061687328338623046, 0.06200368118286133, 0.06178201675415039, 0.062459903717041014, 0.06192281723022461, 0.06195846557617188, 0.06178153610229492, 0.062175457000732424, 0.061980384826660156, 0.0620748176574707, 0.06201744079589844, 0.06210649490356445, 0.06201948928833008, 0.06217443084716797, 0.06236249542236328, 0.06194790267944336, 0.062306304931640626, 0.06215423965454102, 0.06207913589477539, 0.06234502410888672, 0.06249654388427734, 0.06210604858398437, 0.06235372924804688, 0.062429183959960936, 0.062494720458984375, 0.062470142364501956, 0.062266590118408204, 0.062322689056396485, 0.06249756622314453, 0.06245548629760742, 0.062324703216552736, 0.0622655029296875, 0.0625802879333496, 0.06268377685546875, 0.06259267044067383, 0.06265276718139648, 0.06257868957519531, 0.062484161376953125, 0.06261785507202149, 0.06246134567260742, 0.06177382278442383, 0.062104766845703124, 0.06108038330078125, 0.06122038269042969, 0.06127254486083984, 0.061415424346923826, 0.06133484649658203, 0.06130963134765625, 0.061396991729736325, 0.06135910415649414, 0.0613135986328125, 0.061393344879150394, 0.06142566299438477, 0.061548545837402345, 0.061529502868652344, 0.062147167205810545, 0.0616399040222168, 0.06136502456665039, 0.061456382751464846, 0.061841407775878904, 0.061679615020751956, 0.06170214462280273, 0.06160534286499023, 0.062036510467529296, 0.06174310302734375, 0.06171833419799805, 0.06193990325927735, 0.06172051239013672, 0.061558849334716795, 0.06267417526245117, 0.06216780853271484, 0.06209516906738281, 0.06208262252807617, 0.06203251266479492, 0.061966335296630856, 0.06191286468505859, 0.061973953247070314, 0.06200713729858399, 0.06198064041137695, 0.061899742126464846, 0.06225884628295898, 0.0623271369934082, 0.06205974578857422, 0.06230643081665039, 0.06189123153686524, 0.06219571304321289, 0.06230796813964844, 0.06200735855102539, 0.06235168075561524, 0.06231654357910156, 0.06202777481079102, 0.06237139129638672, 0.06225145721435547, 0.062042110443115236, 0.0623636474609375, 0.06239846420288086, 0.062461952209472656, 0.062390239715576175, 0.062351390838623046, 0.062473217010498044, 0.0625301742553711, 0.06266073608398437, 0.0625516471862793]",tokens/s,16.14208065813934,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.747328,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,3552.100352,5264.83456,0.0,4869.586944,4520.068608,s,1,11.08853125,11.08853125,0.0,11.08853125,11.08853125,11.08853125,11.08853125,[11.08853125],,kWh,0.00012195984266666641,1.3445429172698094e-05,5.28433756079999e-05,0.0001882486474473644,,MB,1489.997824,5298.388992,0.0,4888.461312,4194.016256,s,10,1.8471275177001956,0.1847127517700195,0.0002522035746094528,0.18474337768554688,0.18496663665771484,0.18501363754272462,0.1850512382507324,"[0.1843956756591797, 0.18467015075683593, 0.18439353942871095, 0.18472857666015624, 0.18491842651367188, 0.18506063842773438, 0.1847581787109375, 0.18495619201660157, 0.18431417846679687, 0.1849319610595703]",tokens/s,1385.935716656629,kWh,5.451094363425919e-06,6.011560006089856e-07,3.6200286161851924e-06,9.672278980220097e-06,tokens/kWh,26467392.07207758,MB,1494.26176,5306.7776,0.0,4896.84992,4194.018816,s,10,19.221016967773437,1.9221016967773437,0.009146835434526338,1.919507080078125,1.9376085083007812,1.9383769470214842,1.9389916979980468,"[1.9136322021484375, 1.9104893798828124, 1.9193065185546876, 1.9391453857421874, 1.9268912353515626, 1.9148179931640625, 1.9212440185546875, 1.9197076416015626, 1.937437744140625, 1.9183448486328125]",tokens/s,32.77662160416787,kWh,5.602947284699082e-05,6.179889168601416e-06,3.400133481421482e-05,9.621069682980705e-05,tokens/kWh,654812.8438508717,,s,630,19.21847046661375,0.030505508677164714,0.0005148219707423202,0.030384927749633788,0.030966467475891112,0.03126797103881836,0.03259703628540041,"[0.030611455917358397, 0.0304967041015625, 0.03031612777709961, 0.03030825614929199, 0.030020160675048826, 0.03007916831970215, 0.02998806381225586, 0.03006729507446289, 0.030107648849487304, 0.03014860725402832, 0.03042099189758301, 0.03059926414489746, 0.03043315124511719, 0.030795808792114257, 0.03019161605834961, 0.029981887817382813, 0.03011257553100586, 0.03006800079345703, 0.030456544876098633, 0.030057823181152344, 0.030179967880249025, 0.03121148872375488, 0.030181087493896485, 0.030123680114746094, 0.03037798309326172, 0.030918527603149414, 0.030266048431396485, 0.03034441566467285, 0.03014748764038086, 0.03022233581542969, 0.03014816093444824, 0.030793407440185546, 0.03023641586303711, 0.030313472747802734, 0.03048979187011719, 0.0304167366027832, 0.030376352310180665, 0.030234367370605468, 0.03030054473876953, 0.030327232360839843, 0.030291584014892577, 0.03019523239135742, 0.030243679046630858, 0.030310144424438478, 0.03008505630493164, 0.030300384521484376, 0.030256288528442383, 0.0303175048828125, 0.03044528007507324, 0.030292255401611328, 0.03018547248840332, 0.03046553611755371, 0.030894559860229494, 0.03169833564758301, 0.03156156730651855, 0.030501535415649414, 0.03039580726623535, 0.03025302314758301, 0.03061631965637207, 0.030244287490844728, 0.030304832458496092, 0.030232576370239257, 0.030437376022338865, 0.03140604782104492, 0.030475711822509764, 0.03024947166442871, 0.030238624572753905, 0.03006278419494629, 0.030111167907714845, 0.030222623825073243, 0.030118207931518554, 0.030168895721435548, 0.0301628475189209, 0.030060800552368164, 0.03014656066894531, 0.03018511962890625, 0.030277023315429686, 0.030139520645141603, 0.030576448440551757, 0.030291807174682616, 0.031021120071411133, 0.030328832626342773, 0.03016713523864746, 0.030324703216552736, 0.030236703872680664, 0.030521343231201172, 0.03006185531616211, 0.03010223960876465, 0.03005571174621582, 0.03045449638366699, 0.030038015365600586, 0.030361600875854492, 0.030174463272094727, 0.03027587127685547, 0.03036412811279297, 0.03016089630126953, 0.03032678413391113, 0.030220064163208007, 0.03021436882019043, 0.030103296279907227, 0.030090911865234375, 0.030140960693359375, 0.03009440040588379, 0.030193824768066407, 0.030183744430541993, 0.030509599685668944, 0.030469120025634764, 0.030327808380126952, 0.03028540802001953, 0.030431455612182617, 0.030309919357299805, 0.030304927825927735, 0.030203903198242187, 0.0317130241394043, 0.03052569580078125, 0.030285663604736328, 0.03050511932373047, 0.030242687225341798, 0.030770816802978516, 0.030367584228515626, 0.03044576072692871, 0.030266847610473633, 0.030605663299560548, 0.03019843292236328, 0.030209344863891603, 0.030167743682861327, 0.030597343444824218, 0.030380319595336915, 0.03059542465209961, 0.0301844482421875, 0.030128448486328126, 0.030419647216796877, 0.030078975677490235, 0.030016992568969728, 0.030314687728881837, 0.030212448120117186, 0.03016089630126953, 0.03022371292114258, 0.030679712295532225, 0.030266687393188475, 0.031111871719360352, 0.03082854461669922, 0.030248960494995116, 0.030345216751098632, 0.030269439697265626, 0.03068025588989258, 0.030255935668945313, 0.030277631759643556, 0.030208000183105467, 0.030242816925048828, 0.030633983612060548, 0.029933536529541015, 0.030087200164794922, 0.03015872001647949, 0.0317664966583252, 0.03291126251220703, 0.03051241683959961, 0.030219232559204102, 0.030239744186401366, 0.030258176803588867, 0.030234624862670898, 0.030229888916015624, 0.030140863418579102, 0.03004419136047363, 0.029991071701049806, 0.030099456787109374, 0.030066495895385743, 0.030138559341430664, 0.030103424072265624, 0.03129766464233399, 0.03032281684875488, 0.030350784301757812, 0.03137750434875488, 0.031072608947753905, 0.031014495849609375, 0.030578432083129884, 0.03056889533996582, 0.030410144805908205, 0.03062380790710449, 0.030356224060058595, 0.03048963165283203, 0.030395456314086914, 0.030265247344970703, 0.030717376708984376, 0.030202432632446288, 0.030410144805908205, 0.03052729606628418, 0.03093788719177246, 0.03063382339477539, 0.03118275260925293, 0.03058492851257324, 0.030607423782348632, 0.03052739143371582, 0.030689311981201173, 0.030881919860839845, 0.030838655471801757, 0.030696767807006836, 0.03129206466674805, 0.030791711807250977, 0.030519296646118164, 0.0315773754119873, 0.030790367126464845, 0.03060105514526367, 0.030642175674438478, 0.031178911209106444, 0.03192422485351563, 0.03083395195007324, 0.030804704666137696, 0.030619199752807618, 0.030751104354858397, 0.03188115119934082, 0.03060736083984375, 0.03053785514831543, 0.03124224090576172, 0.030826496124267577, 0.030605600357055663, 0.03043008041381836, 0.03022729682922363, 0.030490591049194337, 0.030468128204345704, 0.03041187286376953, 0.030379104614257812, 0.030431039810180666, 0.03054591941833496, 0.030697471618652345, 0.030644224166870116, 0.03058892822265625, 0.03056844711303711, 0.030535680770874023, 0.030697471618652345, 0.030904031753540038, 0.030615840911865234, 0.03139750480651855, 0.03066713523864746, 0.030514528274536133, 0.03111737632751465, 0.03050761604309082, 0.031086591720581053, 0.030651424407958986, 0.030878688812255858, 0.030686784744262695, 0.030537216186523438, 0.030923263549804687, 0.030676511764526366, 0.0346899528503418, 0.030697471618652345, 0.03087273597717285, 0.030311264038085938, 0.030062559127807618, 0.029908544540405275, 0.02998524856567383, 0.03007014465332031, 0.031105152130126955, 0.03035158348083496, 0.030337087631225584, 0.03022198486328125, 0.030728256225585938, 0.030312448501586913, 0.030133855819702147, 0.03121308708190918, 0.030201919555664064, 0.0301279354095459, 0.030178207397460938, 0.03059107208251953, 0.03018547248840332, 0.03327590560913086, 0.0354029426574707, 0.030475072860717774, 0.030375999450683595, 0.03066873550415039, 0.030102880477905273, 0.03002169609069824, 0.03034720039367676, 0.03043600082397461, 0.030636032104492186, 0.030402559280395508, 0.03019161605834961, 0.030438783645629883, 0.030186111450195313, 0.030207103729248046, 0.030231039047241212, 0.030253440856933593, 0.030195520401000975, 0.030369632720947264, 0.03022662353515625, 0.030741695404052735, 0.030847967147827147, 0.031092735290527345, 0.031047679901123046, 0.030496768951416016, 0.030525440216064452, 0.03132803153991699, 0.0305948486328125, 0.030443199157714845, 0.030224992752075196, 0.030177440643310547, 0.030492767333984375, 0.03016387176513672, 0.030243839263916016, 0.030523168563842774, 0.030275808334350587, 0.030661983489990233, 0.030265567779541015, 0.03020025634765625, 0.030352415084838866, 0.030823328018188476, 0.030553632736206055, 0.030460447311401368, 0.03033888053894043, 0.03011599922180176, 0.030244319915771485, 0.030324384689331053, 0.03070044708251953, 0.031247840881347657, 0.0308897590637207, 0.03153033638000488, 0.03061222457885742, 0.03055548858642578, 0.031179328918457032, 0.0301976318359375, 0.03037811279296875, 0.03037705612182617, 0.03068156814575195, 0.030427583694458006, 0.030507104873657227, 0.030369440078735353, 0.030424575805664062, 0.03028438377380371, 0.03105299186706543, 0.030401248931884766, 0.030175487518310548, 0.02999817657470703, 0.03037648010253906, 0.030269792556762695, 0.030133920669555662, 0.030085216522216796, 0.0300097599029541, 0.02999692726135254, 0.030054399490356445, 0.030224384307861327, 0.030529535293579102, 0.03021004867553711, 0.0300948486328125, 0.0299967041015625, 0.030170175552368166, 0.030283008575439453, 0.03041539192199707, 0.030515199661254884, 0.030770591735839844, 0.030912351608276368, 0.03222195053100586, 0.03142422485351563, 0.03087286376953125, 0.030434303283691407, 0.030088384628295897, 0.030649375915527344, 0.030649343490600587, 0.030597984313964845, 0.03044550323486328, 0.030382080078125, 0.030166463851928713, 0.03075129508972168, 0.030023263931274413, 0.030110111236572267, 0.02999603271484375, 0.030320928573608397, 0.0302653751373291, 0.030071487426757814, 0.02999091148376465, 0.02999888038635254, 0.030226367950439453, 0.03009564781188965, 0.030091232299804687, 0.030115583419799804, 0.029980960845947265, 0.030042015075683593, 0.030230079650878906, 0.03013644790649414, 0.03187132835388184, 0.03054739189147949, 0.030852735519409178, 0.030753471374511718, 0.03092710494995117, 0.030672895431518556, 0.03037183952331543, 0.030285823822021486, 0.03029747200012207, 0.0302139835357666, 0.03052796745300293, 0.030175552368164063, 0.030325983047485353, 0.030103935241699218, 0.03027395248413086, 0.030060543060302734, 0.03061555290222168, 0.030087167739868165, 0.03005232048034668, 0.030088512420654297, 0.030175968170166014, 0.03040870475769043, 0.03060736083984375, 0.030324735641479493, 0.030318208694458008, 0.030363872528076173, 0.030349472045898437, 0.030619327545166015, 0.030816064834594727, 0.03069958305358887, 0.031072608947753905, 0.030594175338745117, 0.0304486083984375, 0.03116009521484375, 0.030369279861450195, 0.03093987274169922, 0.03052470397949219, 0.030745311737060545, 0.0303636474609375, 0.0305930233001709, 0.030492671966552733, 0.030668800354003906, 0.030799680709838868, 0.030869056701660156, 0.030283872604370116, 0.030300703048706055, 0.03027507209777832, 0.030757375717163086, 0.030457855224609375, 0.030432479858398438, 0.030546079635620116, 0.030389184951782226, 0.03034819221496582, 0.03034601593017578, 0.030511104583740234, 0.030489952087402343, 0.03041961669921875, 0.030304256439208983, 0.030320640563964843, 0.03027974319458008, 0.03046803283691406, 0.03036787223815918, 0.030250591278076173, 0.031180320739746095, 0.031064544677734375, 0.0312740478515625, 0.03057561683654785, 0.03028540802001953, 0.030256959915161134, 0.030503456115722655, 0.029980640411376953, 0.030416608810424805, 0.03062175941467285, 0.030799615859985353, 0.03031068801879883, 0.0303635196685791, 0.030277984619140625, 0.03021414375305176, 0.030228479385375977, 0.030256959915161134, 0.030097440719604494, 0.030074911117553713, 0.029995231628417968, 0.030224288940429687, 0.03016499137878418, 0.03010291290283203, 0.030349952697753906, 0.030225439071655272, 0.032750240325927736, 0.030828672409057616, 0.030382303237915038, 0.03193795204162598, 0.031382080078125, 0.030369792938232422, 0.030279680252075194, 0.03035308837890625, 0.030306623458862304, 0.03018288040161133, 0.0312956485748291, 0.030235008239746095, 0.030103551864624024, 0.030146175384521485, 0.03048838424682617, 0.030067520141601564, 0.03021798324584961, 0.03007814407348633, 0.03024675178527832, 0.030100448608398438, 0.03024508857727051, 0.030223648071289064, 0.030235008239746095, 0.030326271057128908, 0.030328895568847655, 0.030364448547363282, 0.03023849678039551, 0.030316543579101563, 0.030277631759643556, 0.03018320083618164, 0.030655807495117187, 0.03055094337463379, 0.03061350440979004, 0.030510719299316407, 0.031060575485229492, 0.0304136962890625, 0.030366016387939454, 0.030956127166748046, 0.031328128814697265, 0.030971904754638672, 0.030769151687622072, 0.0314368953704834, 0.03103324890136719, 0.030501184463500978, 0.03054707145690918, 0.031260543823242185, 0.031103679656982422, 0.030983423233032225, 0.030567167282104492, 0.03056572723388672, 0.03061417579650879, 0.03046505546569824, 0.03080086326599121, 0.030818304061889647, 0.030668256759643554, 0.03075494384765625, 0.03051510429382324, 0.030677696228027344, 0.030580543518066407, 0.030556095123291015, 0.030853151321411133, 0.030576671600341797, 0.03041689682006836, 0.031282432556152345, 0.030485248565673827, 0.030951648712158202, 0.030715808868408204, 0.03064944076538086, 0.03096598434448242, 0.030384384155273437, 0.030718271255493163, 0.030664703369140626, 0.030631872177124025, 0.03115014457702637, 0.030367647171020508, 0.030263391494750977, 0.03472812652587891, 0.03085433578491211, 0.030767744064331054, 0.030790847778320314, 0.03068191909790039, 0.030478336334228515, 0.030332063674926756, 0.030435583114624024, 0.03021660804748535, 0.030455711364746094, 0.030801216125488282, 0.03039740753173828, 0.03041587257385254, 0.03041587257385254, 0.030411903381347655, 0.030601247787475586, 0.030695711135864258, 0.030782079696655272, 0.030457792282104493, 0.030520704269409178, 0.03143948745727539, 0.030482431411743165, 0.03033497619628906, 0.030459903717041017, 0.030648319244384766, 0.03125923156738281, 0.030444896697998047, 0.030462623596191406, 0.030713472366333008, 0.030172576904296877, 0.03026972770690918, 0.03018822479248047, 0.029996320724487303, 0.030165151596069337, 0.03036422348022461, 0.02992243194580078, 0.030058624267578125, 0.030214912414550783, 0.030038015365600586, 0.030147743225097657, 0.0304116153717041, 0.030117887496948242, 0.030084703445434572, 0.02998271942138672, 0.030419231414794922, 0.030262880325317383, 0.030174816131591797, 0.030970815658569337, 0.030072832107543947, 0.029845504760742186, 0.03015679931640625, 0.03033087921142578, 0.030048255920410157, 0.030242816925048828, 0.030380064010620118, 0.030223808288574218, 0.030192159652709962, 0.030849023818969725, 0.03017932891845703, 0.030269439697265626, 0.03038969612121582, 0.030593791961669923, 0.03027315139770508, 0.0333776969909668, 0.03040540885925293, 0.030117887496948242, 0.030264768600463867, 0.030300735473632812, 0.030349184036254882, 0.03065011215209961, 0.03065894317626953, 0.03038547134399414, 0.030784191131591795, 0.03055615997314453, 0.030610912322998046, 0.03055039978027344, 0.030518688201904298, 0.03083139228820801, 0.03067286491394043, 0.03058278465270996, 0.03106390380859375, 0.03065667152404785, 0.0305664005279541, 0.030693376541137695, 0.030803359985351563, 0.03039673614501953, 0.03093849563598633, 0.030460832595825195]",tokens/s,32.78096459832393,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,4929.241088,8001.61792,0.0,7606.370304,6988.678144,s,1,13.9430087890625,13.9430087890625,0.0,13.9430087890625,13.9430087890625,13.9430087890625,13.9430087890625,[13.9430087890625],,kWh,0.00018919454070411725,2.0861765625099773e-05,7.585894957595762e-05,0.00028591525590517463,,MB,2811.670528,8018.395136,0.0,7608.467456,6915.136,s,10,2.2374570465087893,0.22374570465087892,0.00037969995005639054,0.22365451049804688,0.22425113830566407,0.22430920104980467,0.2243556512451172,"[0.2242382354736328, 0.22316246032714843, 0.2236561279296875, 0.22365289306640626, 0.2232406463623047, 0.22382080078125, 0.22364569091796874, 0.22354595947265626, 0.22412696838378907, 0.2243672637939453]",tokens/s,1144.1560426800104,kWh,6.628699578146822e-06,7.308311389348398e-07,4.388984992666057e-06,1.174851570974772e-05,tokens/kWh,21789986.609763592,MB,2815.934464,8020.492288,0.0,7610.564608,6915.13856,s,10,21.531000488281247,2.153100048828125,0.007082121175120396,2.1534468994140625,2.160139599609375,2.16175107421875,2.16304025390625,"[2.163362548828125, 2.148183349609375, 2.159029052734375, 2.158015625, 2.152819580078125, 2.1382724609375, 2.146603515625, 2.150858642578125, 2.15407421875, 2.159781494140625]",tokens/s,29.260135883740855,kWh,6.264239718852457e-05,6.908405714485793e-06,4.09264957041343e-05,0.00011047729860714466,tokens/kWh,570252.9007703827,,s,630,21.528521755218502,0.03417225675431509,0.0004772735727576938,0.034099822998046876,0.034559771347045895,0.03492978286743164,0.036041361808776866,"[0.03500147247314453, 0.034740222930908206, 0.034495361328125, 0.03453286361694336, 0.034643905639648434, 0.03560095977783203, 0.0345456657409668, 0.0344535026550293, 0.03457843017578125, 0.03551027297973633, 0.03474227142333984, 0.034320384979248046, 0.03415225601196289, 0.03415059280395508, 0.034148353576660156, 0.0339021110534668, 0.03396860885620117, 0.034050048828125, 0.03395100784301758, 0.03398851013183594, 0.03423657608032227, 0.034116256713867185, 0.0344002571105957, 0.03446768188476562, 0.03504089736938477, 0.03408950424194336, 0.033906688690185545, 0.03392015838623047, 0.03406934356689453, 0.034244350433349606, 0.03400048065185547, 0.03381264114379883, 0.03404822540283203, 0.03400265502929688, 0.035770721435546875, 0.035071712493896484, 0.034488609313964844, 0.03390630340576172, 0.03434979248046875, 0.03430796813964844, 0.03406028747558594, 0.033996353149414064, 0.03391532897949219, 0.03397241592407227, 0.03383276748657227, 0.03497983932495117, 0.03401660919189453, 0.035269279479980466, 0.034062335968017575, 0.034078399658203126, 0.03434531021118164, 0.034009056091308595, 0.03379404830932617, 0.034915809631347654, 0.03383126449584961, 0.03385158538818359, 0.03407257461547852, 0.03450265502929688, 0.03432015991210938, 0.03445782470703125, 0.03434700775146484, 0.0343633918762207, 0.03436860656738281, 0.03498441696166992, 0.034446815490722656, 0.034339359283447266, 0.034041568756103514, 0.03405033493041992, 0.0340503044128418, 0.03397135925292969, 0.034079326629638675, 0.03425894546508789, 0.03415859222412109, 0.03398451232910156, 0.03394057464599609, 0.034184062957763675, 0.03442416000366211, 0.03383980941772461, 0.034200801849365234, 0.03394220733642578, 0.03419548797607422, 0.03410131072998047, 0.03403504180908203, 0.0338348159790039, 0.03388297653198242, 0.03376038360595703, 0.033737598419189455, 0.033899936676025394, 0.03386355209350586, 0.03382041549682617, 0.03406124877929687, 0.03406646347045898, 0.034009056091308595, 0.034449440002441406, 0.03379548645019531, 0.033823326110839845, 0.03385283279418945, 0.033804702758789065, 0.03389894485473633, 0.03419520187377929, 0.034369537353515625, 0.0342072639465332, 0.03445830535888672, 0.03427920150756836, 0.03439616012573242, 0.03544876861572266, 0.034280574798583985, 0.034076736450195315, 0.03388703918457031, 0.034181182861328124, 0.034673694610595704, 0.0341401596069336, 0.03388924789428711, 0.033701854705810545, 0.03373468780517578, 0.033841152191162106, 0.03382672119140625, 0.03381462478637695, 0.034342910766601564, 0.034455486297607425, 0.03456764984130859, 0.033964160919189454, 0.034286048889160155, 0.03388396835327148, 0.03368569564819336, 0.03358678436279297, 0.034888031005859375, 0.03437097549438477, 0.03468540954589844, 0.03486646270751953, 0.03432944107055664, 0.03436524963378906, 0.03421817779541016, 0.03419955062866211, 0.034078720092773435, 0.034178272247314456, 0.034337566375732424, 0.034301952362060545, 0.034307167053222655, 0.0343515510559082, 0.03446646499633789, 0.03416044616699219, 0.03413593673706055, 0.03396006393432617, 0.034041664123535154, 0.03444140625, 0.03431945419311523, 0.03423529434204101, 0.03400089645385742, 0.0342279052734375, 0.034294078826904294, 0.03449168014526367, 0.03436003112792969, 0.034418689727783204, 0.034560001373291016, 0.034364639282226564, 0.0344420166015625, 0.03475046539306641, 0.034240222930908205, 0.03397046279907227, 0.034078720092773435, 0.034179073333740234, 0.03495935821533203, 0.03406438446044922, 0.033927230834960936, 0.033810367584228514, 0.033716129302978515, 0.03370947265625, 0.0339152946472168, 0.03394793701171875, 0.03406438446044922, 0.03421731185913086, 0.03440822219848633, 0.03414863967895508, 0.034769439697265626, 0.03423206329345703, 0.0341558723449707, 0.034216926574707034, 0.03425075149536133, 0.03415363311767578, 0.03407740783691406, 0.03414028930664063, 0.0340184326171875, 0.0340447998046875, 0.03424367904663086, 0.03376220703125, 0.03385958480834961, 0.03414425659179687, 0.03622092819213867, 0.03504364776611328, 0.03420003128051758, 0.033901023864746097, 0.033963489532470706, 0.033786048889160154, 0.03412572860717773, 0.03389039993286133, 0.03376537704467773, 0.033906688690185545, 0.03408710479736328, 0.03403142547607422, 0.03420159912109375, 0.03418316650390625, 0.03428700637817383, 0.03429846572875977, 0.03431366348266601, 0.03441926574707031, 0.034369537353515625, 0.03412582397460937, 0.034097152709960936, 0.03508582305908203, 0.034211582183837894, 0.03406105422973633, 0.03389235305786133, 0.033971904754638675, 0.0340136947631836, 0.03394950485229492, 0.03412947082519531, 0.03392345428466797, 0.033942913055419924, 0.03416339111328125, 0.03429584121704102, 0.03426095962524414, 0.03397836685180664, 0.03372032165527344, 0.03406412887573242, 0.03381260681152344, 0.03470963287353516, 0.03421388626098633, 0.03429315185546875, 0.03431484985351563, 0.036357246398925784, 0.03464265441894531, 0.03457449722290039, 0.03446169662475586, 0.037515262603759765, 0.03416844940185547, 0.03435763168334961, 0.034086910247802735, 0.034427902221679685, 0.03390127944946289, 0.03387011337280273, 0.03370134353637695, 0.03377983856201172, 0.03403772735595703, 0.03399929428100586, 0.0341401596069336, 0.034236415863037106, 0.03421184158325195, 0.034402305603027344, 0.034146305084228515, 0.034375679016113284, 0.034342849731445316, 0.03502735900878906, 0.034437183380126954, 0.03430636978149414, 0.034014751434326175, 0.03372652816772461, 0.03375145721435547, 0.03371212768554688, 0.034014720916748044, 0.03373263931274414, 0.03448393630981445, 0.034369537353515625, 0.034187393188476564, 0.034374271392822266, 0.03439616012573242, 0.034288894653320315, 0.034366207122802736, 0.034313953399658204, 0.03460739135742188, 0.03433184051513672, 0.03528345489501953, 0.034472000122070315, 0.03425817489624024, 0.034025630950927734, 0.03423660659790039, 0.03455657577514649, 0.03448368072509766, 0.034719905853271484, 0.03409958267211914, 0.03410739135742188, 0.033906688690185545, 0.03398767852783203, 0.03391171264648438, 0.03385513687133789, 0.03404131317138672, 0.034356094360351565, 0.03441664123535156, 0.03420950317382813, 0.03394179153442383, 0.033972000122070314, 0.03383113479614258, 0.03378732681274414, 0.03394822311401367, 0.03376031875610352, 0.03403366470336914, 0.03424761581420899, 0.03373020935058594, 0.033882465362548825, 0.03378768157958984, 0.035227519989013675, 0.03499657440185547, 0.03404729461669922, 0.03385990524291992, 0.033806720733642576, 0.0338895378112793, 0.03372518539428711, 0.03401113510131836, 0.033810367584228514, 0.03377363204956055, 0.03367731094360352, 0.033691646575927735, 0.03363779067993164, 0.03615190505981445, 0.03398451232910156, 0.034047073364257815, 0.03386403274536133, 0.033813121795654294, 0.03367731094360352, 0.03372780990600586, 0.03376614379882813, 0.033724353790283206, 0.03383840179443359, 0.03355830383300781, 0.033632896423339845, 0.033794334411621094, 0.03381657409667969, 0.033642494201660156, 0.03395721435546875, 0.033712799072265626, 0.03359247970581054, 0.0340076789855957, 0.033587425231933594, 0.03365683364868164, 0.033742847442626955, 0.03360873413085937, 0.03368854522705078, 0.03375718307495117, 0.03377907180786133, 0.03376812744140625, 0.033796031951904296, 0.033775615692138675, 0.033898048400878907, 0.034046592712402346, 0.0339986572265625, 0.03414028930664063, 0.03432620620727539, 0.03448646545410156, 0.034252799987792966, 0.034432926177978516, 0.03434672164916992, 0.034344352722167966, 0.03414115142822265, 0.03386982345581055, 0.0343485107421875, 0.03441923141479492, 0.03411763381958008, 0.03445145416259766, 0.033889694213867186, 0.033899105072021485, 0.03377084732055664, 0.03390041732788086, 0.03396278381347656, 0.03388809585571289, 0.03457382583618164, 0.034226943969726566, 0.03392416000366211, 0.03410006332397461, 0.03378995132446289, 0.033807903289794924, 0.03390521621704102, 0.0339967041015625, 0.034453472137451174, 0.033823806762695315, 0.033627105712890626, 0.03379129409790039, 0.03394367980957031, 0.03382041549682617, 0.034459648132324217, 0.033964031219482424, 0.03420159912109375, 0.03379404830932617, 0.03373862457275391, 0.0337655029296875, 0.03371212768554688, 0.033661022186279296, 0.03368540954589844, 0.03369779205322266, 0.03369779205322266, 0.033739871978759765, 0.033719200134277344, 0.03377260971069336, 0.03382742309570313, 0.03375084686279297, 0.033985057830810544, 0.03391606521606445, 0.03368534469604492, 0.03368447875976562, 0.033791999816894534, 0.03392102432250976, 0.03358924865722656, 0.033617919921875, 0.033888256072998044, 0.03405228805541992, 0.03435295867919922, 0.03421500778198242, 0.03433295822143555, 0.03432511901855469, 0.03437363052368164, 0.03494623947143555, 0.0342413444519043, 0.03468492889404297, 0.03434883117675781, 0.0341874885559082, 0.03403148651123047, 0.033939422607421874, 0.03407900619506836, 0.034126815795898435, 0.03393119812011719, 0.03466918563842773, 0.03381283187866211, 0.034407615661621094, 0.03397100830078125, 0.03384320068359375, 0.034430431365966796, 0.03387801742553711, 0.03420828628540039, 0.034415935516357424, 0.034226303100585935, 0.034326431274414065, 0.0342570571899414, 0.034301471710205075, 0.03410646438598633, 0.03435951995849609, 0.03425798416137695, 0.03417103958129883, 0.034165184020996095, 0.0341319694519043, 0.034255870819091795, 0.03446470260620117, 0.034250816345214846, 0.03496575927734375, 0.034304542541503905, 0.034285568237304685, 0.03435737609863281, 0.03418918228149414, 0.03443711853027344, 0.03451903915405274, 0.034179073333740234, 0.034113536834716796, 0.03403571319580078, 0.034381855010986326, 0.034057247161865235, 0.03394451141357422, 0.03422412872314453, 0.034127872467041014, 0.03398012924194336, 0.03413020706176758, 0.03384729766845703, 0.033772865295410154, 0.03394582366943359, 0.03379248046875, 0.03385343933105469, 0.03395555114746094, 0.034206016540527344, 0.03440841674804687, 0.034106433868408205, 0.03422281646728516, 0.03376969528198242, 0.033710079193115236, 0.03370393753051758, 0.033767425537109375, 0.03422959899902344, 0.03371484756469727, 0.0337523193359375, 0.033837825775146484, 0.034840576171875, 0.03376128005981445, 0.03383222579956055, 0.03380831909179687, 0.03373299026489258, 0.033970592498779296, 0.03371212768554688, 0.03381398391723633, 0.033843742370605466, 0.0337448959350586, 0.03371212768554688, 0.03369331359863281, 0.03369375991821289, 0.033742591857910155, 0.03384121704101563, 0.03403417587280273, 0.033705440521240235, 0.033753631591796875, 0.03378790283203125, 0.033797824859619144, 0.033754463195800784, 0.036950782775878904, 0.03401087951660156, 0.03388227081298828, 0.03456777572631836, 0.03432089614868164, 0.03899824142456055, 0.03446099090576172, 0.035622337341308596, 0.03448070526123047, 0.034195457458496094, 0.03408486557006836, 0.033836673736572266, 0.033685504913330076, 0.0337694091796875, 0.033640830993652344, 0.03368966293334961, 0.03451859283447266, 0.03430643081665039, 0.0343573112487793, 0.034176063537597653, 0.034185600280761716, 0.034255424499511716, 0.034148353576660156, 0.034318336486816405, 0.034264225006103516, 0.03430486297607422, 0.03420569610595703, 0.033982463836669925, 0.033882110595703126, 0.03400028610229492, 0.034049697875976566, 0.034415550231933593, 0.034369281768798825, 0.03366323089599609, 0.03423564910888672, 0.033739521026611326, 0.03404390335083008, 0.033783809661865234, 0.03368489456176758, 0.03406857681274414, 0.03441692733764649, 0.034260734558105466, 0.03516668701171875, 0.03430809783935547, 0.0343337287902832, 0.03453148651123047, 0.034324382781982424, 0.034781631469726564, 0.03451862335205078, 0.03420659255981445, 0.03448937606811524, 0.03391177749633789, 0.03383091354370117, 0.034113536834716796, 0.03407807922363281, 0.03400563049316406, 0.03426128005981445, 0.03448336029052734, 0.03394822311401367, 0.03389235305786133, 0.03379609680175781, 0.033879966735839845, 0.03381375885009766, 0.03389523315429688, 0.03450991821289062, 0.03428643035888672, 0.0347300796508789, 0.03423139190673828, 0.03446467208862305, 0.0343633918762207, 0.03522608184814453, 0.034343006134033204, 0.03402751922607422, 0.033957023620605466, 0.03380828857421875, 0.03375545501708985, 0.034398849487304685, 0.03545439910888672, 0.03400761413574219, 0.03418931198120117, 0.03446755218505859, 0.03452342224121094, 0.034514942169189454, 0.035471359252929685, 0.034382911682128904, 0.036981697082519534, 0.034551807403564457, 0.034290878295898435, 0.03474214553833008, 0.03419750213623047, 0.034941215515136716, 0.03399481582641602, 0.03410800170898438, 0.03426115036010742, 0.03400688171386719, 0.03398025512695312, 0.033742401123046876, 0.03409196853637695, 0.0342457275390625, 0.03417555236816406, 0.03401232147216797, 0.034154975891113284, 0.034361217498779295, 0.03442124938964844, 0.0344637451171875, 0.034559745788574216, 0.03472000122070312, 0.034731552124023436, 0.03440035247802734, 0.03462115097045899, 0.034570911407470706, 0.03450806427001953, 0.0345115852355957, 0.034506752014160154, 0.034655681610107424, 0.03453190231323242, 0.03434451293945313, 0.03454000091552734, 0.033992992401123044, 0.03404915237426758, 0.033874496459960934, 0.03384729766845703, 0.03371417617797851, 0.03360492706298828, 0.03363827133178711, 0.03369043350219727, 0.0337367057800293, 0.03366041564941406, 0.03369830322265625, 0.03362406539916992, 0.03362198257446289, 0.033640480041503905, 0.03369779205322266]",tokens/s,29.263504812971576,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,881.225728,697.237504,0.0,301.989888,282.769408,s,1,8.115388671875,8.115388671875,0.0,8.115388671875,8.115388671875,8.115388671875,8.115388671875,[8.115388671875],,kWh,2.7295277566675697e-05,3.003633164850173e-06,8.516117923990096e-06,3.881502865551596e-05,,MB,1224.056832,751.763456,0.0,341.835776,318.94528,s,14,0.19344745635986327,0.013817675454275948,7.485959433415556e-05,0.013801631927490236,0.01389693422317505,0.013936592054367066,0.013989008321762085,"[0.013711968421936034, 0.013839743614196777, 0.013796352386474609, 0.013757760047912598, 0.01400211238861084, 0.013886719703674317, 0.013901311874389649, 0.013763296127319335, 0.013855520248413087, 0.01378662395477295, 0.01382975959777832, 0.013800127983093262, 0.013803135871887208, 0.013713024139404296]",tokens/s,18526.994706680533,kWh,4.08081155355449e-07,4.500398054212974e-08,2.2897026864617723e-07,6.82055404543756e-07,tokens/kWh,375336077.23736876,MB,1260.453888,776.92928,0.0,367.0016,318.94784,s,14,10.310911926269531,0.7364937090192523,0.00821429704017505,0.733315673828125,0.7502115112304688,0.7535501770019531,0.7550960217285156,"[0.731744384765625, 0.7309376831054688, 0.7525097045898438, 0.7390380249023437, 0.7554824829101563, 0.7297362060546875, 0.7448490600585937, 0.7279030151367187, 0.7310556640625, 0.7342277221679687, 0.7320957641601562, 0.73263671875, 0.73399462890625, 0.7347008666992187]",tokens/s,85.54044553061233,kWh,2.142347510089356e-05,2.3626636365616697e-06,8.067167320637933e-06,3.1853306058093164e-05,tokens/kWh,1977816.6789061823,,s,882,10.304133318901059,0.011682690837756306,0.00026965252532891844,0.011595471858978272,0.012006979274749755,0.012102902269363402,0.01250075856208801,"[0.011332127571105957, 0.011513216018676758, 0.011523903846740723, 0.011514687538146972, 0.011567104339599609, 0.011506976127624512, 0.011500255584716797, 0.0115032958984375, 0.01157145595550537, 0.0115, 0.011535967826843262, 0.011537504196166993, 0.01150044822692871, 0.011519519805908204, 0.0116843204498291, 0.011497119903564453, 0.01149955177307129, 0.011528512001037598, 0.011507712364196777, 0.011517951965332032, 0.011522047996520996, 0.011536383628845214, 0.011618304252624511, 0.011584639549255371, 0.011707263946533203, 0.011537440299987793, 0.011595840454101563, 0.011653375625610352, 0.011615232467651367, 0.011658559799194336, 0.01165552043914795, 0.011603872299194335, 0.011617440223693848, 0.011604096412658691, 0.011585856437683105, 0.011633152008056641, 0.011616095542907715, 0.011680031776428223, 0.011685759544372558, 0.011615551948547364, 0.012221119880676269, 0.012294143676757812, 0.011687935829162598, 0.0116627197265625, 0.011536416053771972, 0.011536928176879882, 0.011585408210754395, 0.011544768333435058, 0.011587583541870117, 0.011808095932006836, 0.01165385627746582, 0.011642399787902833, 0.011640895843505859, 0.011566911697387695, 0.011641375541687012, 0.011624544143676759, 0.011571104049682618, 0.011580960273742676, 0.01161411190032959, 0.011581727981567383, 0.011815199851989746, 0.011599871635437011, 0.011587648391723633, 0.011315327644348144, 0.011515968322753907, 0.011562463760375976, 0.01171078395843506, 0.011559200286865235, 0.011632351875305176, 0.011551039695739746, 0.011545696258544923, 0.011518176078796388, 0.01150819206237793, 0.011536288261413574, 0.01151961612701416, 0.011553152084350586, 0.011536479949951172, 0.011547967910766602, 0.011608384132385253, 0.011607647895812988, 0.011583935737609864, 0.011637248039245606, 0.01163424015045166, 0.011589216232299806, 0.01148953628540039, 0.011503968238830566, 0.011534336090087891, 0.011538432121276856, 0.011568927764892578, 0.0115, 0.011554559707641602, 0.011624287605285645, 0.011762911796569825, 0.012058688163757323, 0.011580479621887208, 0.01152835178375244, 0.011554143905639648, 0.01158176040649414, 0.011716608047485352, 0.011552032470703124, 0.011535072326660156, 0.01155840015411377, 0.011536288261413574, 0.011457344055175782, 0.011515456199645996, 0.011505887985229493, 0.011585536003112793, 0.011697440147399903, 0.01161308765411377, 0.011927359580993653, 0.011561152458190917, 0.011576319694519043, 0.011567935943603516, 0.01153228759765625, 0.01159718418121338, 0.011604000091552734, 0.011825887680053711, 0.0116778564453125, 0.011640095710754394, 0.011684288024902343, 0.011738880157470703, 0.011684160232543945, 0.011620287895202636, 0.01164083194732666, 0.011595168113708497, 0.011536992073059082, 0.01131503963470459, 0.011548831939697265, 0.011655232429504395, 0.01158950424194336, 0.011568544387817382, 0.011610783576965333, 0.011775296211242675, 0.01202451229095459, 0.01194598388671875, 0.012062399864196778, 0.012091103553771973, 0.012057151794433594, 0.012130335807800293, 0.01204428768157959, 0.01205008029937744, 0.012075584411621093, 0.012076160430908204, 0.012047136306762695, 0.012345120429992676, 0.012214367866516113, 0.012165472030639649, 0.012131999969482422, 0.012068863868713378, 0.012277728080749511, 0.012010720252990723, 0.012003840446472168, 0.012003775596618652, 0.012091360092163085, 0.012036224365234376, 0.012035136222839355, 0.012047072410583497, 0.012056032180786132, 0.01209603214263916, 0.012072352409362793, 0.012259936332702637, 0.012213600158691407, 0.012230751991271972, 0.012175200462341308, 0.012130144119262696, 0.01219264030456543, 0.012103263854980468, 0.012030367851257323, 0.01241702365875244, 0.011921759605407714, 0.011951552391052247, 0.011896608352661132, 0.011863840103149415, 0.01187446403503418, 0.011878911972045898, 0.011974656105041503, 0.011937791824340821, 0.011804672241210937, 0.01170956802368164, 0.011801471710205078, 0.011701536178588868, 0.011725407600402832, 0.011652544021606445, 0.011580096244812012, 0.01157027244567871, 0.011555744171142577, 0.011534144401550293, 0.0115033597946167, 0.011524543762207032, 0.011343199729919434, 0.011522591590881348, 0.012023584365844726, 0.011730112075805665, 0.011649087905883789, 0.011541248321533203, 0.011573247909545899, 0.011509759902954102, 0.011574655532836915, 0.01151363182067871, 0.0115, 0.011467136383056641, 0.011556832313537598, 0.01162393569946289, 0.012193408012390137, 0.01223737621307373, 0.011980192184448242, 0.012771552085876464, 0.011760095596313476, 0.01160217571258545, 0.011673407554626465, 0.011946175575256348, 0.012265472412109376, 0.01174726390838623, 0.01163987159729004, 0.01158358383178711, 0.011530207633972169, 0.011858880043029785, 0.011736767768859863, 0.01168569564819336, 0.01163868808746338, 0.011626239776611327, 0.011565919876098632, 0.011524191856384277, 0.01151369571685791, 0.011475071907043458, 0.011552096366882324, 0.011550944328308105, 0.01182755184173584, 0.011616288185119629, 0.011571328163146972, 0.01153990364074707, 0.011553215980529784, 0.011589728355407714, 0.011542431831359863, 0.011595775604248047, 0.011716608047485352, 0.011577343940734864, 0.011577695846557616, 0.011578335762023925, 0.011540767669677734, 0.011717023849487305, 0.012767168045043945, 0.012411231994628907, 0.011845343589782715, 0.011935744285583496, 0.01173468780517578, 0.01175369644165039, 0.011649151802062989, 0.011759360313415528, 0.011724672317504882, 0.011703968048095703, 0.01190112018585205, 0.011593376159667969, 0.011972288131713867, 0.01187724781036377, 0.011890175819396973, 0.012007328033447265, 0.012232447624206543, 0.012208767890930176, 0.012140768051147461, 0.011964735984802246, 0.012011072158813477, 0.011962431907653809, 0.01389475154876709, 0.012280799865722657, 0.011997183799743653, 0.012040191650390625, 0.012062687873840332, 0.012019424438476563, 0.012007583618164062, 0.012017312049865723, 0.011946463584899902, 0.011941184043884277, 0.01198703956604004, 0.012292736053466797, 0.0120381441116333, 0.012011520385742188, 0.012001279830932618, 0.012007424354553223, 0.012056063652038575, 0.012044960021972657, 0.01201360034942627, 0.011940863609313965, 0.01200111961364746, 0.011930368423461913, 0.012023039817810059, 0.011953120231628417, 0.011968159675598144, 0.011946335792541504, 0.011949888229370117, 0.011941056251525878, 0.011911520004272461, 0.011969344139099121, 0.01201750373840332, 0.012036447525024413, 0.012076607704162598, 0.012013471603393555, 0.012150976181030273, 0.012000864028930663, 0.012104096412658692, 0.011859968185424804, 0.011877663612365722, 0.011769760131835937, 0.011840255737304687, 0.01169651222229004, 0.01168992042541504, 0.011776864051818848, 0.01173580837249756, 0.011754816055297851, 0.011872320175170899, 0.011838560104370116, 0.011685888290405273, 0.011697855949401856, 0.011759615898132325, 0.011684096336364747, 0.01166207981109619, 0.012231936454772949, 0.012161952018737793, 0.01175551986694336, 0.01164083194732666, 0.011558912277221679, 0.011542528152465821, 0.011533408164978028, 0.011537311553955078, 0.011517663955688477, 0.011511520385742187, 0.01152467155456543, 0.011520000457763671, 0.011620415687561034, 0.011564448356628418, 0.011528736114501953, 0.011510880470275878, 0.01150864028930664, 0.011520000457763671, 0.011505120277404786, 0.011620927810668945, 0.011525888442993164, 0.011514080047607423, 0.01162649631500244, 0.011543711662292481, 0.01148316764831543, 0.011576416015625, 0.011519712448120117, 0.01151580810546875, 0.011475040435791015, 0.011517951965332032, 0.011468128204345703, 0.01154319953918457, 0.01158950424194336, 0.01152233600616455, 0.011503456115722656, 0.0115316801071167, 0.011618304252624511, 0.01157910442352295, 0.01156604766845703, 0.011505023956298828, 0.011503423690795898, 0.011542431831359863, 0.011625280380249023, 0.011480287551879883, 0.011540287971496583, 0.011539423942565918, 0.011518272399902343, 0.011556480407714844, 0.011644191741943359, 0.011641632080078125, 0.011630592346191406, 0.01161734390258789, 0.011641792297363281, 0.011567104339599609, 0.011521696090698242, 0.011560640335083007, 0.0115513916015625, 0.011511136054992675, 0.011535008430480958, 0.011539744377136231, 0.011581279754638673, 0.011590527534484864, 0.011395071983337402, 0.011618304252624511, 0.011712512016296387, 0.01175551986694336, 0.011811967849731445, 0.011758208274841309, 0.01182102394104004, 0.011841823577880859, 0.011727968215942382, 0.011727519989013672, 0.011652511596679687, 0.011747872352600098, 0.011711071968078614, 0.011814271926879883, 0.011823391914367676, 0.011880703926086426, 0.011900992393493652, 0.011986687660217285, 0.011876352310180664, 0.011781503677368164, 0.011715200424194336, 0.011671039581298828, 0.01183561611175537, 0.011622976303100586, 0.011896896362304687, 0.011615903854370117, 0.01163263988494873, 0.011630271911621094, 0.011764032363891602, 0.011763584136962891, 0.011780447959899902, 0.01176915168762207, 0.011719136238098145, 0.011647168159484864, 0.011626303672790527, 0.011620287895202636, 0.011577407836914062, 0.011646688461303711, 0.011890975952148438, 0.013569855690002441, 0.011784192085266113, 0.011737600326538086, 0.01165392017364502, 0.011698207855224609, 0.011684224128723145, 0.011962880134582519, 0.012072480201721192, 0.012175456047058105, 0.012153056144714356, 0.011876640319824218, 0.01184921646118164, 0.011894240379333496, 0.011815199851989746, 0.011923839569091797, 0.0117838716506958, 0.011772480010986328, 0.011993056297302246, 0.012031167984008788, 0.011957088470458984, 0.011843584060668945, 0.011685888290405273, 0.0120381441116333, 0.011640735626220703, 0.011487456321716308, 0.011500639915466309, 0.01153660774230957, 0.011543007850646972, 0.011582688331604004, 0.01153536033630371, 0.011565919876098632, 0.011690976142883301, 0.01158739185333252, 0.011526432037353516, 0.011648896217346192, 0.011470848083496094, 0.011530495643615723, 0.011460639953613282, 0.011527296066284179, 0.011508543968200684, 0.011511839866638184, 0.01148697566986084, 0.011498720169067382, 0.011485568046569824, 0.011500127792358398, 0.011495519638061523, 0.011516960144042968, 0.011516448020935058, 0.011573408126831055, 0.011765151977539063, 0.011571840286254883, 0.0115730562210083, 0.011597984313964844, 0.011526240348815917, 0.011577247619628906, 0.0115382080078125, 0.011534784317016602, 0.01153606414794922, 0.011613823890686036, 0.011520480155944824, 0.011546624183654786, 0.011493535995483398, 0.011507776260375977, 0.01162012767791748, 0.011538335800170899, 0.011587679862976074, 0.011708064079284668, 0.01150921630859375, 0.011522111892700195, 0.01178502368927002, 0.011551936149597168, 0.011539263725280762, 0.011499423980712891, 0.011528287887573242, 0.01153212833404541, 0.011534496307373047, 0.011478848457336426, 0.011493120193481445, 0.011534784317016602, 0.011536383628845214, 0.011517951965332032, 0.011537887573242188, 0.01152451229095459, 0.011483200073242187, 0.011541855812072755, 0.011548416137695312, 0.01157868766784668, 0.011396991729736327, 0.011537023544311523, 0.011621824264526367, 0.011586112022399902, 0.011549823760986328, 0.011544575691223144, 0.011541600227355958, 0.011531583786010742, 0.011555295944213867, 0.01155628776550293, 0.011544480323791503, 0.011502495765686035, 0.011529088020324707, 0.011513888359069824, 0.011526880264282226, 0.01162604808807373, 0.0116778564453125, 0.011740639686584472, 0.011762175559997558, 0.01168124771118164, 0.011699168205261231, 0.011655391693115235, 0.01164675235748291, 0.011538432121276856, 0.011552767753601074, 0.011548064231872558, 0.011569408416748047, 0.011609727859497071, 0.011680480003356933, 0.011663359642028808, 0.011616255760192871, 0.01163263988494873, 0.011616255760192871, 0.011575296401977539, 0.011560959815979004, 0.011608063697814941, 0.011528191566467285, 0.01159552001953125, 0.011566559791564941, 0.011600543975830078, 0.01161023998260498, 0.011583488464355468, 0.011576416015625, 0.011649824142456055, 0.011569024085998535, 0.011546879768371583, 0.011523648262023926, 0.01151961612701416, 0.011710559844970703, 0.01160582447052002, 0.011629376411437988, 0.01158902359008789, 0.011659168243408203, 0.01159177589416504, 0.011670080184936524, 0.011669887542724609, 0.011583488464355468, 0.011606975555419922, 0.011600319862365722, 0.011616640090942383, 0.011644448280334473, 0.011595487594604492, 0.011618176460266113, 0.011320863723754883, 0.011556511878967285, 0.011838272094726562, 0.011838848114013673, 0.011623040199279785, 0.011645024299621581, 0.01162019157409668, 0.01157481575012207, 0.01157583999633789, 0.01150169563293457, 0.011534208297729493, 0.011498592376708984, 0.011508480072021484, 0.011548831939697265, 0.011511808395385742, 0.011521663665771485, 0.011630784034729004, 0.011590047836303711, 0.01153769588470459, 0.011499456405639649, 0.01150972843170166, 0.011551008224487305, 0.011491647720336914, 0.011468223571777344, 0.01149395179748535, 0.011523455619812011, 0.011504544258117675, 0.012517087936401368, 0.012246591567993164, 0.011889087677001952, 0.011517760276794433, 0.01162054443359375, 0.01155510425567627, 0.011564224243164063, 0.011545120239257812, 0.011607328414916991, 0.011733663558959962, 0.011609312057495116, 0.01159654426574707, 0.011566240310668946, 0.011588607788085938, 0.011586784362792969, 0.011612895965576173, 0.012664352416992188, 0.011820704460144042, 0.01181116771697998, 0.011679679870605468, 0.011630240440368653, 0.011608960151672364, 0.011599007606506348, 0.011879263877868652, 0.011613887786865234, 0.011565343856811523, 0.011607328414916991, 0.011614975929260253, 0.011653120040893555, 0.011548095703125, 0.011572992324829102, 0.011595871925354004, 0.011946720123291016, 0.01164236831665039, 0.01163929557800293, 0.011675647735595703, 0.01133135986328125, 0.01159001636505127, 0.011608256340026855, 0.011569439888000488, 0.011585151672363281, 0.011748895645141602, 0.011701087951660156, 0.011600128173828125, 0.011696063995361328, 0.011553824424743652, 0.011541536331176758, 0.01151097583770752, 0.01154105567932129, 0.011599552154541016, 0.011548992156982421, 0.011579392433166504, 0.011636735916137696, 0.011538432121276856, 0.011517951965332032, 0.011525471687316895, 0.01165334415435791, 0.011605631828308106, 0.01167033576965332, 0.01181056022644043, 0.012308735847473145, 0.01163263988494873, 0.01151200008392334, 0.011519743919372558, 0.011556672096252442, 0.011782400131225586, 0.011712608337402343, 0.011829440116882323, 0.011890399932861328, 0.01166748809814453, 0.01164425563812256, 0.011558943748474122, 0.011532032012939453, 0.011584287643432617, 0.011538335800170899, 0.011622688293457032, 0.011929471969604493, 0.011593824386596679, 0.011601087570190429, 0.01155072021484375, 0.011549152374267579, 0.01158739185333252, 0.01166096019744873, 0.011571999549865722, 0.011548704147338867, 0.011544544219970703, 0.011554816246032714, 0.011589119911193848, 0.01154099178314209, 0.011610112190246581, 0.011581631660461425, 0.011583423614501953, 0.011599552154541016, 0.011527456283569337, 0.011563424110412598, 0.011548768043518067, 0.011579808235168456, 0.011499520301818847, 0.011547807693481446, 0.011324224472045898, 0.011605631828308106, 0.011601535797119141, 0.011564064025878906, 0.011595456123352051, 0.011597791671752929, 0.01163811206817627, 0.011672287940979005, 0.011622400283813476, 0.011603232383728027, 0.0115894718170166, 0.011871328353881836, 0.011739263534545899, 0.011597087860107422, 0.011570624351501465, 0.011594688415527344, 0.01157692813873291, 0.011547039985656739, 0.011569087982177734, 0.011532383918762207, 0.011570879936218262, 0.011544927597045899, 0.01155679988861084, 0.011587583541870117, 0.011525728225708009, 0.011614432334899902, 0.011542719841003418, 0.011679136276245117, 0.011667648315429687, 0.011620063781738282, 0.011629440307617188, 0.011591487884521485, 0.011610112190246581, 0.011587039947509765, 0.011653663635253907, 0.011584799766540527, 0.011586272239685058, 0.011583488464355468, 0.011591679573059082, 0.011585536003112793, 0.011562368392944337, 0.011555456161499023, 0.011595168113708497, 0.011614879608154298, 0.011576319694519043, 0.011750304222106933, 0.011598976135253906, 0.011638848304748536, 0.011655488014221192, 0.011684384346008301, 0.011706368446350097, 0.01215283203125, 0.01173846435546875, 0.011721376419067384, 0.011668512344360352, 0.011633952140808106, 0.011560416221618653, 0.011532511711120606, 0.011565055847167969, 0.011964415550231934, 0.011577343940734864, 0.011609279632568359, 0.011574175834655762, 0.011409952163696288, 0.011574560165405274, 0.011592576026916504, 0.011579039573669434, 0.015159584045410157, 0.011781120300292968, 0.011895520210266113, 0.012040543556213378, 0.011678912162780762, 0.012071392059326173, 0.011631967544555664, 0.01158784008026123, 0.01154911994934082, 0.011625760078430176, 0.011839327812194824, 0.011596639633178711, 0.01169321632385254, 0.011586175918579102, 0.011512127876281739, 0.011509663581848144, 0.011529919624328613, 0.011582783699035645, 0.011521023750305176, 0.011522047996520996, 0.0118538236618042, 0.011583488464355468, 0.011623871803283692, 0.011559103965759277, 0.011528479576110839, 0.011558112144470214, 0.011527039527893067, 0.011599871635437011, 0.011497792243957519, 0.011528127670288086, 0.011503680229187012, 0.011514880180358887, 0.011549375534057616, 0.011481087684631347, 0.011513855934143067, 0.011479040145874024, 0.01144217586517334, 0.0117391357421875, 0.011477215766906738, 0.011482591629028321, 0.011491168022155761, 0.01149180793762207, 0.011520288467407226, 0.011497407913208009, 0.011453568458557129, 0.011601951599121094, 0.01155123233795166, 0.0115316162109375, 0.011485183715820312, 0.011467552185058594, 0.01149948787689209, 0.011585568428039551, 0.01153228759765625, 0.011511808395385742, 0.011530336380004882, 0.011515263557434083, 0.01188054370880127, 0.011671327590942382, 0.011587488174438476, 0.011274720191955566, 0.01171664047241211, 0.011661791801452636, 0.01176576042175293, 0.013993472099304198, 0.011579744338989258, 0.01155072021484375, 0.011563167572021485, 0.011564127922058106, 0.01154963207244873, 0.01157475185394287, 0.011520511627197265, 0.011618399620056152, 0.011521471977233886, 0.011629023551940919, 0.011529888153076172, 0.011524383544921874, 0.01148902416229248, 0.011532352447509766, 0.01157260799407959, 0.011989279747009278, 0.01283465576171875, 0.012496928215026855, 0.01159068775177002, 0.011574975967407227, 0.011568991661071778, 0.011477375984191894, 0.011532416343688965, 0.01158249568939209, 0.011495743751525879, 0.011513343811035156, 0.011567520141601563, 0.011548895835876465, 0.011514368057250977, 0.011483039855957031, 0.011503168106079101, 0.01147494411468506, 0.011527520179748535, 0.011537280082702637, 0.011521183967590332, 0.011533087730407714, 0.011484895706176758, 0.011499327659606933, 0.011513728141784668, 0.011469471931457519, 0.011448320388793945, 0.011683296203613282, 0.011519840240478516, 0.011526368141174316, 0.011581600189208984, 0.011582880020141602, 0.011568127632141113, 0.011606176376342774, 0.011882240295410157, 0.011601119995117188, 0.011520575523376465, 0.011532192230224609, 0.01196678352355957, 0.011513855934143067, 0.011581439971923829, 0.011605792045593262, 0.01234556770324707, 0.011591679573059082]",tokens/s,85.59671858885318,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1559.187456,1861.156864,0.0,1465.909248,1358.169088,s,1,9.04726953125,9.04726953125,0.0,9.04726953125,9.04726953125,9.04726953125,9.04726953125,[9.04726953125],,kWh,5.35661705249898e-05,5.901440951654836e-06,1.904251523400513e-05,7.851012671064976e-05,,MB,1643.769856,1882.128384,0.0,1472.200704,1356.544512,s,10,0.5091492156982422,0.050914921569824224,0.0001894923756759238,0.050911760330200194,0.05102181968688965,0.05121476535797119,0.051369121894836424,"[0.051407711029052734, 0.05073353576660156, 0.05079391860961914, 0.050764545440673825, 0.05096934509277344, 0.0509356803894043, 0.05097894287109375, 0.050953472137451175, 0.05072422409057617, 0.05088784027099609]",tokens/s,5027.9955680364565,kWh,1.540151951425279e-06,1.6985144640651348e-07,1.0198487690947573e-06,2.7298521669265497e-06,tokens/kWh,93777971.97282736,MB,1654.398976,1882.128384,0.0,1472.200704,1409.728,s,10,14.181490478515625,1.4181490478515626,0.011949045983867831,1.4157312622070313,1.4365609375,1.4378442504882814,1.4388709008789062,"[1.4074014892578126, 1.4060621337890624, 1.4138895263671876, 1.4052296142578125, 1.4362757568359374, 1.4266898193359374, 1.4391275634765626, 1.4070614013671876, 1.417572998046875, 1.42218017578125]",tokens/s,44.424103443458506,kWh,4.134897928982142e-05,4.56039416331762e-06,1.8563500230904383e-05,6.447287368404344e-05,tokens/kWh,977155.1413814527,,s,630,14.178197593688962,0.022505075545538038,0.00040603970772232217,0.022582143783569333,0.022888503074645995,0.02301265926361084,0.02379591606140137,"[0.022025472640991212, 0.024424575805664064, 0.022487903594970705, 0.021987327575683592, 0.02176527976989746, 0.021830495834350587, 0.021835775375366212, 0.021753856658935547, 0.02182476806640625, 0.021838592529296874, 0.022016000747680665, 0.02205286407470703, 0.021999488830566405, 0.021917823791503907, 0.02184339141845703, 0.02192348861694336, 0.02226883125305176, 0.02215705680847168, 0.02219647979736328, 0.022005760192871093, 0.02232499122619629, 0.022150783538818358, 0.02206300735473633, 0.022018783569335936, 0.02191974449157715, 0.021850112915039063, 0.022033536911010742, 0.022864288330078125, 0.021902912139892577, 0.02205174446105957, 0.02206105613708496, 0.02249648094177246, 0.022553375244140625, 0.02262575912475586, 0.022878751754760743, 0.022747135162353514, 0.022773759841918945, 0.022775583267211914, 0.022773056030273436, 0.02280745506286621, 0.022767616271972657, 0.02275328063964844, 0.022804479598999023, 0.02289254379272461, 0.022929311752319336, 0.022975648880004883, 0.02287481689453125, 0.02300339126586914, 0.022779903411865234, 0.0225218563079834, 0.022543680191040038, 0.022575584411621094, 0.022440160751342773, 0.022427648544311524, 0.022368064880371095, 0.022573247909545898, 0.022187679290771485, 0.021956960678100587, 0.021991424560546875, 0.021909280776977538, 0.021983455657958985, 0.022040576934814454, 0.022013343811035157, 0.022180288314819337, 0.022029375076293944, 0.022129184722900392, 0.022351999282836914, 0.0223251838684082, 0.02210646438598633, 0.022005760192871093, 0.021977088928222657, 0.021941343307495118, 0.021882976531982422, 0.02197587203979492, 0.021921791076660157, 0.022042623519897463, 0.022045984268188476, 0.022045408248901367, 0.02203788757324219, 0.022080127716064452, 0.022013952255249023, 0.021942272186279296, 0.021981184005737304, 0.021897216796875, 0.022409215927124023, 0.023537664413452147, 0.022390272140502928, 0.022047231674194336, 0.02215116882324219, 0.022595552444458007, 0.022800416946411134, 0.02263859176635742, 0.022691776275634765, 0.023191616058349608, 0.022660991668701173, 0.0226243839263916, 0.02269958305358887, 0.022587007522583007, 0.02265990447998047, 0.022719648361206053, 0.022598112106323242, 0.02258367919921875, 0.022648832321166993, 0.022747135162353514, 0.022845439910888672, 0.022841344833374022, 0.022548128128051757, 0.02243414306640625, 0.022372352600097657, 0.02271027183532715, 0.02254643249511719, 0.02230271911621094, 0.022202367782592772, 0.022136831283569337, 0.022349279403686525, 0.022045215606689452, 0.02214271926879883, 0.022296831130981444, 0.02212633514404297, 0.02199782371520996, 0.02186147117614746, 0.02204742431640625, 0.021997791290283203, 0.022111295700073242, 0.02199443244934082, 0.021991680145263672, 0.02243027114868164, 0.022374399185180666, 0.022461856842041016, 0.022299232482910155, 0.022331167221069335, 0.02237017631530762, 0.02234511947631836, 0.02222585678100586, 0.02230790328979492, 0.02226652717590332, 0.022452512741088868, 0.0222740478515625, 0.02230271911621094, 0.022353919982910156, 0.02228144073486328, 0.02257695960998535, 0.022541120529174806, 0.02243190383911133, 0.02247043228149414, 0.022374624252319335, 0.022322528839111327, 0.0224586238861084, 0.02235219192504883, 0.022370208740234376, 0.02240121650695801, 0.02275071907043457, 0.02273126411437988, 0.02265497589111328, 0.02288844871520996, 0.02272051239013672, 0.022824960708618162, 0.022879392623901366, 0.0226549129486084, 0.02276972770690918, 0.022592351913452147, 0.02269491195678711, 0.022760448455810548, 0.02269718360900879, 0.022734880447387695, 0.02268441581726074, 0.022773439407348633, 0.022752960205078124, 0.022680000305175783, 0.022331584930419923, 0.02239897537231445, 0.02233344078063965, 0.022540288925170897, 0.022743040084838868, 0.022401023864746093, 0.02230886459350586, 0.022353567123413087, 0.022224767684936524, 0.02214681625366211, 0.02213711929321289, 0.02232979202270508, 0.02206924819946289, 0.022071296691894532, 0.022130687713623046, 0.022261760711669923, 0.022223871231079103, 0.02202012825012207, 0.021980127334594726, 0.022016000747680665, 0.022521568298339845, 0.0226014404296875, 0.022661695480346678, 0.022699712753295898, 0.022622528076171874, 0.02255580711364746, 0.022566944122314452, 0.022553407669067382, 0.02259328079223633, 0.022585119247436523, 0.022950368881225584, 0.02253004837036133, 0.022605823516845702, 0.02267136001586914, 0.02269388771057129, 0.022617120742797852, 0.02236675262451172, 0.022175647735595702, 0.022188575744628906, 0.022128448486328126, 0.022143104553222655, 0.022090879440307618, 0.022109119415283204, 0.021996639251708985, 0.02191187286376953, 0.021719648361206056, 0.021909503936767577, 0.02170809555053711, 0.021684928894042967, 0.02166783905029297, 0.02184806442260742, 0.021761920928955077, 0.021956735610961915, 0.022029504776000977, 0.022192960739135743, 0.0221265926361084, 0.022183935165405275, 0.022234176635742186, 0.022211519241333008, 0.022177791595458983, 0.022597631454467772, 0.022261760711669923, 0.02209587287902832, 0.022064191818237305, 0.0220067195892334, 0.022370304107666016, 0.02206278419494629, 0.02201398468017578, 0.02213257598876953, 0.022106559753417968, 0.022130239486694337, 0.022097471237182618, 0.02212505531311035, 0.02219251251220703, 0.022603776931762694, 0.022711423873901366, 0.022727615356445314, 0.02267513656616211, 0.02264703941345215, 0.022657024383544923, 0.022636415481567383, 0.022685728073120116, 0.022685279846191408, 0.022894176483154297, 0.022700223922729492, 0.022828800201416016, 0.02282748794555664, 0.022845439910888672, 0.023136255264282226, 0.02306153678894043, 0.022729375839233398, 0.022667583465576173, 0.02257846450805664, 0.022614688873291017, 0.022614080429077147, 0.022498336791992188, 0.022250463485717773, 0.022112255096435548, 0.02211020851135254, 0.023009279251098632, 0.022681215286254882, 0.0228089599609375, 0.022718143463134766, 0.022681312561035158, 0.02267401695251465, 0.02265100860595703, 0.022617984771728515, 0.02268083190917969, 0.02249372863769531, 0.02262873649597168, 0.022667104721069337, 0.022677312850952147, 0.022606016159057617, 0.022679040908813477, 0.022497791290283203, 0.022652799606323243, 0.022810111999511717, 0.02309744071960449, 0.022707775115966798, 0.022531040191650392, 0.024663904190063476, 0.023474016189575196, 0.022782207489013672, 0.022845504760742187, 0.022794240951538085, 0.02269171142578125, 0.0227063045501709, 0.022633695602416994, 0.02263734436035156, 0.024266271591186522, 0.023129823684692383, 0.022687807083129882, 0.024545984268188478, 0.022742080688476562, 0.022741952896118165, 0.02274086380004883, 0.0225316162109375, 0.02318191909790039, 0.022849536895751952, 0.022609920501708985, 0.022771711349487304, 0.022652032852172852, 0.022637439727783204, 0.022767616271972657, 0.02250294494628906, 0.02233558464050293, 0.022548479080200197, 0.02246019172668457, 0.022376672744750976, 0.023231584548950194, 0.022500255584716796, 0.022660991668701173, 0.02270220756530762, 0.02299622344970703, 0.02256972885131836, 0.022374399185180666, 0.022278144836425783, 0.02215116882324219, 0.022278144836425783, 0.02255183982849121, 0.02218448066711426, 0.022164800643920898, 0.022283136367797853, 0.022401023864746093, 0.02247372817993164, 0.022764543533325195, 0.022657024383544923, 0.022560768127441407, 0.022612127304077148, 0.022603103637695313, 0.02255238342285156, 0.022575456619262694, 0.022579008102416993, 0.02257155227661133, 0.023085056304931642, 0.022783424377441405, 0.02270470428466797, 0.022714111328125, 0.022560672760009767, 0.02268172836303711, 0.022687679290771486, 0.02260201644897461, 0.022664352416992186, 0.022698272705078126, 0.022561023712158203, 0.022602048873901368, 0.022591487884521484, 0.022684671401977538, 0.022704288482666014, 0.022637407302856447, 0.02270207977294922, 0.022606975555419923, 0.022790815353393553, 0.022673631668090822, 0.02267670440673828, 0.02262915229797363, 0.022773759841918945, 0.022906688690185546, 0.022951839447021484, 0.02275542449951172, 0.0229050235748291, 0.02272051239013672, 0.023015424728393553, 0.023259136199951173, 0.02276291275024414, 0.022679584503173828, 0.02269241523742676, 0.022626304626464845, 0.022560768127441407, 0.022639232635498045, 0.022566911697387695, 0.0226693115234375, 0.022580255508422853, 0.022606367111206054, 0.02264313507080078, 0.022689792633056642, 0.02265907287597656, 0.022773759841918945, 0.022673023223876952, 0.022634016036987305, 0.022575040817260743, 0.022557823181152344, 0.02268137550354004, 0.022603776931762694, 0.022556671142578123, 0.022484512329101564, 0.02264931106567383, 0.022716032028198243, 0.02269808006286621, 0.022745376586914064, 0.022648832321166993, 0.024188383102416992, 0.023607423782348633, 0.022935680389404297, 0.022765087127685546, 0.022803199768066405, 0.02371148872375488, 0.022826400756835938, 0.022664031982421874, 0.022700031280517577, 0.02264192008972168, 0.022629119873046874, 0.02262835121154785, 0.0226810245513916, 0.022730783462524416, 0.023067167282104492, 0.022788095474243163, 0.022662656784057617, 0.022745311737060545, 0.022775680541992187, 0.023114143371582033, 0.023830400466918946, 0.022810752868652345, 0.022872032165527342, 0.02282908821105957, 0.022794240951538085, 0.022996992111206056, 0.022949472427368164, 0.0231243839263916, 0.023142400741577147, 0.02348236846923828, 0.022992767333984376, 0.022977983474731446, 0.02319152069091797, 0.022915807723999024, 0.02281062316894531, 0.02315216064453125, 0.02276710319519043, 0.022629344940185547, 0.022706239700317384, 0.022527008056640624, 0.02256121635437012, 0.022552127838134765, 0.022764127731323244, 0.02269593620300293, 0.022751232147216797, 0.02279347229003906, 0.0227457275390625, 0.022622112274169923, 0.022910655975341795, 0.022888992309570314, 0.02272051239013672, 0.022915071487426757, 0.022863872528076173, 0.0230830078125, 0.022769535064697265, 0.022935680389404297, 0.022721824645996095, 0.022681760787963866, 0.02266579246520996, 0.02262015914916992, 0.022662784576416014, 0.02277174377441406, 0.02265532875061035, 0.022665216445922853, 0.022867456436157226, 0.02297702407836914, 0.02288640022277832, 0.022857183456420897, 0.02253385543823242, 0.022377151489257813, 0.022327423095703125, 0.022450176239013672, 0.02219340705871582, 0.02232729530334473, 0.02229324722290039, 0.02224742317199707, 0.021979040145874023, 0.0218768310546875, 0.021886016845703123, 0.021813823699951173, 0.021788543701171875, 0.02178713607788086, 0.0218603515625, 0.021909503936767577, 0.021893119812011717, 0.02186854362487793, 0.021900352478027345, 0.02193212890625, 0.021842784881591797, 0.021824607849121092, 0.0221246395111084, 0.021883487701416016, 0.02184828758239746, 0.022161407470703123, 0.021790719985961913, 0.021770015716552734, 0.022009311676025392, 0.021860639572143556, 0.021967103958129883, 0.021910783767700195, 0.02201193618774414, 0.021898176193237303, 0.021762048721313477, 0.02186854362487793, 0.021831487655639647, 0.02183635139465332, 0.02170460891723633, 0.021683744430541992, 0.021822015762329103, 0.02190540885925293, 0.021865983963012696, 0.021797376632690428, 0.02192793655395508, 0.02312396812438965, 0.02424831962585449, 0.021944128036499023, 0.02227203178405762, 0.022223007202148436, 0.022149120330810547, 0.022326271057128907, 0.02234060859680176, 0.022577152252197266, 0.022237152099609376, 0.022161376953125, 0.02207084846496582, 0.022168064117431642, 0.02231091117858887, 0.022581247329711913, 0.022587072372436522, 0.0224051513671875, 0.02247612762451172, 0.022264768600463867, 0.02229452705383301, 0.022321151733398437, 0.022441247940063476, 0.022389408111572265, 0.022468671798706055, 0.022350912094116212, 0.022524768829345704, 0.022634592056274414, 0.02273689651489258, 0.02272172737121582, 0.02284422492980957, 0.022634368896484375, 0.022601568222045898, 0.022578752517700196, 0.022637279510498046, 0.022666303634643555, 0.022696895599365233, 0.022729888916015625, 0.022762336730957032, 0.02268876838684082, 0.022732831954956054, 0.022776800155639647, 0.022867168426513672, 0.02280268859863281, 0.022831647872924805, 0.02276051139831543, 0.022831520080566405, 0.02285196876525879, 0.02304630470275879, 0.022956031799316406, 0.02294281578063965, 0.022876127243041992, 0.02274604797363281, 0.022771711349487304, 0.022746496200561524, 0.02287968063354492, 0.02282111930847168, 0.022941024780273437, 0.02272764778137207, 0.022705631256103517, 0.022778400421142576, 0.022854976654052735, 0.022808671951293946, 0.022874719619750978, 0.022840799331665038, 0.022945632934570314, 0.022790847778320314, 0.022715423583984377, 0.022658016204833983, 0.022931455612182617, 0.023023616790771483, 0.02290483283996582, 0.022953344345092774, 0.022860416412353514, 0.02270412826538086, 0.02271023941040039, 0.022704160690307618, 0.022654880523681642, 0.022585439682006835, 0.022654783248901366, 0.02251740837097168, 0.022485536575317384, 0.02246451187133789, 0.022679391860961913, 0.022311071395874023, 0.022376447677612304, 0.02232867240905762, 0.022169376373291017, 0.022576000213623045, 0.022347776412963868, 0.02224947166442871, 0.022300416946411133, 0.022276351928710938, 0.02225152015686035, 0.022192127227783204, 0.02208358383178711, 0.0222740478515625, 0.022021215438842775, 0.021990304946899415, 0.021925888061523437, 0.021999616622924805, 0.022171648025512695, 0.022177791595458983, 0.02225971221923828, 0.022427648544311524, 0.022556543350219727, 0.0226243839263916, 0.02245996856689453, 0.022423999786376953, 0.022560768127441407, 0.022583040237426757, 0.02259174346923828, 0.02283513641357422, 0.02273695945739746, 0.022779903411865234, 0.02290483283996582, 0.022861568450927735, 0.022970624923706055]",tokens/s,44.43442093658133,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen2,Qwen/Qwen1.5-1.8B,Qwen/Qwen1.5-1.8B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,1947.234304,2886.664192,0.0,2491.416576,2425.650176,s,1,9.868951171875,9.868951171875,0.0,9.868951171875,9.868951171875,9.868951171875,9.868951171875,[9.868951171875],,kWh,8.412446947921428e-05,9.272347202353469e-06,3.036085762200824e-05,0.00012375767430357598,,MB,1804.824576,3115.25376,0.0,2705.32608,2606.127616,s,10,0.5254048614501953,0.05254048614501953,0.0002773108713715526,0.05247787284851074,0.05277234954833984,0.05302678260803223,0.05323032905578613,"[0.05328121566772461, 0.05221033477783203, 0.05246867370605469, 0.05251366424560547, 0.0524870719909668, 0.05235599899291992, 0.05256211090087891, 0.05242294311523438, 0.052387039184570314, 0.0527158088684082]",tokens/s,4872.433027996773,kWh,1.6226055165509899e-06,1.7894313008149922e-07,1.076778639200039e-06,2.878327285832528e-06,tokens/kWh,88940545.87192453,MB,1813.352448,3115.25376,0.0,2705.32608,2606.130176,s,10,15.354823120117187,1.535482312011719,0.007036248048440928,1.5348394165039063,1.544455078125,1.546784729003906,1.5486484497070312,"[1.541155517578125, 1.53364013671875, 1.5439373779296874, 1.5290914306640624, 1.5290074462890626, 1.5267310791015625, 1.5369732666015625, 1.5360386962890624, 1.5491143798828124, 1.5291337890625]",tokens/s,41.02945342135545,kWh,4.509273926678394e-05,4.972856784603393e-06,2.1461878280601573e-05,7.15274743319889e-05,tokens/kWh,880780.4356070323,,s,630,15.352706731796268,0.024369375764755976,0.0005069567612018552,0.024270143508911133,0.024675234603881834,0.024878185272216795,0.026551848983764656,"[0.02491788864135742, 0.024565887451171876, 0.0244770565032959, 0.024411935806274414, 0.02661030387878418, 0.024983264923095702, 0.0243450870513916, 0.024272768020629883, 0.024323999404907228, 0.024269023895263673, 0.02430771255493164, 0.024483808517456053, 0.024207040786743163, 0.024369504928588866, 0.024154111862182616, 0.024174591064453126, 0.024319583892822266, 0.02429497528076172, 0.02416057586669922, 0.024277536392211915, 0.024219648361206055, 0.02426211166381836, 0.02471785545349121, 0.02448588752746582, 0.024548864364624022, 0.024412511825561523, 0.024575359344482423, 0.024414079666137695, 0.02506166458129883, 0.02458687973022461, 0.024655744552612303, 0.024481311798095703, 0.02459062385559082, 0.024663936614990233, 0.024816064834594725, 0.024495519638061524, 0.024796831130981446, 0.024913919448852538, 0.02438265609741211, 0.024207103729248048, 0.024388896942138673, 0.024675039291381835, 0.024383487701416014, 0.02430486488342285, 0.0243023681640625, 0.024389535903930663, 0.024212671279907227, 0.024167327880859374, 0.024248064041137694, 0.02421513557434082, 0.0241814079284668, 0.024163328170776367, 0.02418307113647461, 0.0242511043548584, 0.024449024200439453, 0.024363008499145508, 0.024188928604125977, 0.02473574447631836, 0.024076255798339843, 0.02415119934082031, 0.024197248458862303, 0.025268543243408204, 0.02424671936035156, 0.02435686492919922, 0.024133600234985352, 0.024125471115112304, 0.024120927810668946, 0.024091039657592773, 0.0241213436126709, 0.02452479934692383, 0.024240224838256837, 0.02448784065246582, 0.024131584167480468, 0.024469152450561523, 0.024158559799194335, 0.024456960678100586, 0.02447529602050781, 0.024283071517944337, 0.02412816047668457, 0.02415001678466797, 0.02444595146179199, 0.0241530876159668, 0.025384960174560548, 0.024129247665405272, 0.024501663208007812, 0.024080543518066405, 0.024268960952758788, 0.024013376235961913, 0.024236032485961914, 0.024066047668457033, 0.024022687911987306, 0.024242496490478514, 0.024862943649291994, 0.024149824142456054, 0.024110815048217774, 0.024254751205444337, 0.024385055541992186, 0.0242935676574707, 0.0243973445892334, 0.02438015937805176, 0.02449612808227539, 0.024347679138183594, 0.02435183906555176, 0.024362880706787108, 0.02440323257446289, 0.024329984664916992, 0.02425663948059082, 0.024199840545654296, 0.024123584747314453, 0.024694688796997072, 0.02494268798828125, 0.025475072860717773, 0.024489984512329102, 0.024132768630981447, 0.02417897605895996, 0.02423846435546875, 0.024490175247192384, 0.024427743911743165, 0.02427532768249512, 0.02434486389160156, 0.024114496231079103, 0.024268672943115233, 0.024243135452270508, 0.024573951721191405, 0.02438924789428711, 0.024439168930053712, 0.02679408073425293, 0.024801471710205077, 0.024880863189697264, 0.02500422477722168, 0.02476851272583008, 0.02455129623413086, 0.024522880554199218, 0.02453913688659668, 0.024527008056640626, 0.02464726448059082, 0.02409075164794922, 0.02443481636047363, 0.024141023635864258, 0.024180896759033205, 0.024269439697265624, 0.02411110305786133, 0.02403459167480469, 0.024122079849243163, 0.024184959411621094, 0.02412460708618164, 0.02407049560546875, 0.02406844711303711, 0.026804224014282226, 0.024825599670410155, 0.02429567909240723, 0.024436895370483398, 0.024587488174438475, 0.024371679306030274, 0.024483776092529295, 0.02472368049621582, 0.02511052894592285, 0.026408735275268554, 0.024821792602539063, 0.024660160064697265, 0.02470639991760254, 0.02463212776184082, 0.024571264266967773, 0.024616512298583984, 0.024423328399658203, 0.02430143928527832, 0.024234111785888673, 0.02428108787536621, 0.024268640518188476, 0.02439593505859375, 0.02534809684753418, 0.02426006317138672, 0.02417513656616211, 0.024055871963500976, 0.024084415435791016, 0.024159904479980468, 0.024237823486328126, 0.024143552780151366, 0.02419126319885254, 0.024114816665649415, 0.02407526397705078, 0.024625152587890626, 0.024453119277954103, 0.024302783966064452, 0.024208192825317384, 0.024152095794677735, 0.02412950325012207, 0.02411724853515625, 0.024061119079589844, 0.024164608001708984, 0.024045568466186523, 0.023969791412353517, 0.024037376403808593, 0.023969791412353517, 0.024252256393432616, 0.024037023544311524, 0.02409926414489746, 0.02412073516845703, 0.024058528900146484, 0.024093759536743163, 0.024011711120605468, 0.024014848709106446, 0.024059616088867187, 0.024013120651245116, 0.023990463256835938, 0.023992095947265625, 0.024006656646728516, 0.02395516777038574, 0.024068191528320314, 0.024070335388183595, 0.024271936416625978, 0.024287904739379883, 0.024219104766845703, 0.024275552749633788, 0.025054527282714845, 0.02495580863952637, 0.024461311340332033, 0.02451059150695801, 0.024422271728515625, 0.024414207458496092, 0.024213504791259766, 0.02425651168823242, 0.024061952590942383, 0.024023040771484375, 0.02414406394958496, 0.024198720932006836, 0.024229631423950196, 0.024041919708251952, 0.024020448684692382, 0.02403388786315918, 0.024102752685546874, 0.02405392074584961, 0.02427494430541992, 0.02405299186706543, 0.024089344024658205, 0.02427510452270508, 0.024893280029296874, 0.024829727172851562, 0.024539552688598632, 0.024463104248046874, 0.02458336067199707, 0.02460697555541992, 0.024541824340820313, 0.024600479125976564, 0.024522207260131837, 0.024422431945800783, 0.02445577621459961, 0.02449612808227539, 0.024416255950927734, 0.024532991409301756, 0.024619007110595705, 0.024383487701416014, 0.02473119926452637, 0.02456825637817383, 0.02431795120239258, 0.024145919799804686, 0.024079551696777345, 0.02426755142211914, 0.024102943420410156, 0.024147136688232422, 0.02408121681213379, 0.02422777557373047, 0.024100479125976564, 0.024144319534301757, 0.024143039703369142, 0.024445215225219728, 0.02456438446044922, 0.02451363182067871, 0.024351520538330076, 0.024325824737548827, 0.024243776321411132, 0.024346752166748045, 0.024236671447753905, 0.02416640090942383, 0.024270847320556642, 0.02425961685180664, 0.024072288513183594, 0.024208255767822266, 0.02413155174255371, 0.024219551086425782, 0.024234111785888673, 0.02422528076171875, 0.02413030433654785, 0.024110336303710938, 0.024125951766967774, 0.024100351333618163, 0.02402707290649414, 0.02399087905883789, 0.024426464080810548, 0.024178688049316405, 0.024363168716430662, 0.02418262481689453, 0.024164352416992187, 0.02413542366027832, 0.02446518325805664, 0.024483392715454102, 0.024111072540283204, 0.024242591857910157, 0.024030815124511717, 0.02402400016784668, 0.023950368881225585, 0.024064064025878906, 0.0242653751373291, 0.024221952438354493, 0.02444803237915039, 0.02432249641418457, 0.02445100784301758, 0.024355424880981445, 0.02449308776855469, 0.024634336471557616, 0.02464156723022461, 0.02458006477355957, 0.02456166458129883, 0.024377344131469726, 0.02426192092895508, 0.024551616668701173, 0.024418304443359375, 0.02443894386291504, 0.02487491226196289, 0.024250303268432617, 0.024170495986938476, 0.024422399520874022, 0.02424028778076172, 0.024110944747924804, 0.02407769584655762, 0.024058496475219727, 0.024041471481323243, 0.024016895294189454, 0.024401376724243164, 0.02448134422302246, 0.024273248672485353, 0.024083072662353516, 0.024023040771484375, 0.024041471481323243, 0.024012800216674804, 0.023941375732421874, 0.02399411201477051, 0.024029375076293946, 0.02405958366394043, 0.023966943740844727, 0.02407072067260742, 0.024025279998779295, 0.02399247932434082, 0.0240883846282959, 0.024090240478515625, 0.024138463973999023, 0.0239532470703125, 0.024010751724243166, 0.023991712570190428, 0.024105247497558595, 0.0239619197845459, 0.02464358329772949, 0.024160255432128908, 0.024006752014160155, 0.024213407516479494, 0.02408982467651367, 0.02416111946105957, 0.0240250244140625, 0.024039424896240235, 0.024345983505249025, 0.024253183364868165, 0.024173824310302735, 0.02447132873535156, 0.02424233627319336, 0.024338144302368164, 0.024071392059326173, 0.024104736328125, 0.024045536041259766, 0.02416640090942383, 0.024143871307373048, 0.024447200775146484, 0.024690464019775392, 0.024743295669555663, 0.0246954231262207, 0.02458380889892578, 0.02504742431640625, 0.024618431091308592, 0.024568384170532226, 0.024742528915405272, 0.024450111389160156, 0.024130495071411132, 0.024127008438110352, 0.024111583709716798, 0.02427903938293457, 0.02424998474121094, 0.024211839675903322, 0.024123008728027345, 0.024150400161743163, 0.024207359313964845, 0.024465408325195313, 0.02418214416503906, 0.02460678482055664, 0.02413420867919922, 0.024373439788818358, 0.02439727973937988, 0.02429952049255371, 0.024465503692626952, 0.02458415985107422, 0.02464531135559082, 0.024703584671020507, 0.024583839416503907, 0.024742240905761718, 0.026992639541625976, 0.0249715518951416, 0.024709888458251953, 0.02475257682800293, 0.024504480361938478, 0.02460300827026367, 0.02446272087097168, 0.024658559799194336, 0.02448355293273926, 0.02441859245300293, 0.024284320831298827, 0.024478368759155274, 0.02457209587097168, 0.024401920318603516, 0.024336383819580077, 0.02411315155029297, 0.024164608001708984, 0.02422915267944336, 0.024218080520629882, 0.02418467140197754, 0.024252576828002928, 0.024236032485961914, 0.02422515106201172, 0.024207231521606445, 0.024414560317993165, 0.024518112182617187, 0.024367391586303713, 0.024200927734375, 0.02420832061767578, 0.02421766471862793, 0.024131519317626953, 0.02406399917602539, 0.02428486442565918, 0.024400192260742186, 0.024139776229858398, 0.024030975341796875, 0.024069631576538086, 0.024177408218383788, 0.024125471115112304, 0.024378591537475586, 0.02430668830871582, 0.02407414436340332, 0.02410710334777832, 0.024014015197753907, 0.024122175216674806, 0.024137216567993162, 0.024224256515502928, 0.024209407806396483, 0.024223743438720705, 0.024122432708740236, 0.024537439346313476, 0.02452537536621094, 0.02471529579162598, 0.02416204833984375, 0.024046079635620117, 0.025385887145996093, 0.024093536376953124, 0.024141504287719728, 0.02414419174194336, 0.02427244758605957, 0.024266271591186522, 0.024355424880981445, 0.024312095642089845, 0.024427999496459962, 0.024455135345458984, 0.024408672332763674, 0.024533151626586914, 0.02455331230163574, 0.024594432830810548, 0.024424448013305664, 0.024371200561523438, 0.024679807662963866, 0.02475289535522461, 0.024424320220947267, 0.024319616317749024, 0.024344415664672853, 0.02436355209350586, 0.02431782341003418, 0.024361087799072267, 0.024178016662597657, 0.024369375228881836, 0.024133407592773437, 0.02411587142944336, 0.02410495948791504, 0.024376447677612306, 0.02413043212890625, 0.024188928604125977, 0.02450227165222168, 0.024898591995239257, 0.025083999633789062, 0.024423295974731446, 0.024483840942382814, 0.024465408325195313, 0.024423488616943358, 0.024568767547607423, 0.024407840728759764, 0.0245100154876709, 0.02451318359375, 0.024414207458496092, 0.024391679763793944, 0.024535039901733398, 0.024408063888549804, 0.024434528350830077, 0.024848543167114257, 0.03255497741699219, 0.02445913505554199, 0.024359071731567383, 0.02433987236022949, 0.024180511474609374, 0.02423891258239746, 0.024481792449951172, 0.024482879638671875, 0.024099775314331055, 0.02409881591796875, 0.024102144241333008, 0.024246240615844728, 0.02449398422241211, 0.024426591873168944, 0.024271007537841796, 0.024676992416381837, 0.024213375091552733, 0.025733247756958007, 0.02564499282836914, 0.024516096115112306, 0.024421024322509765, 0.024242080688476563, 0.024379392623901368, 0.02418227195739746, 0.02401055908203125, 0.023927488327026368, 0.02445516777038574, 0.02411337661743164, 0.02411065673828125, 0.024148191452026367, 0.02405344009399414, 0.027379167556762694, 0.024437599182128907, 0.024240127563476564, 0.024256607055664063, 0.02449807929992676, 0.024294559478759765, 0.02436751937866211, 0.024376960754394533, 0.02448214340209961, 0.024354848861694336, 0.02804278373718262, 0.024661088943481447, 0.024467199325561524, 0.024287296295166017, 0.024188287734985353, 0.02416908836364746, 0.024227840423583984, 0.024358911514282225, 0.024211456298828125, 0.024205312728881836, 0.024326047897338866, 0.024227519989013672, 0.024127904891967773, 0.02417568016052246, 0.024312768936157227, 0.024505952835083007, 0.02422559928894043, 0.02405846405029297, 0.02419478416442871, 0.024277280807495118, 0.02467875289916992, 0.024370111465454102, 0.024073152542114257, 0.02409267234802246, 0.0240579833984375, 0.02412646484375, 0.024182687759399413, 0.024072288513183594, 0.0239847354888916, 0.024076576232910155, 0.024061952590942383, 0.02400048065185547, 0.02405379295349121, 0.024061952590942383, 0.02405075263977051, 0.024011711120605468, 0.024035327911376952, 0.024068063735961914, 0.02400464057922363, 0.02408198356628418, 0.024074432373046874, 0.024084735870361328, 0.023984128952026368, 0.023998464584350586, 0.024231712341308595, 0.024223167419433592, 0.024347423553466797, 0.024008512496948242, 0.024084512710571288, 0.024098560333251952, 0.024241888046264648, 0.024070207595825194, 0.024179328918457033, 0.024851999282836913, 0.02484272003173828, 0.024449024200439453, 0.024577823638916016, 0.024501920700073242, 0.024541471481323244, 0.024518943786621093, 0.02466815948486328, 0.02464143943786621, 0.024663488388061525, 0.024627231597900392, 0.02489936065673828, 0.02526908874511719, 0.024513792037963868, 0.024355392456054687, 0.02428447914123535, 0.02428761672973633, 0.02427471923828125, 0.024236255645751954, 0.024314367294311523, 0.024212799072265624, 0.02418659210205078, 0.02428003120422363, 0.024075647354125977, 0.024154752731323243, 0.024156160354614258, 0.024112672805786134, 0.02406857681274414, 0.0242259521484375, 0.02435206413269043]",tokens/s,41.03510937880659,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Qwen/Qwen-14B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Qwen/Qwen-14B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 613, in resolve_trust_remote_code answer = input( EOFError: EOF when reading a line During handling of the above exception, another exception occurred: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking backend.load() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load self.load_transformers_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 149, in load_transformers_model self.create_no_weights_model() File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 269, in create_no_weights_model meta_model = self.automodel_loader.from_config(self.pretrained_config) File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 419, in from_config trust_remote_code = resolve_trust_remote_code( File ""/usr/local/lib/python3.10/dist-packages/transformers/dynamic_module_utils.py"", line 626, in resolve_trust_remote_code raise ValueError( ValueError: The repository for Qwen/Qwen-7B contains custom code which must be executed to correctly load the model. You can inspect the repository content at https://hf.co/Qwen/Qwen-7B. Please pass the argument `trust_remote_code=True` to allow custom code to be run. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,awq,4,exllama,2,64,1,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz,8,33163.75552,Linux,x86_64,Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['Tesla T4'],1,16106127360,0.4.0,,4.43.4,,0.33.0,,,,1.21.2,,,,0.12.0,,,MB,877.613056,615.448576,0.0,220.20096,205.438976,s,1,8.500099609375,8.500099609375,0.0,8.500099609375,8.500099609375,8.500099609375,8.500099609375,[8.500099609375],,kWh,2.8432755895839057e-05,3.1291399853608897e-06,8.50334013600268e-06,4.0065236017202625e-05,,MB,1265.762304,669.974528,0.0,260.046848,226.386944,s,17,0.21296153736114498,0.012527149256537942,0.0002239725954963185,0.01243660831451416,0.012815871810913086,0.012923238182067871,0.013149460105895997,"[0.01247321605682373, 0.012356096267700196, 0.012852543830871583, 0.012580448150634766, 0.012397248268127442, 0.013206015586853028, 0.012528351783752441, 0.012396703720092774, 0.012384832382202149, 0.012378560066223145, 0.012320768356323243, 0.012509632110595702, 0.012791423797607422, 0.01259267234802246, 0.01243660831451416, 0.012429247856140137, 0.012327168464660644]",tokens/s,20435.615059539035,kWh,3.6659520940118344e-07,4.0428975814141996e-08,2.1475729073366726e-07,6.217814759489927e-07,tokens/kWh,411720210.2383004,MB,1301.495808,684.654592,0.0,274.726912,226.389504,s,17,10.28993084716797,0.60529004983341,0.006259915020451108,0.6042811279296875,0.6105053955078125,0.6131617797851562,0.6216492016601562,"[0.6237710571289062, 0.60166943359375, 0.6040850830078125, 0.6031799926757813, 0.6103824462890625, 0.610502685546875, 0.6029651489257812, 0.6058961791992188, 0.5993203125, 0.5990332641601562, 0.6070938110351562, 0.6058821411132812, 0.606194580078125, 0.6105094604492187, 0.6042811279296875, 0.6007789306640625, 0.5943851928710937]",tokens/s,104.08233212711671,kWh,1.7470339347216228e-05,1.9266556545660242e-06,6.675967176089875e-06,2.6072962177872127e-05,tokens/kWh,2416296.2217414444,,s,1071,10.281280707359333,0.009599701874285074,0.0002051811333571774,0.009560256004333497,0.009805824279785156,0.009942975997924805,0.010313395404815674,"[0.009171008110046387, 0.009627840042114259, 0.009655872344970702, 0.009693375587463379, 0.009803775787353516, 0.009846783638000489, 0.009953280448913575, 0.009963104248046875, 0.010200960159301758, 0.010041824340820312, 0.009928768157958984, 0.009994239807128906, 0.009959199905395507, 0.010000032424926758, 0.01023977565765381, 0.010158592224121094, 0.010066207885742187, 0.01001471996307373, 0.010053119659423827, 0.01003980827331543, 0.00995952033996582, 0.009994144439697266, 0.009936960220336914, 0.010014656066894532, 0.010170080184936524, 0.010049823760986329, 0.010022496223449707, 0.010047679901123046, 0.010000800132751465, 0.009991999626159667, 0.009944128036499024, 0.009989055633544922, 0.009928704261779785, 0.009945152282714843, 0.009989888191223144, 0.009937088012695312, 0.00991436767578125, 0.009963520050048828, 0.009947168350219726, 0.010061440467834472, 0.010079999923706055, 0.009996895790100097, 0.009924351692199707, 0.009973919868469239, 0.009924703598022461, 0.009928704261779785, 0.009844672203063965, 0.009871264457702637, 0.009721823692321777, 0.009646400451660157, 0.009768832206726074, 0.009744000434875489, 0.009773632049560547, 0.009792448043823242, 0.009788031578063966, 0.009751999855041504, 0.0097259521484375, 0.009770079612731934, 0.009736063957214356, 0.00963753604888916, 0.009565952301025391, 0.009583104133605956, 0.009512703895568847, 0.009132831573486329, 0.009488384246826171, 0.009463839530944824, 0.009680191993713378, 0.009472736358642578, 0.009467840194702148, 0.009449600219726562, 0.009438207626342773, 0.009354080200195312, 0.00944099235534668, 0.009469792366027832, 0.009453920364379883, 0.009395456314086914, 0.009422911643981933, 0.009462592124938964, 0.009458687782287598, 0.009466879844665528, 0.009483360290527343, 0.009440128326416015, 0.009436575889587403, 0.009460351943969726, 0.009475263595581054, 0.00959558391571045, 0.00969536018371582, 0.009668607711791993, 0.009870495796203613, 0.009701567649841309, 0.009704095840454102, 0.009498623847961426, 0.009460000038146973, 0.009428735733032226, 0.009385439872741699, 0.009327103614807129, 0.009369855880737304, 0.009355072021484376, 0.009361344337463379, 0.009365056037902832, 0.009372096061706543, 0.009408415794372559, 0.00946390438079834, 0.00951910400390625, 0.009660096168518066, 0.00961308765411377, 0.009667103767395019, 0.009676095962524413, 0.009593536376953125, 0.009637887954711915, 0.009713919639587402, 0.009682687759399414, 0.009721856117248535, 0.009625887870788575, 0.009573663711547852, 0.009586175918579102, 0.009679807662963866, 0.009644031524658203, 0.009625696182250976, 0.009793439865112304, 0.009809920310974121, 0.009721599578857421, 0.009717375755310058, 0.009685471534729003, 0.009660575866699219, 0.00963379192352295, 0.009323552131652832, 0.009723967552185058, 0.009725631713867188, 0.009746432304382324, 0.009736096382141114, 0.009617504119873046, 0.009623488426208497, 0.009539711952209472, 0.009451456069946289, 0.00952444839477539, 0.009515456199645996, 0.009551360130310058, 0.009421567916870117, 0.009523200035095216, 0.009402496337890625, 0.00951852798461914, 0.009396896362304688, 0.009446944236755371, 0.009421183586120605, 0.009555232048034669, 0.009522111892700195, 0.00954054355621338, 0.009506912231445312, 0.009521951675415039, 0.009457280158996582, 0.009498527526855469, 0.0095217924118042, 0.009507712364196777, 0.009554240226745605, 0.00945206356048584, 0.009523263931274414, 0.009390080451965332, 0.00941055965423584, 0.009467904090881347, 0.009729887962341309, 0.009640031814575196, 0.010800640106201171, 0.009769023895263672, 0.009482751846313477, 0.009523200035095216, 0.009584671974182129, 0.0095600004196167, 0.009488224029541015, 0.009507231712341309, 0.009494303703308105, 0.00944495964050293, 0.009464223861694336, 0.009465855598449707, 0.009479488372802734, 0.009427647590637207, 0.009532416343688965, 0.009415264129638673, 0.009462176322937011, 0.009967616081237793, 0.00963804817199707, 0.009981792449951172, 0.009711615562438965, 0.009736319541931153, 0.00965824031829834, 0.009650176048278808, 0.009789440155029297, 0.00976300811767578, 0.00973958396911621, 0.009398367881774903, 0.009805824279785156, 0.009750687599182128, 0.009750528335571289, 0.00970355224609375, 0.00965839958190918, 0.009682175636291503, 0.00963753604888916, 0.009605152130126953, 0.009683456420898438, 0.00967910385131836, 0.009699007987976075, 0.009595264434814453, 0.009670880317687989, 0.0096212158203125, 0.009589887619018555, 0.00955072021484375, 0.009581791877746581, 0.009560864448547363, 0.00960870361328125, 0.009638400077819824, 0.009682656288146973, 0.0098306884765625, 0.009820159912109374, 0.0095862398147583, 0.009612031936645508, 0.009519840240478515, 0.009719008445739747, 0.010307040214538574, 0.00950819206237793, 0.009497407913208009, 0.009441408157348632, 0.009545120239257812, 0.009455583572387695, 0.00946457576751709, 0.009462944030761719, 0.009445695877075196, 0.009462176322937011, 0.00957148838043213, 0.009549759864807128, 0.00962172794342041, 0.009554304122924805, 0.009519200325012207, 0.009607647895812988, 0.009540575981140137, 0.009475040435791015, 0.009420607566833496, 0.009484383583068847, 0.00943446445465088, 0.00946233558654785, 0.009470080375671387, 0.009502592086791993, 0.009398271560668945, 0.009438719749450684, 0.009425408363342285, 0.009391743659973145, 0.009480256080627442, 0.009470175743103028, 0.009416159629821777, 0.009356096267700195, 0.009371359825134278, 0.00942255973815918, 0.009480575561523437, 0.00919539165496826, 0.009489055633544921, 0.009371808052062988, 0.009392127990722657, 0.009570143699645995, 0.00944099235534668, 0.009560256004333497, 0.009598464012145995, 0.00963046360015869, 0.009744319915771484, 0.00986950397491455, 0.009901951789855958, 0.009772255897521972, 0.009700127601623535, 0.009725824356079101, 0.00981824016571045, 0.009711615562438965, 0.009695199966430664, 0.009738112449645996, 0.009740544319152831, 0.009679776191711426, 0.009862272262573243, 0.009678303718566894, 0.00970748805999756, 0.009693535804748535, 0.00978166389465332, 0.00976863956451416, 0.00973414421081543, 0.009760767936706542, 0.009801888465881347, 0.009803296089172363, 0.009849120140075683, 0.009814271926879883, 0.009702752113342285, 0.00965062427520752, 0.00962764835357666, 0.009643872261047364, 0.009717920303344726, 0.009703424453735352, 0.009750656127929688, 0.009701248168945312, 0.00992198371887207, 0.009770848274230958, 0.009894240379333495, 0.009717791557312012, 0.01021561622619629, 0.009672127723693847, 0.009613632202148437, 0.009574879646301269, 0.009551712036132812, 0.009599072456359863, 0.009494560241699219, 0.009596799850463867, 0.009506912231445312, 0.009508447647094726, 0.009507231712341309, 0.009516480445861817, 0.009534015655517578, 0.009778592109680176, 0.00960377597808838, 0.009641504287719727, 0.009687423706054688, 0.009850879669189454, 0.009324543952941895, 0.00964515209197998, 0.009589632034301758, 0.00955504035949707, 0.009611840248107911, 0.00963212776184082, 0.009727999687194825, 0.00969711971282959, 0.009580703735351562, 0.009620863914489746, 0.00960307216644287, 0.009740960121154786, 0.009661664009094238, 0.009672575950622559, 0.009761664390563965, 0.009670399665832519, 0.009658623695373535, 0.009602432250976562, 0.009578623771667481, 0.009638400077819824, 0.009733375549316406, 0.00976089572906494, 0.009736831665039063, 0.009725664138793945, 0.009684927940368652, 0.00978774356842041, 0.009649920463562011, 0.009681023597717286, 0.009743743896484375, 0.009789343833923339, 0.00978121566772461, 0.009813119888305664, 0.009731840133666993, 0.009657855987548827, 0.009503199577331542, 0.009594911575317383, 0.009537311553955078, 0.009617631912231445, 0.009587008476257325, 0.00969491195678711, 0.00971571159362793, 0.009670207977294922, 0.009662431716918945, 0.009676575660705566, 0.0098056640625, 0.009737055778503418, 0.009699071884155273, 0.009733440399169922, 0.00975318431854248, 0.009720000267028808, 0.009915871620178223, 0.009802592277526856, 0.010184032440185547, 0.00979196834564209, 0.009754655838012695, 0.00963276767730713, 0.009667584419250488, 0.00958182430267334, 0.00965494441986084, 0.009599072456359863, 0.009582688331604004, 0.009639840126037597, 0.00960921573638916, 0.00929792022705078, 0.009689087867736817, 0.009654080390930176, 0.009646271705627442, 0.009649503707885742, 0.009935263633728028, 0.009639231681823731, 0.009681056022644043, 0.009609248161315918, 0.009585408210754395, 0.009542783737182618, 0.009476991653442382, 0.009478143692016602, 0.009482272148132324, 0.00952950382232666, 0.009523008346557617, 0.009414655685424805, 0.009484224319458008, 0.00949459171295166, 0.009535488128662109, 0.009529343605041504, 0.009584832191467284, 0.009631039619445801, 0.009661984443664551, 0.009627967834472657, 0.009737055778503418, 0.009662336349487304, 0.009559328079223633, 0.009552000045776367, 0.009482591629028321, 0.009488032341003419, 0.009537823677062988, 0.009518336296081543, 0.009475071907043458, 0.009479680061340333, 0.009483039855957031, 0.009533151626586914, 0.00955907154083252, 0.009726943969726563, 0.00952905559539795, 0.009521439552307128, 0.009510687828063965, 0.009373920440673828, 0.009457152366638183, 0.009426560401916504, 0.009427840232849122, 0.009445376396179199, 0.009512479782104493, 0.009518912315368652, 0.009427616119384766, 0.00944057559967041, 0.009456000328063964, 0.009515328407287598, 0.010233759880065918, 0.00963980770111084, 0.009683168411254883, 0.009689087867736817, 0.009656319618225098, 0.009720864295959473, 0.009599967956542969, 0.00951103973388672, 0.009512928009033203, 0.009490336418151855, 0.009165056228637695, 0.009415936470031739, 0.00938265609741211, 0.009409536361694336, 0.00937241554260254, 0.00933027172088623, 0.009364447593688964, 0.009411935806274413, 0.009388383865356445, 0.009390080451965332, 0.009383487701416015, 0.009430815696716308, 0.00937564754486084, 0.009410304069519043, 0.009468928337097168, 0.009412863731384277, 0.00934876823425293, 0.009426464080810546, 0.009412384033203125, 0.009421183586120605, 0.009427103996276855, 0.009400896072387695, 0.009358112335205077, 0.009396479606628419, 0.009394944190979004, 0.009483776092529296, 0.009447392463684081, 0.009490079879760742, 0.00958240032196045, 0.009833120346069335, 0.009627776145935059, 0.00975481605529785, 0.011046208381652833, 0.01058067226409912, 0.00982425594329834, 0.009805567741394043, 0.009711039543151856, 0.009804512023925781, 0.009725664138793945, 0.009687423706054688, 0.00982534408569336, 0.009741215705871583, 0.009658656120300292, 0.009708576202392578, 0.009941823959350586, 0.00987331199645996, 0.009747455596923828, 0.00971059226989746, 0.009843008041381836, 0.009844256401062011, 0.00981827163696289, 0.009767135620117188, 0.009688447952270508, 0.009711135864257812, 0.009737183570861816, 0.009630847930908203, 0.009642784118652343, 0.009627264022827148, 0.009607040405273437, 0.009544192314147949, 0.009564160346984863, 0.00953916835784912, 0.009517760276794433, 0.009058624267578126, 0.00949232006072998, 0.0094782075881958, 0.009377568244934083, 0.009380895614624024, 0.009429984092712403, 0.009412832260131836, 0.009476896286010741, 0.009476608276367187, 0.009590592384338379, 0.009569055557250976, 0.009490400314331056, 0.009492480278015136, 0.009602848052978516, 0.009642399787902833, 0.009704416275024413, 0.009544672012329102, 0.009545087814331055, 0.009611712455749512, 0.009443327903747559, 0.011451680183410645, 0.010408672332763672, 0.009653280258178711, 0.00951961612701416, 0.009482560157775878, 0.009478400230407715, 0.009477312088012696, 0.009429727554321289, 0.00946787166595459, 0.009475744247436524, 0.009490367889404296, 0.009535103797912598, 0.009470784187316894, 0.009422016143798829, 0.009470848083496094, 0.009419808387756348, 0.0094749755859375, 0.009498623847961426, 0.00941875171661377, 0.009441439628601073, 0.009448479652404786, 0.009462176322937011, 0.00950496006011963, 0.009538975715637207, 0.009491264343261719, 0.009500672340393066, 0.009441280364990234, 0.00942460823059082, 0.00943283176422119, 0.009398719787597657, 0.009399999618530273, 0.009406335830688476, 0.00933353614807129, 0.00936524772644043, 0.009364992141723634, 0.009334752082824707, 0.009383456230163574, 0.009349856376647949, 0.00933071994781494, 0.009394399642944336, 0.009390111923217773, 0.009396224021911622, 0.00944320011138916, 0.00910422420501709, 0.00940329647064209, 0.009640543937683106, 0.009717984199523926, 0.009580544471740723, 0.009473504066467285, 0.009511455535888672, 0.009469599723815917, 0.009451871871948242, 0.00944921588897705, 0.009576959609985352, 0.009397120475769042, 0.009390975952148438, 0.009457056045532226, 0.00946886444091797, 0.009419551849365234, 0.009382783889770508, 0.009373791694641113, 0.009432991981506348, 0.009418368339538574, 0.009483776092529296, 0.009499039649963379, 0.009457695960998535, 0.009476096153259277, 0.009492128372192384, 0.009571104049682618, 0.009510911941528321, 0.009511199951171875, 0.009462559700012207, 0.00970748805999756, 0.009495743751525879, 0.009486271858215331, 0.00948243236541748, 0.009498271942138672, 0.00945907211303711, 0.009527935981750488, 0.009486335754394531, 0.009486335754394531, 0.009474047660827637, 0.010303296089172364, 0.009507007598876953, 0.009545663833618163, 0.009566271781921386, 0.00951910400390625, 0.009592831611633301, 0.009515007972717286, 0.009516256332397461, 0.009462559700012207, 0.009488384246826171, 0.009514623641967774, 0.009496959686279297, 0.009467904090881347, 0.009456831932067871, 0.009416576385498046, 0.009629695892333985, 0.009510047912597656, 0.00952246379852295, 0.009447808265686036, 0.009457216262817383, 0.009454143524169923, 0.00950489616394043, 0.009461631774902344, 0.009506815910339356, 0.00906054401397705, 0.009447232246398925, 0.009499615669250488, 0.009506815910339356, 0.009447423934936524, 0.009434720039367676, 0.009488832473754883, 0.009455583572387695, 0.009481951713562012, 0.00935478401184082, 0.009437472343444823, 0.00941104030609131, 0.009596223831176759, 0.00964406394958496, 0.010400287628173828, 0.010328224182128906, 0.009663519859313964, 0.009513248443603515, 0.009525919914245605, 0.009545087814331055, 0.009527935981750488, 0.009539039611816406, 0.009586655616760253, 0.009589311599731445, 0.009667776107788085, 0.009604991912841797, 0.009605759620666504, 0.00966483211517334, 0.009691136360168457, 0.009743935585021972, 0.009705920219421387, 0.009610879898071288, 0.009609919548034668, 0.009643168449401855, 0.009634336471557618, 0.009661503791809082, 0.00980678367614746, 0.009740511894226074, 0.009710880279541015, 0.009898624420166015, 0.009799391746520995, 0.009852928161621094, 0.009749759674072265, 0.00981004810333252, 0.009923359870910645, 0.009751615524291992, 0.00973027229309082, 0.009787615776062011, 0.009729824066162109, 0.009724255561828613, 0.009680255889892578, 0.00961023998260498, 0.00962559986114502, 0.009615360260009765, 0.009559776306152344, 0.009611552238464355, 0.009512160301208496, 0.009536288261413573, 0.009527296066284179, 0.009443327903747559, 0.009508864402770996, 0.009500127792358398, 0.009505311965942382, 0.009476703643798828, 0.009670399665832519, 0.009631744384765625, 0.009697279930114745, 0.00969859218597412, 0.009681632041931152, 0.00971951961517334, 0.0096014404296875, 0.00956332778930664, 0.009609184265136718, 0.009517760276794433, 0.00952950382232666, 0.009523008346557617, 0.009516192436218261, 0.009671392440795899, 0.009494720458984375, 0.00947993564605713, 0.009517312049865723, 0.009430944442749023, 0.009511008262634277, 0.009662431716918945, 0.00947612762451172, 0.009488320350646973, 0.009642144203186036, 0.009512864112854003, 0.009500672340393066, 0.009456704139709473, 0.009518143653869628, 0.009435232162475587, 0.009445152282714843, 0.00973414421081543, 0.009494784355163575, 0.009483488082885742, 0.009506367683410644, 0.009413663864135741, 0.00960095977783203, 0.009500415802001953, 0.009463775634765624, 0.009595168113708497, 0.009612544059753418, 0.009621312141418457, 0.009638431549072266, 0.009648192405700683, 0.009674592018127441, 0.009671199798583984, 0.009633631706237792, 0.009644160270690918, 0.009672191619873047, 0.009816384315490723, 0.009676095962524413, 0.00974118423461914, 0.009797216415405274, 0.009735615730285644, 0.009796575546264648, 0.009756735801696777, 0.009629023551940919, 0.009669216156005859, 0.009676223754882812, 0.009752832412719726, 0.00971951961517334, 0.009730112075805665, 0.00979203224182129, 0.00975603199005127, 0.009398431777954101, 0.009790783882141113, 0.009690143585205079, 0.009721759796142579, 0.009706912040710449, 0.009738559722900391, 0.009764512062072754, 0.009703807830810547, 0.00963584041595459, 0.009580256462097168, 0.009599264144897462, 0.00960307216644287, 0.009569536209106446, 0.009566783905029297, 0.00950496006011963, 0.009641759872436524, 0.009528927803039551, 0.0094684476852417, 0.009706656455993652, 0.009982912063598633, 0.009622655868530273, 0.009667136192321778, 0.009652159690856933, 0.009632191658020019, 0.00969529628753662, 0.009756544113159179, 0.009693183898925782, 0.009708864212036133, 0.00965401554107666, 0.009642208099365234, 0.00963798427581787, 0.009586560249328612, 0.009584799766540527, 0.00955452823638916, 0.009522527694702148, 0.00953321647644043, 0.009482368469238282, 0.009499263763427735, 0.009439359664916993, 0.009482239723205567, 0.009531519889831542, 0.00957049560546875, 0.009444543838500977, 0.009574912071228027, 0.009498623847961426, 0.009477215766906738, 0.009527199745178223, 0.009548704147338867, 0.00949187183380127, 0.009691712379455566, 0.009611200332641602, 0.009554112434387207, 0.00957049560546875, 0.009463616371154785, 0.00952889633178711, 0.009559712409973144, 0.009497376441955566, 0.00961945629119873, 0.009672736167907715, 0.009834176063537597, 0.009779040336608886, 0.009763263702392578, 0.009883232116699218, 0.009333087921142577, 0.00965766429901123, 0.009728704452514648, 0.00973583984375, 0.009791839599609374, 0.010076160430908204, 0.009791487693786622, 0.009736448287963868, 0.009621312141418457, 0.009738176345825196, 0.00990822410583496, 0.009815199851989746, 0.009700127601623535, 0.009820223808288574, 0.009844736099243164, 0.009727999687194825, 0.009672767639160156, 0.009756223678588867, 0.009771391868591309, 0.009772992134094239, 0.009874784469604492, 0.009904864311218262, 0.009795743942260743, 0.009840767860412597, 0.009761631965637207, 0.009865856170654297, 0.009761216163635254, 0.009797087669372559, 0.009746784210205079, 0.009711615562438965, 0.00971776008605957, 0.009672287940979005, 0.009761343955993653, 0.009731936454772949, 0.009594047546386719, 0.009575231552124024, 0.009521056175231933, 0.00959273624420166, 0.009543264389038086, 0.009558815956115723, 0.009531455993652344, 0.009545472145080567, 0.009571423530578613, 0.009505696296691894, 0.009676256179809571, 0.009631808280944825, 0.00972060775756836, 0.009700480461120606, 0.009706048011779786, 0.009653599739074707, 0.009653120040893555, 0.009598112106323243, 0.009710207939147949, 0.009540736198425293, 0.009595775604248047, 0.009674752235412597, 0.009569279670715332, 0.00951097583770752, 0.009465056419372559, 0.009446720123291015, 0.009476511955261231, 0.009577983856201172, 0.009595392227172851, 0.009239904403686524, 0.009540255546569825, 0.009547200202941895, 0.009482815742492675, 0.00951427173614502, 0.009537728309631347, 0.009586496353149414, 0.009578495979309083, 0.009547967910766602, 0.009552160263061523, 0.009576383590698243, 0.009619839668273926, 0.009606911659240722, 0.009691328048706054, 0.009698719978332519, 0.009806336402893067, 0.00965340805053711, 0.009854144096374511, 0.009723744392395019, 0.009740351676940918, 0.009783072471618652, 0.00978313636779785, 0.009615584373474121, 0.009639936447143555, 0.009508864402770996, 0.009486335754394531, 0.009538592338562012, 0.010194111824035644, 0.009602784156799316, 0.010584416389465333, 0.009709280014038086, 0.010919103622436524, 0.009595775604248047, 0.009570240020751953, 0.009543680191040039, 0.009529343605041504, 0.009500672340393066, 0.009517215728759765, 0.009490592002868652, 0.00948192024230957, 0.009484416007995605, 0.009461600303649903, 0.009410592079162597, 0.009459712028503419, 0.009705216407775878, 0.009410431861877441, 0.009435296058654785, 0.009408672332763672, 0.009432576179504394, 0.009409279823303222, 0.009449440002441406, 0.009489343643188477, 0.009435263633728028, 0.009345088005065918, 0.009380576133728027, 0.009381855964660644, 0.00942905616760254, 0.009455360412597657, 0.009363679885864258, 0.009416704177856445, 0.009450943946838378, 0.009417344093322753, 0.009479583740234375, 0.009153792381286621, 0.009463871955871582, 0.009493247985839844, 0.009561823844909669, 0.009521439552307128, 0.009453568458557129, 0.009467904090881347, 0.009397503852844238, 0.009399040222167969, 0.009575679779052734, 0.00956492805480957, 0.01002905559539795, 0.010358943939208984, 0.009647968292236329, 0.009621631622314454, 0.009529215812683105, 0.01051206398010254, 0.00967523193359375, 0.00961315155029297, 0.009598976135253906, 0.00970956802368164, 0.009591072082519532, 0.009576064109802246, 0.009640031814575196, 0.009443327903747559, 0.009504799842834472, 0.009521120071411133, 0.009453568458557129, 0.009465855598449707, 0.009466143608093261, 0.009469663619995117, 0.0094269437789917, 0.009799679756164551, 0.009549823760986328, 0.009744511604309083, 0.00953331184387207, 0.009433279991149903, 0.009445183753967286, 0.009465408325195313, 0.009503007888793946, 0.009449631690979003, 0.009445376396179199, 0.009414655685424805, 0.009391679763793945, 0.00943558406829834, 0.00941055965423584, 0.009531392097473144, 0.009400256156921387, 0.009432703971862792, 0.009427007675170898, 0.009468255996704101, 0.009432255744934082, 0.009419615745544434, 0.009453120231628418, 0.009413056373596192, 0.009462016105651855, 0.009608960151672364, 0.009394207954406738, 0.009431008338928223, 0.009375231742858887, 0.009355584144592284, 0.009371904373168946, 0.009400159835815429, 0.009089088439941407, 0.00941868782043457, 0.00937990379333496, 0.009520511627197265, 0.009367551803588867, 0.009364255905151367, 0.009434975624084473, 0.00942080020904541, 0.00941055965423584, 0.009392127990722657, 0.009465951919555664, 0.009379743576049804, 0.009383328437805176, 0.009396832466125488, 0.0094651517868042, 0.009476160049438477, 0.009392767906188966, 0.009380096435546876, 0.009408448219299316, 0.009378687858581543, 0.009448160171508789, 0.009445599555969238, 0.009660415649414063, 0.009490431785583496, 0.009516127586364746, 0.00942307186126709, 0.00946662425994873, 0.009393216133117676, 0.009413503646850585, 0.009426591873168945, 0.009359711647033692, 0.009447423934936524, 0.009539584159851074, 0.009402112007141113, 0.009417023658752442, 0.009500032424926757, 0.00936415958404541, 0.00945081615447998, 0.00944595241546631, 0.009459327697753907, 0.00941811180114746, 0.009458687782287598, 0.00949884796142578, 0.009457440376281738, 0.009502752304077149, 0.00949187183380127, 0.009359935760498046, 0.009407648086547852, 0.009374784469604493, 0.009389920234680176, 0.009488320350646973, 0.00941055965423584, 0.009414112091064453, 0.009381952285766602, 0.009384415626525878, 0.009381888389587402, 0.009561471939086914, 0.009379520416259765, 0.009413567543029785, 0.009455360412597657, 0.009412416458129883, 0.009456064224243164, 0.009404704093933106]",tokens/s,104.16990163816664,,