diff --git "a/perf-df-gptq-1xA10.csv" "b/perf-df-gptq-1xA10.csv" --- "a/perf-df-gptq-1xA10.csv" +++ "b/perf-df-gptq-1xA10.csv" @@ -340,6 +340,33 @@ ChildProcessError: Traceback (most recent call last): OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,2101.9648,5971.116032,0.0,5324.668928,5080.87552,s,10,3.1623598327636717,0.31623598327636715,0.00024235336102111678,0.3162159729003906,0.3165054656982422,0.31660202178955077,0.31667926666259766,"[0.31669857788085937, 0.3162686462402344, 0.3164840087890625, 0.31633709716796876, 0.3158043518066406, 0.31612255859375, 0.3161309814453125, 0.3161632995605469, 0.31597808837890623, 0.31637222290039063]",tokens/s,809.5220453653268,kWh,3.7342462792164754e-06,2.046175760932556e-06,1.8911707837687697e-05,2.4692129877836726e-05,tokens/kWh,10367675.905908046,MB,2101.9648,5971.116032,0.0,5324.668928,5290.117632,s,10,173.26880078125,17.326880078125,0.0008046980240511191,17.3267978515625,17.327645117187497,17.32819267578125,17.32863072265625,"[17.3270703125, 17.32675, 17.326052734375, 17.327296875, 17.3275234375, 17.32606640625, 17.3265234375, 17.325931640625, 17.326845703125, 17.328740234375]",tokens/s,3.6359690674801186,kWh,0.0002045667352755037,0.00011211794793604098,0.0010186654833345132,0.0013353501665460579,tokens/kWh,47178.636419353796,,s,629,175.83934860229496,0.2795538133581795,0.03756451887205511,0.27501873779296876,0.2752235412597656,0.27533250122070313,0.591240341796875,"[0.275198974609375, 0.2747781066894531, 0.2748385314941406, 0.2747607116699219, 0.27491326904296876, 0.2749624328613281, 0.27463168334960936, 0.27481497192382814, 0.274840576171875, 0.2751907958984375, 0.27510272216796877, 0.2749347839355469, 0.27483544921875, 0.2749849548339844, 0.2753403015136719, 0.27486309814453125, 0.27506381225585935, 0.2750320739746094, 0.2749327392578125, 0.27479449462890626, 0.27507815551757814, 0.27491122436523435, 0.2749091796875, 0.2749859924316406, 0.2749819030761719, 0.2750750732421875, 0.2750402526855469, 0.27504434204101563, 0.2749245300292969, 0.2748487548828125, 0.2750873718261719, 0.27546316528320314, 0.2752255859375, 0.2750679016113281, 0.2751549377441406, 0.2749347839355469, 0.27496038818359375, 0.2750351257324219, 0.2749020080566406, 0.27519692993164063, 0.27499520874023436, 0.2750474243164063, 0.2749327392578125, 0.27501669311523436, 0.27505767822265625, 0.27501055908203126, 0.27503411865234373, 0.2750894165039062, 0.2751805419921875, 0.27512115478515625, 0.27502081298828124, 0.2750965881347656, 0.27512216186523436, 0.2753597412109375, 0.2751068115234375, 0.275093505859375, 0.27523275756835935, 0.275135498046875, 0.2751375427246094, 0.27496551513671874, 0.27523583984375, 0.27499520874023436, 0.5913477172851562, 0.274830322265625, 0.2748323974609375, 0.27500338745117187, 0.27521127319335936, 0.27528704833984374, 0.27483544921875, 0.2747678833007812, 0.27484979248046876, 0.2747781066894531, 0.274882568359375, 0.2750955505371094, 0.2749358215332031, 0.2748938293457031, 0.2748334045410156, 0.27500851440429686, 0.27480267333984376, 0.27513958740234373, 0.27495013427734377, 0.27490304565429685, 0.275267578125, 0.2749644775390625, 0.27491326904296876, 0.2748948364257813, 0.27508120727539065, 0.2749266052246094, 0.27535052490234374, 0.2749358215332031, 0.2749992980957031, 0.2749921264648437, 0.27484774780273435, 0.27500543212890627, 0.27491943359375, 0.2750003051757812, 0.2750115966796875, 0.2751180725097656, 0.27498699951171873, 0.27493170166015624, 0.2749542541503906, 0.2748968811035156, 0.2757120056152344, 0.2750218200683594, 0.2752235412597656, 0.2749921264648437, 0.2749245300292969, 0.27508428955078124, 0.27499520874023436, 0.2750064697265625, 0.2751129455566406, 0.27507403564453126, 0.2751037292480469, 0.2750556030273438, 0.27498086547851563, 0.27507098388671875, 0.27502590942382815, 0.27505255126953126, 0.27506585693359376, 0.2751016845703125, 0.2752962646484375, 0.27508837890625, 0.275515380859375, 0.2752000122070313, 0.2751385498046875, 0.5916641235351563, 0.27478631591796876, 0.27482009887695313, 0.27491226196289065, 0.2747607116699219, 0.27492556762695314, 0.2749245300292969, 0.27482110595703124, 0.274798583984375, 0.27490609741210936, 0.2748671875, 0.27493991088867187, 0.27506689453125, 0.27507302856445315, 0.2748938293457031, 0.27494708251953126, 0.2748016662597656, 0.2749941711425781, 0.27487640380859374, 0.2748692626953125, 0.27501055908203126, 0.27505050659179686, 0.2749010009765625, 0.2748334045410156, 0.27492556762695314, 0.274956298828125, 0.2749706115722656, 0.2750597229003906, 0.2748968811035156, 0.27507916259765625, 0.27507199096679685, 0.27513037109375, 0.27511911010742185, 0.2748968811035156, 0.274951171875, 0.27523992919921875, 0.27490304565429685, 0.2749982604980469, 0.2750699462890625, 0.27497369384765624, 0.27497369384765624, 0.27502490234375, 0.27506381225585935, 0.275009521484375, 0.2749306945800781, 0.27501055908203126, 0.2751016845703125, 0.27510272216796877, 0.275230712890625, 0.27520306396484373, 0.27552459716796873, 0.27511398315429686, 0.2750894165039062, 0.2750750732421875, 0.2749798278808594, 0.27514981079101564, 0.27515802001953127, 0.2751447143554687, 0.2750351257324219, 0.27518875122070313, 0.27545803833007815, 0.27511911010742185, 0.2752235412597656, 0.5914030151367188, 0.27480679321289064, 0.27478631591796876, 0.27494091796875, 0.27484262084960936, 0.27500851440429686, 0.274798583984375, 0.2749900817871094, 0.274830322265625, 0.27475662231445314, 0.2748590087890625, 0.27493374633789064, 0.2748610534667969, 0.27496038818359375, 0.2748938293457031, 0.2751068115234375, 0.27488357543945313, 0.275167236328125, 0.274872314453125, 0.274840576171875, 0.27497369384765624, 0.27494503784179686, 0.2751068115234375, 0.2749245300292969, 0.27491021728515624, 0.27501873779296876, 0.27484262084960936, 0.2748876647949219, 0.27505255126953126, 0.27506381225585935, 0.27500851440429686, 0.2750965881347656, 0.27498291015625, 0.27501669311523436, 0.27518362426757814, 0.2751795349121094, 0.2750402526855469, 0.27513446044921874, 0.2750003051757812, 0.27497268676757813, 0.27517132568359376, 0.2752593994140625, 0.27508428955078124, 0.27551028442382813, 0.2751467590332031, 0.27510784912109376, 0.2750074768066406, 0.2751068115234375, 0.27511602783203126, 0.2750750732421875, 0.2751539306640625, 0.2751119384765625, 0.2749624328613281, 0.2750822448730469, 0.2752163696289063, 0.275198974609375, 0.2751201171875, 0.27526144409179687, 0.27553280639648436, 0.2751682434082031, 0.2751150207519531, 0.2751877136230469, 0.2750679016113281, 0.5908869018554688, 0.2749245300292969, 0.27489279174804687, 0.27481292724609374, 0.2747852783203125, 0.27498699951171873, 0.2748948364257813, 0.27476171875, 0.27514776611328123, 0.2748016662597656, 0.2749440002441406, 0.27507098388671875, 0.274988037109375, 0.27523175048828125, 0.2749696044921875, 0.27488461303710937, 0.2750412902832031, 0.27491226196289065, 0.2748374938964844, 0.2750269470214844, 0.2748794860839844, 0.2749992980957031, 0.27503103637695314, 0.2748118896484375, 0.2750074768066406, 0.27487539672851563, 0.27500851440429686, 0.27484262084960936, 0.27503616333007813, 0.27514981079101564, 0.274988037109375, 0.2754109497070312, 0.275030029296875, 0.2749224853515625, 0.27513037109375, 0.2750474243164063, 0.2750494689941406, 0.27498904418945314, 0.27496038818359375, 0.27510580444335936, 0.27506689453125, 0.2750975952148437, 0.275125244140625, 0.27500338745117187, 0.27499725341796877, 0.2753341369628906, 0.27504638671875, 0.2751037292480469, 0.2751518859863281, 0.2750351257324219, 0.2752010192871094, 0.2751590270996094, 0.275209228515625, 0.2754037780761719, 0.2751528930664062, 0.27510272216796877, 0.2750330810546875, 0.27525222778320313, 0.27512832641601564, 0.27523275756835935, 0.27520306396484373, 0.27509451293945314, 0.275167236328125, 0.591182861328125, 0.2749153137207031, 0.27479244995117186, 0.2752962646484375, 0.27491738891601564, 0.2749798278808594, 0.2749900817871094, 0.2748671875, 0.27480780029296875, 0.27507302856445315, 0.27497882080078123, 0.2749624328613281, 0.2749849548339844, 0.27487335205078123, 0.27491943359375, 0.27487026977539064, 0.27495834350585935, 0.2748538818359375, 0.27482009887695313, 0.27472076416015623, 0.2749286499023437, 0.27513345336914063, 0.27491738891601564, 0.27482827758789063, 0.2753013610839844, 0.27489791870117186, 0.2749071350097656, 0.274956298828125, 0.27478427124023436, 0.2748241882324219, 0.2750679016113281, 0.2749542541503906, 0.2750382080078125, 0.2749859924316406, 0.2752041015625, 0.2751119384765625, 0.27506689453125, 0.274914306640625, 0.2751201171875, 0.2753064880371094, 0.2750136413574219, 0.27502285766601564, 0.2750218200683594, 0.2749214782714844, 0.27502899169921874, 0.2750044250488281, 0.275093505859375, 0.27505767822265625, 0.2751201171875, 0.27507815551757814, 0.27517031860351565, 0.2749706115722656, 0.2751231994628906, 0.27507403564453126, 0.2749992980957031, 0.2751016845703125, 0.2751488037109375, 0.275135498046875, 0.2754703369140625, 0.27516619873046877, 0.27518362426757814, 0.27513958740234373, 0.27514163208007814, 0.5912626953125, 0.2746531982421875, 0.2748385314941406, 0.27486309814453125, 0.27493887329101563, 0.2749706115722656, 0.2748252258300781, 0.2748938293457031, 0.2749163513183594, 0.2748385314941406, 0.2748856201171875, 0.27499725341796877, 0.2751129455566406, 0.27485592651367186, 0.27486209106445314, 0.2748907470703125, 0.2748590087890625, 0.274882568359375, 0.2749521789550781, 0.2750556030273438, 0.27492352294921873, 0.2749521789550781, 0.2749593505859375, 0.27478936767578127, 0.2749358215332031, 0.2749266052246094, 0.2748938293457031, 0.2749296569824219, 0.2749245300292969, 0.274956298828125, 0.2753966064453125, 0.2750392456054688, 0.2749747314453125, 0.2751201171875, 0.27508428955078124, 0.27516928100585936, 0.275093505859375, 0.2749767761230469, 0.27500543212890627, 0.2749276123046875, 0.27500851440429686, 0.27505050659179686, 0.27517031860351565, 0.27502386474609375, 0.27498291015625, 0.2751324157714844, 0.2752061462402344, 0.27502081298828124, 0.275557373046875, 0.27510272216796877, 0.2751098937988281, 0.27516415405273437, 0.27508120727539065, 0.2750771179199219, 0.27513446044921874, 0.27501773071289065, 0.27513650512695315, 0.27518875122070313, 0.2750699462890625, 0.2751385498046875, 0.2755860595703125, 0.27517031860351565, 0.27525531005859377, 0.5912658081054688, 0.27493170166015624, 0.27482827758789063, 0.2749491271972656, 0.27483544921875, 0.27496038818359375, 0.2748876647949219, 0.27492556762695314, 0.2748098449707031, 0.2748590087890625, 0.274946044921875, 0.27496038818359375, 0.2749358215332031, 0.2749859924316406, 0.27488153076171873, 0.2748692626953125, 0.27488461303710937, 0.2749286499023437, 0.27498291015625, 0.2748989562988281, 0.27534335327148435, 0.2750320739746094, 0.27485592651367186, 0.2748610534667969, 0.27494091796875, 0.2748876647949219, 0.2748487548828125, 0.27498086547851563, 0.27499725341796877, 0.27497369384765624, 0.27501873779296876, 0.2750269470214844, 0.2748538818359375, 0.27500851440429686, 0.2749931640625, 0.2751590270996094, 0.27493374633789064, 0.274951171875, 0.2750494689941406, 0.27503717041015624, 0.2751231994628906, 0.2754232177734375, 0.27509451293945314, 0.2750013427734375, 0.27502490234375, 0.2750064697265625, 0.2748436584472656, 0.27507916259765625, 0.2750965881347656, 0.27501055908203126, 0.27513650512695315, 0.2751047668457031, 0.27512728881835935, 0.275093505859375, 0.27511602783203126, 0.2750648193359375, 0.275061767578125, 0.27523175048828125, 0.275167236328125, 0.2750986328125, 0.2752481384277344, 0.27527679443359376, 0.2751201171875, 0.5914869995117188, 0.2747115478515625, 0.27491122436523435, 0.2747750549316406, 0.27482110595703124, 0.2748252258300781, 0.2748968811035156, 0.27479962158203125, 0.2747607116699219, 0.27496038818359375, 0.274872314453125, 0.27533004760742186, 0.27499520874023436, 0.2750648193359375, 0.2749327392578125, 0.27499520874023436, 0.275009521484375, 0.27496755981445314, 0.2748528747558594, 0.274766845703125, 0.2749644775390625, 0.2749266052246094, 0.2748805236816406, 0.2747740173339844, 0.27503411865234373, 0.2748190612792969, 0.27482827758789063, 0.27507916259765625, 0.27513650512695315, 0.27535565185546873, 0.27501873779296876, 0.2749849548339844, 0.27530853271484373, 0.2750218200683594, 0.2750545959472656, 0.27512832641601564, 0.2749819030761719, 0.27498904418945314, 0.27506381225585935, 0.27501055908203126, 0.275061767578125, 0.2750218200683594, 0.2751016845703125, 0.2750802001953125, 0.27516110229492186, 0.2751426696777344, 0.2751098937988281, 0.2752634887695313, 0.275267578125, 0.27502285766601564, 0.2752563171386719, 0.27503103637695314, 0.2752204895019531, 0.2750484619140625, 0.27531674194335937, 0.2751047668457031, 0.275135498046875, 0.27513446044921874, 0.27511602783203126, 0.27506585693359376, 0.2751242370605469, 0.2751047668457031, 0.2751129455566406, 0.5915084838867187, 0.27509451293945314, 0.2749163513183594, 0.27489279174804687, 0.27486004638671874, 0.2749900817871094, 0.27495321655273436, 0.2748989562988281, 0.2748374938964844, 0.27503717041015624, 0.27489791870117186, 0.2749859924316406, 0.2748805236816406, 0.2748518371582031, 0.27495834350585935, 0.27506072998046877, 0.27489791870117186, 0.27492044067382815, 0.2749245300292969, 0.2749839477539063, 0.2749910888671875, 0.2749491271972656, 0.2753484802246094, 0.274946044921875, 0.27497882080078123, 0.27499520874023436, 0.2748989562988281, 0.2750648193359375, 0.2751231994628906, 0.27494503784179686, 0.2751129455566406, 0.27514163208007814, 0.2752122802734375, 0.27512728881835935, 0.2750433349609375, 0.275262451171875, 0.27510272216796877, 0.27497369384765624, 0.2752850036621094, 0.2749706115722656, 0.27505255126953126, 0.27513650512695315, 0.27524197387695315, 0.27487539672851563, 0.27529727172851565, 0.2754805908203125, 0.2750269470214844, 0.27508428955078124, 0.27512933349609375, 0.27514163208007814, 0.2751528930664062, 0.27501055908203126, 0.2751600646972656, 0.27512728881835935, 0.27509963989257813, 0.27514163208007814, 0.27519488525390623, 0.27512728881835935, 0.27538534545898435, 0.2750556030273438, 0.27517745971679686, 0.2751447143554687, 0.2751426696777344]",tokens/s,3.577128811041278,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -560,48 +587,33 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpf3j99m06/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 86.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -702,48 +714,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2_moe,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,mixtral,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1331.101696,1313.34144,0.0,666.894336,583.411712,s,10,0.33520121383667,0.033520121383666994,0.001453907594310569,0.03276590347290039,0.03438638572692871,0.03594858493804932,0.0371983443069458,"[0.037510784149169925, 0.03251792144775391, 0.03257727813720703, 0.03279670333862305, 0.03273510360717773, 0.0338130874633789, 0.0339708480834961, 0.03403923034667969, 0.03259091186523438, 0.03264934539794922]",tokens/s,7637.2038474997435,kWh,3.937127772304747e-07,2.1573539868950317e-07,1.2275167227534153e-06,1.836964898673393e-06,tokens/kWh,139360311.23124692,MB,1331.101696,1313.34144,0.0,666.894336,644.0192,s,10,20.422793212890625,2.0422793212890626,0.020881081923259662,2.038978942871094,2.0718335693359373,2.0766283325195314,2.0804641430664064,"[2.05873486328125, 2.0230191650390625, 2.022578857421875, 2.030239990234375, 2.07076806640625, 2.081423095703125, 2.038813720703125, 2.0391441650390627, 2.013723388671875, 2.044347900390625]",tokens/s,30.847886155080463,kWh,2.5016873122917285e-05,1.3707734888537744e-05,7.079942978764167e-05,0.0001095240377990967,tokens/kWh,575216.1924084906,,s,629,20.68533759689332,0.03288606931143611,0.004037093144175609,0.032176128387451174,0.03341967468261719,0.033662213134765626,0.06502883392333984,"[0.033104896545410156, 0.03315814590454102, 0.0339947509765625, 0.03458047866821289, 0.03360153579711914, 0.03348889541625977, 0.03379516983032226, 0.03395471954345703, 0.034119678497314454, 0.033826911926269534, 0.033135520935058595, 0.03497574234008789, 0.032911422729492185, 0.03253753662109375, 0.03308441543579101, 0.03468185424804687, 0.033547264099121094, 0.03344486236572266, 0.03216896057128906, 0.032756736755371094, 0.032584705352783204, 0.031925247192382815, 0.03165603256225586, 0.03363113784790039, 0.03241676712036133, 0.032061439514160156, 0.03212799835205078, 0.033124351501464845, 0.03342950439453125, 0.03297894287109375, 0.03218739318847656, 0.031764480590820314, 0.03209011077880859, 0.03204198455810547, 0.03206655883789063, 0.03209734344482422, 0.03223238372802734, 0.03212799835205078, 0.031920127868652344, 0.03172249603271484, 0.032205825805664064, 0.0320634880065918, 0.032023681640625, 0.03208793640136719, 0.032194561004638675, 0.0322426872253418, 0.03201536178588867, 0.031942655563354495, 0.03224063873291016, 0.03195699119567871, 0.03216998291015625, 0.03210854339599609, 0.03225907135009766, 0.0318884162902832, 0.032253921508789064, 0.03189452743530274, 0.032215038299560544, 0.03224166488647461, 0.03209625625610352, 0.03184435272216797, 0.03227238464355469, 0.03201536178588867, 0.0649543685913086, 0.03204198455810547, 0.03203276824951172, 0.03199897575378418, 0.03233385467529297, 0.032118751525878904, 0.031925247192382815, 0.03201740646362305, 0.03212799835205078, 0.03179520034790039, 0.03220278549194336, 0.0322088623046875, 0.032091136932373046, 0.03206041717529297, 0.03184236717224121, 0.03198764801025391, 0.03176140785217285, 0.03216486358642578, 0.032, 0.03201740646362305, 0.03221196746826172, 0.03222016143798828, 0.032091136932373046, 0.031916032791137694, 0.031753215789794925, 0.03231132888793945, 0.032330718994140625, 0.03216486358642578, 0.032231456756591795, 0.03199279975891113, 0.03186073684692383, 0.0320819206237793, 0.032210945129394535, 0.03181158447265625, 0.03150028800964356, 0.03199897575378418, 0.03202150344848633, 0.032489471435546875, 0.03226419067382812, 0.03231129455566406, 0.032194561004638675, 0.03220377731323242, 0.032314369201660156, 0.03222016143798828, 0.03188531112670898, 0.03210137557983399, 0.03206553649902344, 0.03210649490356445, 0.03198259162902832, 0.032304126739501955, 0.032077823638916016, 0.03200921630859375, 0.03172147178649903, 0.03178188705444336, 0.03171123123168945, 0.031903743743896484, 0.03326873779296875, 0.033317886352539065, 0.03220694351196289, 0.03213711929321289, 0.0322426872253418, 0.03200921630859375, 0.03232153701782227, 0.06520320129394531, 0.03232767868041992, 0.03204198455810547, 0.032148479461669925, 0.03198873519897461, 0.03217407989501953, 0.03191296005249023, 0.032146430969238284, 0.03197849655151367, 0.032277503967285154, 0.031971328735351565, 0.03213926315307617, 0.03194777679443359, 0.03202969741821289, 0.03183001518249512, 0.031865856170654294, 0.032031742095947266, 0.032123905181884765, 0.03240038299560547, 0.032161792755126956, 0.0322754898071289, 0.03221500778198242, 0.03213516616821289, 0.0319815673828125, 0.03259187316894531, 0.03320217514038086, 0.03232665634155273, 0.032246784210205076, 0.031801343917846676, 0.03212799835205078, 0.03236556625366211, 0.03234921646118164, 0.03226313781738281, 0.03233996963500976, 0.0320634880065918, 0.03211471939086914, 0.03225084686279297, 0.03212595367431641, 0.03177676773071289, 0.03220377731323242, 0.031860767364501955, 0.03182383918762207, 0.03216998291015625, 0.031927295684814457, 0.032008190155029294, 0.03208294296264649, 0.031850496292114255, 0.03204915237426758, 0.03201126480102539, 0.03209222412109375, 0.032090049743652344, 0.03199897575378418, 0.032054271697998044, 0.03199078369140625, 0.03199795150756836, 0.0319815673828125, 0.032059391021728514, 0.0321341438293457, 0.03216998291015625, 0.031866880416870115, 0.03199283218383789, 0.0318525447845459, 0.0316866569519043, 0.06526771545410157, 0.0321003532409668, 0.032156673431396485, 0.0322949104309082, 0.03197337532043457, 0.0318474235534668, 0.03226009750366211, 0.032176128387451174, 0.0321607666015625, 0.03189760017395019, 0.03209830474853516, 0.03213824081420898, 0.032189441680908204, 0.03204095840454101, 0.03174092864990234, 0.03217715072631836, 0.03218022537231445, 0.03219046401977539, 0.032061439514160156, 0.03203583908081055, 0.03203481674194336, 0.03219558334350586, 0.03174399948120117, 0.032053249359130856, 0.03205734252929687, 0.032935935974121096, 0.03235635375976562, 0.03200614547729492, 0.03260518264770508, 0.032069633483886716, 0.03273932647705078, 0.03267804718017578, 0.03222000122070313, 0.03172352027893066, 0.032352256774902347, 0.0326297607421875, 0.0321638412475586, 0.031821823120117186, 0.03215468978881836, 0.03231942367553711, 0.03211673736572265, 0.031959039688110355, 0.031958015441894534, 0.032586753845214846, 0.03233484649658203, 0.032471038818359374, 0.03204915237426758, 0.031835168838500974, 0.032327648162841796, 0.03257753753662109, 0.03249868774414062, 0.032210945129394535, 0.03236249542236328, 0.03205836868286133, 0.03227545547485351, 0.032244735717773435, 0.03242700958251953, 0.03219251251220703, 0.031923200607299806, 0.03297689437866211, 0.03294412612915039, 0.03244441604614258, 0.032438270568847655, 0.06516633605957031, 0.03237887954711914, 0.03210649490356445, 0.0328611831665039, 0.03278335952758789, 0.03263283157348633, 0.03236454391479492, 0.03188531112670898, 0.03201536178588867, 0.032143360137939454, 0.03220479965209961, 0.03231846237182617, 0.032036865234375, 0.03224166488647461, 0.031923200607299806, 0.03245363235473633, 0.03258163070678711, 0.03215052795410156, 0.03210444641113281, 0.03236556625366211, 0.032140289306640625, 0.03235123062133789, 0.031850496292114255, 0.03293491363525391, 0.03199488067626953, 0.03168972778320313, 0.034272254943847655, 0.03420774459838867, 0.033530879974365234, 0.03320633697509766, 0.03329529571533203, 0.03347967910766601, 0.0335810546875, 0.033501312255859374, 0.033164161682128906, 0.03283456039428711, 0.03306086349487305, 0.03298611068725586, 0.033023998260498046, 0.03273011016845703, 0.033073150634765625, 0.03321558380126953, 0.0334744644165039, 0.033670143127441404, 0.03373158264160156, 0.033122303009033204, 0.033137664794921876, 0.032949249267578126, 0.03352166366577149, 0.03325235366821289, 0.03329740905761719, 0.03341721725463867, 0.0329431037902832, 0.032920574188232424, 0.03270041656494141, 0.03291449737548828, 0.03361273574829102, 0.0333383674621582, 0.03300966262817383, 0.033314815521240236, 0.03352576065063476, 0.033484798431396484, 0.03335168075561523, 0.06773554992675782, 0.03331584167480469, 0.032792736053466796, 0.03239920043945312, 0.032656383514404294, 0.032487422943115234, 0.032510974884033206, 0.032625663757324216, 0.033486846923828126, 0.0323133430480957, 0.032674816131591795, 0.03250995254516602, 0.03332198333740234, 0.0331673583984375, 0.03303833770751953, 0.032824321746826174, 0.03312844848632813, 0.03281119918823242, 0.03530630493164062, 0.0341288948059082, 0.03366307067871094, 0.033581985473632815, 0.0333496322631836, 0.03298713684082031, 0.03383500671386719, 0.03362406539916992, 0.03323088073730469, 0.033562591552734375, 0.03322777557373047, 0.03339571380615235, 0.033470462799072266, 0.033393665313720705, 0.03320627212524414, 0.03301683044433594, 0.03308854293823242, 0.03340288162231445, 0.03320111846923828, 0.03311206436157227, 0.03325235366821289, 0.03300556945800781, 0.03261337661743164, 0.032723968505859374, 0.032677886962890625, 0.03328521728515625, 0.033381278991699216, 0.03289907073974609, 0.0326901741027832, 0.03343155288696289, 0.033373344421386716, 0.03307913589477539, 0.03322367858886719, 0.033445888519287106, 0.03355753707885742, 0.03334857559204102, 0.033306625366210936, 0.03201126480102539, 0.032074752807617186, 0.03210137557983399, 0.031854591369628905, 0.032043006896972655, 0.03199590492248535, 0.03199590492248535, 0.03232460784912109, 0.06749696350097656, 0.032715774536132815, 0.03205734252929687, 0.031893535614013674, 0.03176547241210938, 0.031662080764770506, 0.031867904663085936, 0.031474687576293944, 0.03173785591125488, 0.03174502372741699, 0.032121856689453124, 0.032102401733398435, 0.032121856689453124, 0.03208396911621094, 0.03197644805908203, 0.0320706558227539, 0.031995071411132815, 0.032895809173583986, 0.03338956832885742, 0.03222630310058594, 0.03171327972412109, 0.032189441680908204, 0.03259699249267578, 0.03214745712280274, 0.03189555168151856, 0.03158844757080078, 0.031835039138793944, 0.03214745712280274, 0.03194879913330078, 0.032132095336914065, 0.03190784072875977, 0.031838207244873046, 0.03188531112670898, 0.03211468887329102, 0.031869951248168944, 0.032113662719726564, 0.031850496292114255, 0.03192934417724609, 0.03216281509399414, 0.03153203201293946, 0.03204095840454101, 0.03206659317016602, 0.032035808563232425, 0.0327116813659668, 0.03444121551513672, 0.033683456420898435, 0.032330753326416016, 0.03154944038391113, 0.031768575668334964, 0.03169075202941894, 0.032817150115966795, 0.03294105529785156, 0.03330559921264648, 0.03367731094360352, 0.03334463882446289, 0.03294502258300781, 0.033154048919677735, 0.03304246520996094, 0.033652702331542966, 0.033350654602050785, 0.033121280670166016, 0.033290241241455076, 0.033051647186279294, 0.06784102630615234, 0.03334348678588867, 0.03323187255859375, 0.03275980758666992, 0.03303833770751953, 0.033274913787841795, 0.03332502365112305, 0.03250380706787109, 0.03273420715332031, 0.033035263061523434, 0.0333568000793457, 0.034625537872314455, 0.033642528533935546, 0.033573856353759764, 0.0330885124206543, 0.03200723266601563, 0.03196735954284668, 0.031814559936523434, 0.0320357437133789, 0.03220479965209961, 0.03211468887329102, 0.03197644805908203, 0.0320819206237793, 0.03228979110717774, 0.03189049530029297, 0.03215865707397461, 0.03181465530395508, 0.032048126220703126, 0.03180544090270996, 0.031985727310180664, 0.03187910461425781, 0.032320510864257815, 0.03181670379638672, 0.03186892890930176, 0.03202560043334961, 0.03222937774658203, 0.03159452819824219, 0.03166102409362793, 0.032102401733398435, 0.03137347221374512, 0.03133219146728516, 0.03190169525146484, 0.031926271438598636, 0.03193446350097656, 0.03205542373657227, 0.03184217643737793, 0.03156889533996582, 0.03164985656738281, 0.03224262237548828, 0.03217407989501953, 0.03388729476928711, 0.03340281677246094, 0.03364863967895508, 0.032742401123046876, 0.03202764892578125, 0.0319109115600586, 0.03200921630859375, 0.031916032791137694, 0.03180748748779297, 0.031916032791137694, 0.03242086410522461, 0.03213516616821289, 0.03232563018798828, 0.06490930938720703, 0.03211775970458984, 0.03203177642822266, 0.03207574462890625, 0.032105472564697264, 0.03196518325805664, 0.031838207244873046, 0.03207372665405273, 0.03203379058837891, 0.03212192153930664, 0.031962047576904296, 0.03213312149047851, 0.03176140785217285, 0.031899648666381834, 0.03203481674194336, 0.03176243209838867, 0.03202560043334961, 0.03242092895507812, 0.03222214508056641, 0.03180646324157715, 0.03202560043334961, 0.031848512649536136, 0.03206444931030274, 0.03233484649658203, 0.03199590492248535, 0.03216486358642578, 0.03189871978759766, 0.03182275199890137, 0.032094207763671875, 0.03187302398681641, 0.03178291130065918, 0.031719423294067385, 0.031409151077270506, 0.032059391021728514, 0.03167334365844727, 0.03156991958618164, 0.031459327697753905, 0.031955968856811526, 0.031752191543579104, 0.031849472045898435, 0.032089088439941404, 0.03199590492248535, 0.03220991897583008, 0.03219353485107422, 0.032301055908203126, 0.03199708747863769, 0.03167932891845703, 0.031663103103637694, 0.031920127868652344, 0.03268921661376953, 0.031928255081176755, 0.031344671249389645, 0.031846368789672855, 0.03184745597839356, 0.031919071197509766, 0.03223551940917969, 0.032146430969238284, 0.03202252960205078, 0.032074752807617186, 0.031932416915893554, 0.031835136413574217, 0.03167948722839355, 0.0321607666015625, 0.06505779266357421, 0.03212086486816406, 0.03157814407348633, 0.0318535041809082, 0.03200102233886719, 0.03226931381225586, 0.03219968032836914, 0.0319815673828125, 0.0331673583984375, 0.03325235366821289, 0.03343974304199219, 0.03346435165405273, 0.032823265075683596, 0.03463884735107422, 0.03484467315673828, 0.03366092681884766, 0.03305574417114258, 0.03280998229980469, 0.03216588973999023, 0.031710208892822264, 0.03217510223388672, 0.032246784210205076, 0.03222016143798828, 0.03241779327392578, 0.03233996963500976, 0.03221196746826172, 0.032176128387451174, 0.0317573127746582, 0.03218534469604492, 0.03207987213134766, 0.03217203140258789, 0.03193139266967773, 0.03201849746704102, 0.03191494369506836, 0.0320552978515625, 0.03230515289306641, 0.03226828765869141, 0.03278847885131836, 0.03237273788452148, 0.03205836868286133, 0.03208294296264649, 0.032156673431396485, 0.03226521682739258, 0.03197542381286621, 0.0322979850769043, 0.03211673736572265, 0.032233470916748046, 0.03278335952758789, 0.032421886444091795, 0.032421886444091795, 0.03249868774414062, 0.03187609672546387, 0.032476158142089845, 0.032279552459716795, 0.032189441680908204, 0.031854591369628905, 0.032369663238525394, 0.03227340698242188, 0.032464897155761716, 0.03206553649902344, 0.03333529663085937, 0.033263614654541016, 0.0335206413269043]",tokens/s,30.408012296326657,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -1268,48 +1265,33 @@ ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please r ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1273.15968,1017.643008,0.0,371.195904,315.446272,s,10,0.17996480178833008,0.017996480178833006,0.0008474778858308205,0.017858192443847655,0.019447183799743653,0.019550215816497803,0.01963264142990112,"[0.019653247833251952, 0.018108352661132813, 0.017151264190673827, 0.01790598487854004, 0.017810400009155274, 0.01942428779602051, 0.01810348892211914, 0.017385663986206053, 0.017283935546875, 0.01713817596435547]",tokens/s,14225.003859426943,kWh,2.0975892374871728e-07,1.1493715032924598e-07,4.918131031402125e-07,8.165091772181758e-07,tokens/kWh,313529850.17533416,MB,1273.15968,1017.643008,0.0,371.195904,340.901888,s,10,10.844099975585937,1.084409997558594,0.02124332077999351,1.0841959228515625,1.1139094970703125,1.113964208984375,1.114007978515625,"[1.0721380615234375, 1.0624615478515624, 1.086800048828125, 1.081591796875, 1.10824365234375, 1.1140189208984375, 1.1138973388671876, 1.058116455078125, 1.055816650390625, 1.0910155029296875]",tokens/s,58.096107691589154,kWh,1.2963560541835509e-05,7.103481325002172e-06,2.7866098472065074e-05,4.793314033890275e-05,tokens/kWh,1314330.7439189192,,s,629,10.984805355072023,0.01746391948342134,0.00222119169666076,0.017201152801513672,0.01777562217712402,0.01790013427734375,0.03446820938110352,"[0.019532800674438477, 0.017575935363769533, 0.017661951065063478, 0.017559551239013673, 0.017477632522583008, 0.01739776039123535, 0.01701683235168457, 0.01681715202331543, 0.016644096374511717, 0.01683148765563965, 0.016921600341796874, 0.016746496200561522, 0.016950271606445313, 0.016953344345092772, 0.01695232009887695, 0.016910335540771485, 0.01681510353088379, 0.016473247528076173, 0.016454496383666993, 0.01641472053527832, 0.016499711990356446, 0.016715776443481444, 0.01696767997741699, 0.0168724479675293, 0.016736255645751954, 0.016712703704833985, 0.01675468826293945, 0.016615423202514648, 0.01681001663208008, 0.01688265609741211, 0.016876544952392578, 0.01680486488342285, 0.016647167205810547, 0.0168407039642334, 0.017120256423950195, 0.017632255554199217, 0.01778278350830078, 0.01756159973144531, 0.017524736404418945, 0.017369087219238282, 0.01721343994140625, 0.017161216735839844, 0.01666662406921387, 0.016910335540771485, 0.0168724479675293, 0.01687558364868164, 0.016836544036865235, 0.01682329559326172, 0.01680691146850586, 0.016847999572753906, 0.01660198402404785, 0.016702463150024414, 0.01681100845336914, 0.016859136581420898, 0.016879615783691407, 0.01701068878173828, 0.01700556755065918, 0.016697439193725586, 0.016431007385253906, 0.016690176010131837, 0.01664614486694336, 0.017131519317626954, 0.0360447998046875, 0.017485824584960938, 0.01701683235168457, 0.017459199905395507, 0.017299455642700197, 0.01701171112060547, 0.016887807846069337, 0.016791711807250975, 0.016951135635375977, 0.016733184814453125, 0.016893951416015626, 0.017906688690185548, 0.017275903701782228, 0.016907264709472656, 0.016884735107421875, 0.016707584381103514, 0.016904191970825197, 0.016856096267700196, 0.016788448333740234, 0.01700761604309082, 0.016861183166503906, 0.016882848739624024, 0.016833375930786133, 0.017093631744384767, 0.016793600082397463, 0.016761856079101564, 0.016498687744140626, 0.016457727432250976, 0.0166430721282959, 0.016849920272827147, 0.01663385581970215, 0.01641276741027832, 0.016531455993652345, 0.016519071578979493, 0.01657753562927246, 0.016689151763916017, 0.016747520446777343, 0.01677107238769531, 0.0166748161315918, 0.016792575836181642, 0.016821247100830078, 0.0165980167388916, 0.01677004814147949, 0.016877567291259766, 0.01684377670288086, 0.016798751831054688, 0.016947168350219727, 0.01685196876525879, 0.01683046340942383, 0.016855039596557618, 0.01677824020385742, 0.016854015350341797, 0.01680998420715332, 0.016873472213745116, 0.01701888084411621, 0.01681100845336914, 0.01683251190185547, 0.016942079544067384, 0.01684787178039551, 0.016751615524291993, 0.016739328384399413, 0.016631967544555665, 0.01671561622619629, 0.034219009399414066, 0.016704511642456055, 0.016874496459960937, 0.016935935974121095, 0.016522239685058594, 0.016488447189331054, 0.016662527084350585, 0.01700351905822754, 0.018149375915527344, 0.018177024841308592, 0.01762099266052246, 0.017268896102905273, 0.01789423942565918, 0.017338367462158204, 0.017286144256591796, 0.017512447357177736, 0.017254400253295898, 0.01664614486694336, 0.016914432525634765, 0.017583103179931642, 0.01720319938659668, 0.016671743392944336, 0.017498111724853514, 0.017228799819946287, 0.017484800338745117, 0.017487871170043946, 0.017913856506347657, 0.01744486427307129, 0.017209344863891602, 0.016825344085693358, 0.01761689567565918, 0.01739571189880371, 0.01756876754760742, 0.01745408058166504, 0.017344512939453126, 0.01725644874572754, 0.01766092872619629, 0.017331199645996095, 0.01756979179382324, 0.017480703353881837, 0.016858112335205077, 0.01682329559326172, 0.01677824020385742, 0.016903167724609376, 0.017378303527832033, 0.017529855728149413, 0.017398784637451172, 0.01740492820739746, 0.017079296112060546, 0.01699430465698242, 0.01744486427307129, 0.017386495590209963, 0.017386495590209963, 0.0174202880859375, 0.016942079544067384, 0.017035263061523438, 0.017382400512695313, 0.01723289680480957, 0.017343488693237305, 0.01724825668334961, 0.017131519317626954, 0.01740595245361328, 0.016730112075805666, 0.035399776458740234, 0.016864160537719726, 0.016750591278076172, 0.017101823806762697, 0.017321983337402345, 0.0174335994720459, 0.01740287971496582, 0.01740083122253418, 0.01743052864074707, 0.017342464447021484, 0.017386495590209963, 0.017273855209350587, 0.0167956485748291, 0.016850944519042968, 0.016876544952392578, 0.017253376007080077, 0.017183744430541992, 0.016723968505859374, 0.01681407928466797, 0.017091583251953125, 0.016707584381103514, 0.017163263320922852, 0.017362943649291994, 0.017119232177734374, 0.01725644874572754, 0.01681305694580078, 0.017192960739135742, 0.017343488693237305, 0.01740390396118164, 0.01741414451599121, 0.017506303787231444, 0.01739263916015625, 0.017320959091186524, 0.016937984466552734, 0.017947647094726564, 0.01725951957702637, 0.017141759872436522, 0.01663488006591797, 0.017085439682006837, 0.01746227264404297, 0.017201152801513672, 0.017299455642700197, 0.016751615524291993, 0.016669696807861328, 0.01738035202026367, 0.017353727340698243, 0.01720729637145996, 0.017287168502807617, 0.017367071151733398, 0.017203168869018556, 0.01739263916015625, 0.017426431655883787, 0.017375232696533204, 0.017385471343994142, 0.01742438316345215, 0.0170199031829834, 0.01651728057861328, 0.01707811164855957, 0.01718681526184082, 0.016687103271484375, 0.01678745651245117, 0.016920576095581053, 0.017475584030151366, 0.03582566452026367, 0.01760256004333496, 0.017308671951293944, 0.017273855209350587, 0.017546239852905272, 0.017535999298095704, 0.016873504638671873, 0.01689187240600586, 0.016947200775146484, 0.017308671951293944, 0.017551359176635743, 0.017452032089233398, 0.017297407150268555, 0.017474559783935546, 0.017145856857299805, 0.017472511291503907, 0.017175552368164062, 0.017342464447021484, 0.017396896362304688, 0.017432479858398436, 0.017765312194824218, 0.017258495330810548, 0.017160192489624023, 0.01656524848937988, 0.01660927963256836, 0.01661644744873047, 0.01743974494934082, 0.01717452812194824, 0.01724006462097168, 0.017269760131835937, 0.017257471084594727, 0.017367040634155274, 0.017465375900268556, 0.017996768951416015, 0.02202828788757324, 0.018571264266967775, 0.017719295501708983, 0.017763328552246094, 0.017911808013916015, 0.017796096801757814, 0.01760972785949707, 0.017691648483276368, 0.017979391098022462, 0.018464767456054687, 0.01779097557067871, 0.0177039680480957, 0.017647584915161132, 0.017613824844360353, 0.017697792053222656, 0.017661951065063478, 0.017505279541015627, 0.017681407928466796, 0.017803264617919923, 0.017753087997436523, 0.01776742362976074, 0.01788313674926758, 0.017640607833862305, 0.01768739128112793, 0.017741823196411134, 0.017727487564086913, 0.01761484718322754, 0.01775103950500488, 0.01784940719604492, 0.0375428466796875, 0.017771520614624024, 0.017687551498413084, 0.017302528381347656, 0.01760870361328125, 0.017532928466796875, 0.017666048049926757, 0.01763430404663086, 0.01773465538024902, 0.01787392044067383, 0.01775923156738281, 0.017724416732788087, 0.017515520095825195, 0.017945600509643556, 0.017739776611328126, 0.01785958480834961, 0.017694719314575197, 0.01779302406311035, 0.017697824478149413, 0.017821664810180663, 0.017617919921875, 0.017625087738037108, 0.017590272903442384, 0.017541120529174805, 0.017499135971069335, 0.017796096801757814, 0.01761996841430664, 0.01737215995788574, 0.01741312026977539, 0.01742540740966797, 0.017345535278320313, 0.017519615173339845, 0.017633279800415038, 0.01763839912414551, 0.017671167373657228, 0.017538047790527343, 0.017583103179931642, 0.017498111724853514, 0.01763737678527832, 0.01760051155090332, 0.017500160217285156, 0.01765990447998047, 0.01745305633544922, 0.017648639678955077, 0.017542144775390626, 0.017518592834472657, 0.017491968154907226, 0.017584127426147463, 0.017726463317871095, 0.017681407928466796, 0.01785958480834961, 0.017780736923217775, 0.01777561569213867, 0.017698816299438477, 0.01761894416809082, 0.01808896064758301, 0.018035711288452147, 0.019092479705810548, 0.01798041534423828, 0.01765376091003418, 0.017625087738037108, 0.01763942337036133, 0.017736703872680663, 0.03618201446533203, 0.01776639938354492, 0.017732608795166017, 0.017599584579467774, 0.01721743965148926, 0.017369087219238282, 0.01765376091003418, 0.017743871688842772, 0.017519615173339845, 0.017794048309326172, 0.017833984375, 0.017758207321166994, 0.017728511810302734, 0.017733631134033204, 0.017690624237060547, 0.017615936279296876, 0.017680320739746094, 0.01777564811706543, 0.01753094482421875, 0.017258399963378905, 0.01739263916015625, 0.017511423110961915, 0.01762713623046875, 0.017498111724853514, 0.01804800033569336, 0.017732608795166017, 0.01776742362976074, 0.017703935623168944, 0.0176363525390625, 0.01764352035522461, 0.01758515167236328, 0.017477632522583008, 0.017727487564086913, 0.017789951324462892, 0.017121280670166016, 0.01761996841430664, 0.01739571189880371, 0.01764556884765625, 0.01781862449645996, 0.017699840545654297, 0.01785241508483887, 0.017695743560791014, 0.017847295761108398, 0.0178155517578125, 0.01758515167236328, 0.017693695068359376, 0.01768448066711426, 0.01758220863342285, 0.017704832077026368, 0.017727487564086913, 0.01787392044067383, 0.017673215866088866, 0.017552383422851564, 0.01780531120300293, 0.017655807495117186, 0.017495168685913085, 0.018366336822509766, 0.017979391098022462, 0.017574911117553712, 0.017830911636352538, 0.01767628860473633, 0.017675264358520508, 0.01805516815185547, 0.034700286865234374, 0.016712703704833985, 0.01663795280456543, 0.01683251190185547, 0.01682329559326172, 0.016916479110717773, 0.01681920051574707, 0.01677004814147949, 0.016755712509155272, 0.01682022476196289, 0.016842752456665038, 0.017167360305786132, 0.01683148765563965, 0.016741439819335936, 0.016780223846435547, 0.01670047950744629, 0.01698092842102051, 0.016892927169799805, 0.016912384033203123, 0.01663283157348633, 0.016662527084350585, 0.016877567291259766, 0.016871423721313478, 0.016919551849365236, 0.016892927169799805, 0.016753664016723634, 0.01683046340942383, 0.016885759353637696, 0.01666662406921387, 0.016698368072509767, 0.016678911209106445, 0.016545791625976563, 0.016846847534179688, 0.01683251190185547, 0.016841728210449217, 0.016908287048339844, 0.016850944519042968, 0.016827392578125, 0.016845823287963867, 0.016782400131225585, 0.01665939140319824, 0.016681087493896483, 0.016987007141113283, 0.016733184814453125, 0.016630783081054687, 0.01661644744873047, 0.016703487396240235, 0.016707584381103514, 0.016721920013427736, 0.016871423721313478, 0.01681817626953125, 0.016833536148071288, 0.01679462432861328, 0.016657407760620118, 0.016876544952392578, 0.016738304138183592, 0.016494623184204103, 0.016753631591796873, 0.016919551849365236, 0.01676697540283203, 0.016699392318725585, 0.016776191711425782, 0.01684480094909668, 0.034543617248535156, 0.01660313606262207, 0.016696319580078126, 0.016720895767211915, 0.016579584121704103, 0.016665599822998048, 0.016752639770507814, 0.016959487915039064, 0.01680281639099121, 0.016735231399536133, 0.01686425590515137, 0.01663692855834961, 0.016639999389648438, 0.01664204788208008, 0.016903167724609376, 0.016899072647094726, 0.016751615524291993, 0.016857088088989256, 0.016857088088989256, 0.016861183166503906, 0.01684480094909668, 0.01680076789855957, 0.016587776184082033, 0.016640031814575195, 0.016519136428833008, 0.016861183166503906, 0.01686835289001465, 0.016912416458129884, 0.016912351608276366, 0.016954368591308593, 0.016764928817749023, 0.016772096633911132, 0.016903167724609376, 0.016870399475097657, 0.01680588722229004, 0.016681983947753908, 0.016601247787475584, 0.016563039779663086, 0.016570367813110352, 0.016782432556152343, 0.01667695999145508, 0.016713632583618163, 0.01665951919555664, 0.016781183242797853, 0.01680892753601074, 0.016586751937866212, 0.016517120361328123, 0.0165928955078125, 0.016719871520996094, 0.016857088088989256, 0.016723968505859374, 0.016727039337158203, 0.016777215957641603, 0.01683251190185547, 0.016833536148071288, 0.01657548713684082, 0.016712703704833985, 0.016751615524291993, 0.016825344085693358, 0.01685196876525879, 0.01679667282104492, 0.01681100845336914, 0.016792575836181642, 0.034274303436279296, 0.01683558464050293, 0.016763904571533202, 0.016906240463256835, 0.016900096893310547, 0.016665599822998048, 0.01680793571472168, 0.016871423721313478, 0.016895999908447267, 0.01683865547180176, 0.01699635124206543, 0.016720895767211915, 0.016875520706176757, 0.016862207412719727, 0.016883712768554687, 0.016664575576782227, 0.016529407501220703, 0.016408575057983397, 0.016458751678466797, 0.016781312942504883, 0.01683148765563965, 0.016755712509155272, 0.01663283157348633, 0.016681983947753908, 0.017364992141723632, 0.017688575744628905, 0.017581056594848633, 0.01743052864074707, 0.01734758377075195, 0.017449983596801756, 0.017523712158203125, 0.01744895935058594, 0.017846271514892577, 0.017512447357177736, 0.0175861759185791, 0.017509376525878906, 0.01776639938354492, 0.017803264617919923, 0.017290239334106446, 0.01745715141296387, 0.01768550491333008, 0.017538047790527343, 0.017522687911987304, 0.017565696716308594, 0.017643680572509767, 0.017544031143188477, 0.01759334373474121, 0.01777459144592285, 0.018421760559082033, 0.017902591705322265, 0.017549312591552735, 0.017760255813598632, 0.017701887130737306, 0.017852447509765626, 0.01769264030456543, 0.017689599990844726, 0.017560575485229494, 0.0176312313079834, 0.017709056854248048, 0.017730560302734375, 0.01779302406311035, 0.01785139274597168, 0.017896448135375977]",tokens/s,57.26091447852295,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,,cuda,0,42,,,,,False,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,56387f0ed9acb6808de92d1e9f4c4b67cc3f2798,4.40.2,,0.30.0,,,,1.19.1,,,,0.10.0,,,MB,3999.793152,12218.53184,0.0,11572.084736,11067.6992,s,10,6.280759704589844,0.6280759704589844,0.003145377785725941,0.6271241149902345,0.62955556640625,0.6331921508789062,0.6361014184570313,"[0.6368287353515625, 0.6253135986328126, 0.6252825927734375, 0.6265311889648437, 0.627404541015625, 0.6268436889648438, 0.6267838134765625, 0.6283837890625, 0.6287474365234375, 0.6286403198242188]",tokens/s,407.5940046120865,kWh,7.388594808677833e-06,4.048393679443051e-06,4.113359193462507e-05,5.257058042274595e-05,tokens/kWh,4869643.780635059,MB,3999.793152,12218.53184,0.0,11572.084736,11417.442816,s,10,346.04539453125,34.604539453125,0.0047755651489063055,34.605564453125,34.6111265625,34.61120390625,34.61126578125,"[34.611109375, 34.60820703125, 34.60643359375, 34.6046953125, 34.61128125, 34.600859375, 34.59778125, 34.5991171875, 34.5991640625, 34.60674609375]",tokens/s,1.8205703932381252,kWh,0.000408529573318859,0.00022390994749188167,0.002260659416164959,0.0028930989369757,tokens/kWh,21775.957674595473,,s,629,351.13987780761704,0.5582509981043199,0.07457057126364097,0.5492091064453125,0.5496271484374999,0.5498900634765626,1.1765116064453125,"[0.5493729248046875, 0.5499044189453125, 0.5497476196289063, 0.5494589233398437, 0.5499525146484375, 0.5494312744140625, 0.5494948120117188, 0.5496258544921875, 0.5498163452148438, 0.5496616821289062, 0.5496156005859375, 0.5494312744140625, 0.5493534545898437, 0.5489326171875, 0.5494537963867188, 0.5491026000976562, 0.5491671142578125, 0.5491046142578125, 0.5493237915039062, 0.5494691772460938, 0.5490820922851563, 0.5493237915039062, 0.5493032836914062, 0.5493135375976562, 0.549086181640625, 0.5493186645507813, 0.549139404296875, 0.5493739624023437, 0.5494599609375, 0.549264404296875, 0.54923876953125, 0.5489387817382813, 0.54940673828125, 0.5491517944335937, 0.5491987915039063, 0.5492828369140625, 0.54899609375, 0.5492469482421874, 0.5514075927734375, 0.54955517578125, 0.5491506958007812, 0.5490308837890625, 0.5491793823242187, 0.549517333984375, 0.5489141845703125, 0.5489879150390625, 0.5490933837890625, 0.5492295532226562, 0.5493135375976562, 0.549370849609375, 0.5498460083007812, 0.5490524291992187, 0.5492725830078125, 0.549475341796875, 0.5494097900390625, 0.5497507934570313, 0.549017578125, 0.5494312744140625, 0.5492991943359375, 0.549381103515625, 0.5490186157226562, 0.5491671142578125, 1.1763302001953124, 0.5493514404296875, 0.5497794799804687, 0.5493851928710938, 0.5492203369140625, 0.5496012573242187, 0.5492080688476563, 0.5489336547851562, 0.54935546875, 0.549738525390625, 0.5498255615234375, 0.550150146484375, 0.549559326171875, 0.5492346801757813, 0.54894384765625, 0.5491292114257813, 0.5490780029296874, 0.5490667724609375, 0.5491978149414063, 0.5493821411132812, 0.54919677734375, 0.5490882568359375, 0.5489418334960937, 0.54934326171875, 0.5492192993164062, 0.5490953979492188, 0.5489889526367188, 0.5490985107421875, 0.549001220703125, 0.5493043212890625, 0.5491640625, 0.54915380859375, 0.5488732299804687, 0.5488076782226563, 0.5493995361328124, 0.5489950561523438, 0.5490053100585938, 0.549485595703125, 0.5489889526367188, 0.54925927734375, 0.5492408447265625, 0.5491302490234375, 0.549444580078125, 0.5494763793945312, 0.5491292114257813, 0.5491414794921875, 0.5523097534179687, 0.5497333984375, 0.5492971801757812, 0.5498634033203125, 0.5491056518554688, 0.5496801147460938, 0.5490728759765625, 0.5494292602539063, 0.5492265014648438, 0.549843994140625, 0.549517333984375, 0.5492408447265625, 0.5492408447265625, 0.5490811157226563, 0.5495203857421875, 0.5489991455078125, 0.5488455810546875, 1.175055419921875, 0.5490615844726563, 0.5490503540039062, 0.5495418701171875, 0.5491292114257813, 0.5492203369140625, 0.5486981201171875, 0.5490934448242187, 0.5489909057617187, 0.5498101806640625, 0.5493739624023437, 0.5491322631835938, 0.5491988525390625, 0.5490513916015625, 0.5492305908203124, 0.5490985107421875, 0.5496575927734375, 0.5495326538085937, 0.5493944091796875, 0.5496749877929688, 0.5495029907226563, 0.5493422241210938, 0.54955517578125, 0.5509222412109375, 0.549212158203125, 0.5494579467773437, 0.5503784790039062, 0.54936474609375, 0.5502095336914062, 0.5490667724609375, 0.5491107788085937, 0.5490186157226562, 0.5492612915039062, 0.5492469482421874, 0.5495572509765625, 0.54902783203125, 0.5490350341796875, 0.5490288696289063, 0.5492449340820312, 0.5492920532226563, 0.5488394165039062, 0.5488087158203125, 0.5487308959960937, 0.5489879150390625, 0.5494957885742188, 0.5492848510742188, 0.5492490234375, 0.5511434326171875, 0.5490380859375, 0.5490576171875, 0.5489745483398437, 0.549454833984375, 0.5490012817382812, 0.549023681640625, 0.5490616455078124, 0.5491005249023437, 0.5491302490234375, 0.5493145751953125, 0.549339111328125, 0.5492275390625, 0.549212158203125, 0.5491732177734375, 0.5492531127929687, 1.1767552490234374, 0.5493289184570312, 0.5493759765625, 0.5493156127929687, 0.5495326538085937, 0.5494599609375, 0.5490677490234375, 0.5494087524414063, 0.5493575439453126, 0.5493729248046875, 0.5496115112304687, 0.549412841796875, 0.5494271850585938, 0.54919677734375, 0.5490944213867187, 0.5491732177734375, 0.5489755859375, 0.54938623046875, 0.5495531616210938, 0.5493156127929687, 0.549475341796875, 0.5492500610351563, 0.5490134887695313, 0.5494271850585938, 0.5491742553710938, 0.549064697265625, 0.5489920043945312, 0.5490820922851563, 0.5494343872070313, 0.5490769653320312, 0.5493667602539063, 0.5490360107421876, 0.5489786987304688, 0.5504778442382813, 0.5492838134765625, 0.5490104370117187, 0.5495838623046875, 0.5494323120117187, 0.5493759765625, 0.5493156127929687, 0.5489910278320312, 0.5492316284179688, 0.5488772583007813, 0.5490237426757812, 0.5491373901367187, 0.5489981689453125, 0.548853759765625, 0.5494957885742188, 0.5495675048828125, 0.5491425170898437, 0.5489868774414063, 0.5493934326171875, 0.5492817993164063, 0.5491199951171875, 0.54923876953125, 0.549760009765625, 0.5492428588867188, 0.5491691284179687, 0.549011474609375, 0.549317626953125, 0.549106689453125, 0.5491886596679687, 0.54917626953125, 1.1772979736328124, 0.549538818359375, 0.5495776977539063, 0.5493831787109374, 0.5490739135742188, 0.5492469482421874, 0.5491199951171875, 0.5494149169921875, 0.5496207275390625, 0.5495152587890625, 0.5494763793945312, 0.5490155639648437, 0.54917529296875, 0.5495193481445313, 0.5495818481445313, 0.54920703125, 0.5491773681640625, 0.5490872192382813, 0.54920703125, 0.5496873168945312, 0.54965966796875, 0.5492684936523438, 0.5497507934570313, 0.5490616455078124, 0.5492091064453125, 0.5491107788085937, 0.5491712036132812, 0.5494589233398437, 0.5494906616210937, 0.5488271484375, 0.5490391235351563, 0.5495040283203125, 0.5490841674804687, 0.548937744140625, 0.5493248291015626, 0.5489848022460937, 0.5491107788085937, 0.5494896850585937, 0.548895751953125, 0.5493534545898437, 0.5491865844726562, 0.5494159545898437, 0.5492510986328125, 0.549465087890625, 0.5495449829101563, 0.5495828247070312, 0.549433349609375, 0.5491005249023437, 0.5491046142578125, 0.5494537963867188, 0.5493667602539063, 0.54966064453125, 0.549074951171875, 0.5488209838867187, 0.5490780029296874, 0.5495040283203125, 0.5493095092773438, 0.5541016845703125, 0.549232666015625, 0.5494569091796875, 0.5492603149414063, 0.549359619140625, 0.549291015625, 1.176932373046875, 0.5496279296875, 0.5491742553710938, 0.5493401489257812, 0.5498306884765625, 0.549434326171875, 0.5490739135742188, 0.5489838256835937, 0.5493360595703125, 0.5495941162109375, 0.5493084106445313, 0.5494630126953125, 0.54894384765625, 0.549080078125, 0.5490370483398438, 0.549375, 0.5490370483398438, 0.5501358032226562, 0.549159912109375, 0.5493360595703125, 0.5492848510742188, 0.5492951049804687, 0.5494998779296875, 0.5489397583007812, 0.5491414794921875, 0.548674560546875, 0.5486264038085937, 0.5489203491210938, 0.5495234375, 0.5495101318359376, 0.5490944213867187, 0.5489264526367188, 0.5486029052734375, 0.54854248046875, 0.5488701171875, 0.5489879150390625, 0.5493340454101563, 0.5489827880859375, 0.5516820678710938, 0.5491456298828125, 0.5495726318359375, 0.5491005249023437, 0.5491712036132812, 0.5491302490234375, 0.5491107788085937, 0.5489714965820313, 0.5494783935546875, 0.549212158203125, 0.5491210327148438, 0.5490258178710937, 0.5487687377929688, 0.5492500610351563, 0.548906005859375, 0.5495992431640625, 0.5493923950195313, 0.5492091064453125, 0.54873291015625, 0.5487176513671875, 0.5490247192382812, 0.549032958984375, 0.5490923461914062, 0.5490974731445313, 0.5492244262695313, 1.1765821533203125, 0.549148681640625, 0.5489417724609374, 0.5489469604492188, 0.5489080200195312, 0.5490667724609375, 0.5490513916015625, 0.5494691772460938, 0.5497661743164063, 0.5491650390625, 0.5492275390625, 0.548947998046875, 0.5488783569335938, 0.5488609008789063, 0.5490820922851563, 0.5490503540039062, 0.5489367065429688, 0.548890625, 0.5491875610351562, 0.5492265014648438, 0.5492531127929687, 0.5489019775390624, 0.5489847412109375, 0.55086181640625, 0.5490391235351563, 0.5492428588867188, 0.5496053466796875, 0.5490841674804687, 0.54921728515625, 0.54888037109375, 0.5493319702148437, 0.5492100830078125, 0.5487738647460938, 0.5490811157226563, 0.549359619140625, 0.5490216674804688, 0.54917529296875, 0.5490892944335938, 0.5492725830078125, 0.54915380859375, 0.5489417724609374, 0.5489336547851562, 0.5490084228515625, 0.5492111206054687, 0.5502412719726563, 0.549032958984375, 0.5490411376953125, 0.5488516845703125, 0.5487493896484374, 0.549298095703125, 0.5489725341796875, 0.5493114624023437, 0.5493677978515625, 0.5491763305664062, 0.548974609375, 0.5500374755859375, 0.5493206787109375, 0.5489971313476563, 0.5488466186523437, 0.5491189575195312, 0.54900634765625, 0.5489315795898437, 0.5493463134765625, 1.1782021484375, 0.5492080688476563, 0.5489561767578125, 0.549626953125, 0.5489909057617187, 0.5491159057617188, 0.548748291015625, 0.5488844604492188, 0.5489613037109375, 0.5494312744140625, 0.5492080688476563, 0.5488895874023437, 0.5490647583007813, 0.548844482421875, 0.5486868286132812, 0.5488568115234375, 0.550529052734375, 0.5495675048828125, 0.5489367065429688, 0.5490166015625, 0.5491179809570312, 0.549296142578125, 0.5492100830078125, 0.54926953125, 0.54919677734375, 0.5489613647460938, 0.5496729125976563, 0.54991357421875, 0.5499207763671875, 0.5490032348632813, 0.5490953979492188, 0.549391357421875, 0.549359619140625, 0.5493370971679687, 0.549728271484375, 0.5490953979492188, 0.5488455810546875, 0.549022705078125, 0.5487913208007813, 0.549622802734375, 0.5490032348632813, 0.5489090576171874, 0.5489203491210938, 0.5491210327148438, 0.5489152221679687, 0.5491435546875, 0.5498665161132813, 0.5491824340820313, 0.5490042724609375, 0.5491773681640625, 0.549228515625, 0.5496483764648438, 0.54925927734375, 0.5493575439453126, 0.548912109375, 0.548969482421875, 0.5487401123046876, 0.5491445922851562, 0.5492316284179688, 0.5490237426757812, 0.5489817504882812, 0.5489397583007812, 0.5492889404296875, 1.178955810546875, 0.5491865844726562, 0.5490596313476562, 0.5495131225585937, 0.5489755859375, 0.5494159545898437, 0.5494210815429688, 0.5493309326171875, 0.5491107788085937, 0.549123046875, 0.54890087890625, 0.5491640625, 0.549148681640625, 0.5489684448242188, 0.548832275390625, 0.5490023193359375, 0.5489827270507812, 0.5507020874023437, 0.5497640991210937, 0.5493606567382813, 0.5494384765625, 0.5490585327148437, 0.548832275390625, 0.5493831787109374, 0.5493718872070312, 0.5493197021484375, 0.5489633178710938, 0.5492408447265625, 0.5493145751953125, 0.5490401000976562, 0.54955419921875, 0.5490197143554687, 0.5490543823242188, 0.5490513916015625, 0.5493883056640625, 0.5489407958984375, 0.549127197265625, 0.5490299072265625, 0.5489981689453125, 0.5489295654296875, 0.5490093994140625, 0.549212158203125, 0.5490790405273438, 0.5493718872070312, 0.5491322631835938, 0.5489305419921875, 0.5492234497070313, 0.5489827880859375, 0.5493575439453126, 0.5491619873046875, 0.5489735717773437, 0.5492951049804687, 0.5490442504882812, 0.5488650512695312, 0.5490299072265625, 0.5495132446289063, 0.548927490234375, 0.5490134887695313, 0.5489254150390624, 0.5494537963867188, 0.548969482421875, 0.5491578979492188, 0.5493360595703125, 1.178430419921875, 0.548947998046875, 0.549432373046875, 0.5493032836914062, 0.5492889404296875, 0.549538818359375, 0.548990966796875, 0.5494138793945312, 0.5493677978515625, 0.5498685302734375, 0.5495060424804687, 0.54938623046875, 0.5493104858398438, 0.5493667602539063, 0.5493237915039062, 0.54991357421875, 0.5494466552734375, 0.5496187133789062, 0.5493790893554688, 0.549349365234375, 0.549760009765625, 0.54940771484375, 0.5489755859375, 0.5493883056640625, 0.5489315795898437, 0.5488291625976562, 0.5510830078125, 0.5492244262695313, 0.5493923950195313, 0.5486192626953125, 0.548895751953125, 0.5490933837890625, 0.5489889526367188, 0.5491302490234375, 0.5494896850585937, 0.5493043212890625, 0.5490964965820313, 0.5491097412109375, 0.5493883056640625, 0.5490022583007812, 0.549053466796875, 0.5489766235351563, 0.5489459228515625, 0.5492561645507813, 0.549147705078125, 0.5497691650390625, 0.5495582885742187, 0.5489080200195312, 0.5490001831054687, 0.549123046875, 0.5491046142578125, 0.55032421875, 0.5495818481445313, 0.5496268920898437, 0.5491036376953125, 0.5493821411132812, 0.5493032836914062, 0.5495152587890625, 0.5492131958007812, 0.5490473022460938, 0.5491732177734375, 0.5488619384765625, 0.5491926879882812]",tokens/s,1.7913089334291363,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,4915.818496,5956.435968,0.0,5309.988864,5047.4112,s,10,3.514953002929688,0.3514953002929687,0.003265316468333025,0.35043815612792967,0.3522641021728516,0.3567220932006836,0.3602884860229492,"[0.3611800842285156, 0.3499181823730469, 0.35083880615234375, 0.3506215209960937, 0.3510619812011719, 0.3497488098144531, 0.3512734375, 0.34994931030273435, 0.3502547912597656, 0.3501060791015625]",tokens/s,728.316992536247,kWh,4.133434410981292e-06,2.2649312095758246e-06,2.0026433645656355e-05,2.6424799266213472e-05,tokens/kWh,9687869.240593228,MB,4915.818496,5956.435968,0.0,5309.988864,5098.333184,s,10,196.5019375,19.65019375,0.004894283073691026,19.6512177734375,19.655654296875,19.656181640625,19.656603515625,"[19.6550078125, 19.655537109375, 19.644330078125, 19.649078125, 19.64495703125, 19.643501953125, 19.646048828125, 19.65341015625, 19.653357421875, 19.656708984375]",tokens/s,3.206075258163803,kWh,0.00023203698898412251,0.00012717321322392182,0.001095179319629344,0.0014543895218373883,tokens/kWh,43317.14375967834,,s,629,199.34022326660164,0.3169160942235319,0.04163812073537977,0.3118530578613281,0.3122046936035156,0.31237509155273435,0.6620613842773437,"[0.3120076904296875, 0.3119175720214844, 0.3122165832519531, 0.3119708251953125, 0.311731201171875, 0.3118448791503906, 0.3117916259765625, 0.311836669921875, 0.311947265625, 0.3118551025390625, 0.31186944580078124, 0.3120506896972656, 0.311947265625, 0.3119452209472656, 0.3117752380371094, 0.31203021240234374, 0.3118172302246094, 0.31187661743164063, 0.3118172302246094, 0.31209268188476563, 0.31193804931640623, 0.3119237060546875, 0.31178546142578123, 0.3118254089355469, 0.31234561157226565, 0.312079345703125, 0.3118202819824219, 0.31171173095703125, 0.31192678833007814, 0.3117547607421875, 0.3116799926757812, 0.31171173095703125, 0.3117178955078125, 0.311868408203125, 0.31268658447265624, 0.31187966918945315, 0.31214080810546874, 0.31197695922851565, 0.3117424621582031, 0.3118479309082031, 0.3122442321777344, 0.3118254089355469, 0.3116779479980469, 0.31168307495117187, 0.31167486572265624, 0.31168307495117187, 0.31172198486328123, 0.31186944580078124, 0.311731201171875, 0.31188479614257814, 0.31199026489257814, 0.3118960571289062, 0.3121612854003906, 0.311942138671875, 0.3118335876464844, 0.31192474365234374, 0.3122104187011719, 0.31238040161132813, 0.3120220031738281, 0.31192166137695315, 0.31197592163085935, 0.3122083740234375, 0.6620631103515625, 0.31184588623046877, 0.31185614013671875, 0.31188479614257814, 0.31196466064453127, 0.3120138244628906, 0.3117055969238281, 0.31188479614257814, 0.3119503479003906, 0.31250534057617185, 0.31180389404296877, 0.3120015258789062, 0.3118837890625, 0.311942138671875, 0.311657470703125, 0.3117701110839844, 0.3119923095703125, 0.31211520385742186, 0.312453125, 0.312026123046875, 0.31225958251953123, 0.3121766357421875, 0.31197906494140626, 0.31207623291015624, 0.31240704345703124, 0.3126917114257812, 0.3125841979980469, 0.31177426147460935, 0.31191033935546875, 0.3118182373046875, 0.31168511962890627, 0.311794677734375, 0.3120394287109375, 0.3118335876464844, 0.31192779541015625, 0.31178546142578123, 0.3118254089355469, 0.31185614013671875, 0.31187661743164063, 0.31182131958007814, 0.31181112670898437, 0.3120495910644531, 0.3117271118164063, 0.31169329833984377, 0.3130624084472656, 0.3118837890625, 0.31198004150390624, 0.3120107421875, 0.31207424926757815, 0.31199026489257814, 0.3120025634765625, 0.31195755004882814, 0.312280029296875, 0.31251773071289063, 0.3118078918457031, 0.31177728271484373, 0.31187457275390623, 0.3122667541503906, 0.31216433715820313, 0.3118121032714844, 0.3116759033203125, 0.3118663635253906, 0.31180801391601565, 0.6627942504882812, 0.31164007568359375, 0.3119861755371094, 0.31173016357421873, 0.311720947265625, 0.31180902099609376, 0.31198004150390624, 0.31179879760742185, 0.31178750610351563, 0.3117342834472656, 0.3118335876464844, 0.31179367065429686, 0.31170867919921874, 0.31184280395507813, 0.3117824096679688, 0.3117926330566406, 0.31182232666015625, 0.3117926330566406, 0.31173837280273436, 0.3119779968261719, 0.3122431945800781, 0.3120394287109375, 0.3121520690917969, 0.31152435302734377, 0.31191552734375, 0.31201791381835936, 0.31192166137695315, 0.3120025634765625, 0.31182232666015625, 0.311900146484375, 0.3118202819824219, 0.311984130859375, 0.311825439453125, 0.3118980712890625, 0.3114987487792969, 0.3117496337890625, 0.31154791259765624, 0.3116789855957031, 0.3115540466308594, 0.311689208984375, 0.31158477783203126, 0.311741455078125, 0.31166259765625, 0.31153048706054687, 0.3116656494140625, 0.3117424621582031, 0.311689208984375, 0.311741455078125, 0.3118397521972656, 0.3116216430664063, 0.3118674011230469, 0.311736328125, 0.311615478515625, 0.31230361938476564, 0.3117557678222656, 0.3117342834472656, 0.31182437133789065, 0.31218585205078125, 0.31234561157226565, 0.3119175720214844, 0.311878662109375, 0.3116390380859375, 0.3117127685546875, 0.6626898193359375, 0.311836669921875, 0.3117373352050781, 0.31185714721679686, 0.3118940124511719, 0.3116974182128906, 0.31178750610351563, 0.3117178955078125, 0.3116871643066406, 0.31178546142578123, 0.31165234375, 0.3118581848144531, 0.3118305358886719, 0.3117978515625, 0.31175363159179686, 0.31180697631835935, 0.3117230224609375, 0.3122688293457031, 0.3117516479492187, 0.3117926330566406, 0.31178854370117187, 0.31269375610351563, 0.3120660400390625, 0.3118540649414063, 0.3118530578613281, 0.31194622802734373, 0.31193292236328124, 0.31199539184570313, 0.31208755493164064, 0.31201995849609376, 0.31198721313476563, 0.3119032287597656, 0.3118919677734375, 0.3120056457519531, 0.3120332946777344, 0.31201995849609376, 0.31191552734375, 0.31209881591796873, 0.311942138671875, 0.3118674011230469, 0.31210086059570313, 0.31198208618164064, 0.3122421875, 0.3118653564453125, 0.312052734375, 0.31169125366210937, 0.31174041748046877, 0.311615478515625, 0.31168511962890627, 0.3117916259765625, 0.3118940124511719, 0.31168511962890627, 0.31174349975585935, 0.31228414916992187, 0.31193191528320313, 0.31168511962890627, 0.31187966918945315, 0.3119452209472656, 0.3118202819824219, 0.3119431762695313, 0.3119390869140625, 0.3117762451171875, 0.3116421203613281, 0.6631044921875, 0.3115663452148438, 0.31163290405273436, 0.3116759033203125, 0.3118182373046875, 0.31181005859375, 0.3118551025390625, 0.31183462524414063, 0.31163494873046876, 0.3117475891113281, 0.311736328125, 0.31175372314453126, 0.31175985717773436, 0.31170867919921874, 0.31166360473632815, 0.3117046203613281, 0.3116338806152344, 0.311773193359375, 0.3117393798828125, 0.31177215576171874, 0.3116974182128906, 0.3126138916015625, 0.311794677734375, 0.3118899230957031, 0.3119933471679687, 0.3123701782226562, 0.31210699462890623, 0.3118888854980469, 0.3117824096679688, 0.3116175231933594, 0.31168511962890627, 0.3118551025390625, 0.31176702880859375, 0.31181005859375, 0.31184384155273437, 0.3117998046875, 0.31185202026367187, 0.31173016357421873, 0.311841796875, 0.31199642944335937, 0.3118284912109375, 0.311794677734375, 0.31187359619140625, 0.3117526550292969, 0.31203225708007815, 0.31181414794921875, 0.3116369934082031, 0.31174655151367187, 0.31168511962890627, 0.31185714721679686, 0.311630859375, 0.3118131103515625, 0.3116451721191406, 0.31207424926757815, 0.3117066650390625, 0.3116246643066406, 0.31169842529296876, 0.3121131591796875, 0.3120466003417969, 0.3118899230957031, 0.31201177978515626, 0.3121448974609375, 0.31176907348632815, 0.6617681884765625, 0.3116810302734375, 0.31177218627929687, 0.31163491821289063, 0.31164825439453125, 0.3117332458496094, 0.31166156005859375, 0.3117127685546875, 0.3116216430664063, 0.31163290405273436, 0.31180084228515625, 0.31170458984375, 0.3116707763671875, 0.3117004699707031, 0.31171173095703125, 0.3116595153808594, 0.3117619323730469, 0.3117782897949219, 0.3119493103027344, 0.31163494873046876, 0.31175985717773436, 0.312489990234375, 0.3118807067871094, 0.3118407592773437, 0.3117916259765625, 0.3120414733886719, 0.31199026489257814, 0.3117916259765625, 0.31184280395507813, 0.31183154296875, 0.31176602172851564, 0.31183462524414063, 0.3117445678710937, 0.31164306640625, 0.31175167846679686, 0.31193701171875, 0.3118899230957031, 0.311868408203125, 0.3117782897949219, 0.31173016357421873, 0.3117895812988281, 0.31164825439453125, 0.3116779479980469, 0.31167486572265624, 0.3116421203613281, 0.3118305358886719, 0.31164825439453125, 0.31182745361328124, 0.3116974182128906, 0.3116779479980469, 0.3121971130371094, 0.31193804931640623, 0.311804931640625, 0.3120814208984375, 0.3117578125, 0.31161138916015624, 0.31186944580078124, 0.31213772583007815, 0.3120220031738281, 0.31196878051757815, 0.31182745361328124, 0.31183462524414063, 0.3116697692871094, 0.663519287109375, 0.3117413635253906, 0.311762939453125, 0.31178750610351563, 0.3118335876464844, 0.3117916259765625, 0.31170050048828124, 0.31172604370117185, 0.31168307495117187, 0.31174554443359376, 0.3116451721191406, 0.31184384155273437, 0.31170458984375, 0.3117916259765625, 0.31180697631835935, 0.3117025146484375, 0.3117506408691406, 0.3116646423339844, 0.3116288146972656, 0.3120005187988281, 0.3116656494140625, 0.3128586120605469, 0.311773193359375, 0.31196978759765626, 0.3120220031738281, 0.31235379028320315, 0.31208856201171875, 0.31201278686523437, 0.3118499755859375, 0.3117967224121094, 0.31190118408203127, 0.31183770751953127, 0.311900146484375, 0.3118387145996094, 0.31162469482421873, 0.3117496337890625, 0.3116236877441406, 0.3119913024902344, 0.3117844543457031, 0.311630859375, 0.3115704345703125, 0.31169125366210937, 0.311773193359375, 0.3117352905273437, 0.31176907348632815, 0.3115888671875, 0.31168307495117187, 0.3117926330566406, 0.3118397521972656, 0.31179571533203126, 0.31169638061523436, 0.311920654296875, 0.31166156005859375, 0.31240090942382814, 0.3120281677246094, 0.31179571533203126, 0.3116339111328125, 0.3120947265625, 0.31178854370117187, 0.3122022399902344, 0.31194622802734373, 0.31199026489257814, 0.31187149047851564, 0.6620569458007812, 0.31159295654296876, 0.312047607421875, 0.3117424621582031, 0.3119390869140625, 0.3116492919921875, 0.3116697692871094, 0.3118837890625, 0.31192678833007814, 0.3118407592773437, 0.3118663635253906, 0.31164108276367186, 0.3117864990234375, 0.3116707763671875, 0.3117281188964844, 0.3116861572265625, 0.3118612365722656, 0.31181515502929685, 0.31188275146484373, 0.3118202819824219, 0.3120086975097656, 0.3122995300292969, 0.31198309326171875, 0.3120773010253906, 0.3120547790527344, 0.31204556274414064, 0.31210086059570313, 0.31193701171875, 0.31183462524414063, 0.31158578491210936, 0.3116462707519531, 0.31192056274414065, 0.3117752380371094, 0.31220120239257815, 0.31220428466796873, 0.31212646484375, 0.31197491455078125, 0.31204043579101565, 0.31203533935546873, 0.3120588684082031, 0.31178546142578123, 0.31172607421875, 0.3118131103515625, 0.3119257507324219, 0.3119339599609375, 0.3119810485839844, 0.3118970947265625, 0.3117967224121094, 0.31180389404296877, 0.3120977783203125, 0.3121049499511719, 0.3120332946777344, 0.3119574890136719, 0.3141632080078125, 0.31226776123046873, 0.3121868896484375, 0.31191961669921875, 0.31209368896484374, 0.31210086059570313, 0.31201486206054685, 0.31168307495117187, 0.31176806640625, 0.31225958251953123, 0.6623856811523438, 0.3117926330566406, 0.31174349975585935, 0.31176397705078124, 0.31193701171875, 0.31180084228515625, 0.31187249755859375, 0.3118172302246094, 0.3119554443359375, 0.3119237060546875, 0.311973876953125, 0.3118581848144531, 0.3118182373046875, 0.311836669921875, 0.3118284912109375, 0.31187661743164063, 0.3117506408691406, 0.3119585266113281, 0.31187966918945315, 0.3117475891113281, 0.31164620971679685, 0.312406005859375, 0.31215719604492187, 0.3118981018066406, 0.3119810485839844, 0.31215719604492187, 0.31199847412109377, 0.31217868041992186, 0.312121337890625, 0.3118264465332031, 0.31171685791015624, 0.3118878784179687, 0.31180084228515625, 0.31187457275390623, 0.311841796875, 0.311973876953125, 0.312099853515625, 0.3118407592773437, 0.3119308776855469, 0.3118397521972656, 0.3117496337890625, 0.3118407592773437, 0.312110107421875, 0.3120650024414062, 0.311836669921875, 0.3120015258789062, 0.31180697631835935, 0.31198513793945315, 0.31185101318359376, 0.3118695068359375, 0.31196868896484375, 0.3123343505859375, 0.311920654296875, 0.3120220031738281, 0.31227801513671877, 0.31184588623046877, 0.3119636535644531, 0.31258316040039064, 0.3123763122558594, 0.31220632934570314, 0.3120056457519531, 0.31237326049804687, 0.3119923095703125, 0.6621552734375, 0.3120445556640625, 0.3120445556640625, 0.31182232666015625, 0.31201278686523437, 0.3124234313964844, 0.31202407836914064, 0.31181619262695315, 0.31179879760742185, 0.31191043090820314, 0.31195745849609374, 0.3122298889160156, 0.3120445556640625, 0.311841796875, 0.311984130859375, 0.3118264465332031, 0.31190118408203127, 0.3119493103027344, 0.3120660400390625, 0.3122237548828125, 0.31208551025390624, 0.3126251831054688, 0.31243362426757815, 0.31203121948242185, 0.3117332458496094, 0.31224627685546874, 0.31222784423828126, 0.31201995849609376, 0.31172402954101563, 0.31192987060546873, 0.3118817138671875, 0.31197491455078125, 0.3124930419921875, 0.31220632934570314, 0.3121766357421875, 0.3118530578613281, 0.3120005187988281, 0.31181414794921875, 0.3118530578613281, 0.3117834167480469, 0.312121337890625, 0.3120977783203125, 0.31200460815429687, 0.31181005859375, 0.31189093017578123, 0.31181414794921875, 0.311731201171875, 0.3116707763671875, 0.311920654296875, 0.3119257507324219, 0.3117752380371094, 0.3117076416015625, 0.31185614013671875, 0.31231796264648437, 0.3123404846191406, 0.3119360046386719, 0.31179879760742185, 0.3122606201171875, 0.31254937744140626, 0.3121121215820313, 0.3119237060546875, 0.3119339599609375, 0.3121715087890625]",tokens/s,3.15540932829579,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,MB,1820.639232,7019.692032,0.0,6371.147776,6308.565504,s,10,1.3269676208496095,0.13269676208496095,0.00013363241314515013,0.13269898986816409,0.13286409454345702,0.13290929641723634,0.13294545791625978,"[0.13277850341796876, 0.1325766143798828, 0.1327298583984375, 0.13250434875488282, 0.1326087646484375, 0.13266812133789063, 0.13273170471191406, 0.1325611572265625, 0.13295449829101563, 0.1328540496826172]",tokens/s,1929.2106000001147,kWh,1.5680979068530643e-06,8.592434780324572e-07,6.275381482289502e-06,8.702722867175024e-06,tokens/kWh,29416080.910215143,MB,1844.772864,7021.789184,0.0,6373.244928,6308.568064,s,10,19.883057861328126,1.9883057861328126,0.013294463060267709,1.9879600219726563,2.0058687622070313,2.008191766357422,2.010050169677734,"[2.0105147705078124, 2.0053525390625, 1.980619140625, 1.9927418212890624, 1.9865767822265625, 2.000463623046875, 1.97184814453125, 1.972032958984375, 1.9735648193359374, 1.98934326171875]",tokens/s,31.68526714521757,kWh,2.366944106467444e-05,1.2971541711587019e-05,7.098237476831017e-05,0.00010762335754457161,tokens/kWh,585374.787010421,,s,630,19.88116284370422,0.03155740133921305,0.0006153156905783781,0.03132774448394775,0.03240140914916992,0.03260605411529541,0.03326953399658203,"[0.03134771156311035, 0.031154176712036134, 0.03386777496337891, 0.032276481628417966, 0.032704513549804685, 0.031340543746948245, 0.0313436164855957, 0.03112550354003906, 0.0310435848236084, 0.03133337593078613, 0.0320819206237793, 0.031926271438598636, 0.03171737670898438, 0.03251302337646484, 0.033157119750976564, 0.03241062545776367, 0.032418815612792966, 0.032301055908203126, 0.03204201507568359, 0.031199199676513672, 0.031492095947265625, 0.030893056869506837, 0.03075071907043457, 0.031320064544677735, 0.030846975326538087, 0.030935039520263673, 0.03162931251525879, 0.03122585678100586, 0.032140289306640625, 0.03270867156982422, 0.03321644973754883, 0.0317388801574707, 0.031660032272338864, 0.03222016143798828, 0.03199078369140625, 0.031105024337768555, 0.03251609420776367, 0.03134771156311035, 0.032105472564697264, 0.032586753845214846, 0.03225395202636719, 0.03263692855834961, 0.03248332977294922, 0.03259801483154297, 0.031264768600463864, 0.03196211242675781, 0.032467967987060545, 0.03124940872192383, 0.032336894989013674, 0.03187711906433106, 0.03135078430175781, 0.03153510475158691, 0.033274879455566404, 0.03260927963256836, 0.03165798377990723, 0.03118284797668457, 0.03127603149414063, 0.03229695892333984, 0.03199180793762207, 0.032353279113769534, 0.03159244728088379, 0.032840705871582034, 0.03150131225585937, 0.033035263061523434, 0.03124224090576172, 0.03237091064453125, 0.03228956985473633, 0.03127807998657227, 0.03242393493652344, 0.032336894989013674, 0.030916608810424805, 0.0314071044921875, 0.031648767471313476, 0.031148031234741212, 0.03138252830505371, 0.03160576057434082, 0.03128319931030273, 0.031264768600463864, 0.03102207946777344, 0.032333824157714845, 0.03123200035095215, 0.03120025634765625, 0.03220889663696289, 0.030846975326538087, 0.03099238395690918, 0.031793216705322265, 0.03191289520263672, 0.03168972778320313, 0.03189452743530274, 0.031357952117919925, 0.03161497688293457, 0.03141222381591797, 0.03162419128417969, 0.03247206497192383, 0.03136614418029785, 0.03230003356933594, 0.031297536849975584, 0.031748096466064454, 0.031438848495483396, 0.031267839431762694, 0.03257651138305664, 0.03226214218139648, 0.03240140914916992, 0.032522239685058595, 0.031768575668334964, 0.03222937774658203, 0.031281152725219724, 0.03181158447265625, 0.03243318557739258, 0.03237065505981445, 0.03209625625610352, 0.032123905181884765, 0.0323061752319336, 0.03155660820007324, 0.03197235107421875, 0.03072204780578613, 0.03125657653808594, 0.03273625564575195, 0.03249049758911133, 0.03213516616821289, 0.03240140914916992, 0.03260211181640625, 0.03272192001342773, 0.03242803192138672, 0.032074752807617186, 0.03122483253479004, 0.031591424942016604, 0.0325591049194336, 0.032315391540527344, 0.03200102233886719, 0.03083673667907715, 0.031438848495483396, 0.03233587265014649, 0.03115724754333496, 0.03151872062683105, 0.03217407989501953, 0.03226214218139648, 0.03194879913330078, 0.032494590759277346, 0.031237119674682616, 0.03178803253173828, 0.03183718490600586, 0.03150540733337402, 0.03117568016052246, 0.03184639930725098, 0.03264409637451172, 0.03126067161560059, 0.03117158317565918, 0.031180799484252928, 0.03129548835754394, 0.03158937644958496, 0.031410175323486327, 0.031441919326782225, 0.031067136764526368, 0.031058944702148438, 0.030838783264160157, 0.031508480072021484, 0.030988288879394532, 0.030723072052001952, 0.031066112518310547, 0.03148595237731933, 0.030842880249023437, 0.030905344009399413, 0.030966783523559572, 0.03119206428527832, 0.03138457679748535, 0.031222784042358398, 0.031118335723876952, 0.031084543228149415, 0.030866432189941406, 0.031120384216308594, 0.0310118408203125, 0.031453184127807614, 0.031121408462524414, 0.031188991546630858, 0.031335424423217774, 0.03230822372436523, 0.03239014434814453, 0.03125862312316895, 0.031170560836791993, 0.0314654712677002, 0.031152128219604492, 0.03126681518554687, 0.03136204719543457, 0.03151667213439941, 0.03101286315917969, 0.03119206428527832, 0.03131289672851562, 0.03143987274169922, 0.031199327468872072, 0.031212448120117187, 0.03137433624267578, 0.03094528007507324, 0.031117311477661135, 0.03118796730041504, 0.031085567474365236, 0.031180799484252928, 0.031078399658203124, 0.032091136932373046, 0.03240857696533203, 0.03177676773071289, 0.031098880767822266, 0.03121664047241211, 0.03185766410827637, 0.031188991546630858, 0.032290817260742184, 0.032543743133544925, 0.0321710090637207, 0.03119001579284668, 0.03130060768127441, 0.031046688079833986, 0.03083977508544922, 0.03168870353698731, 0.031458303451538085, 0.031060991287231447, 0.03116646385192871, 0.031123455047607423, 0.03129446411132813, 0.031545343399047854, 0.031024127960205077, 0.03240140914916992, 0.03118284797668457, 0.031100927352905275, 0.03121664047241211, 0.031731712341308595, 0.0321003532409668, 0.033807361602783206, 0.032625663757324216, 0.03197747230529785, 0.03180646324157715, 0.031135744094848632, 0.031243263244628908, 0.03131494331359863, 0.03114188766479492, 0.03192422485351563, 0.03227443313598633, 0.03244441604614258, 0.03214131164550781, 0.03206041717529297, 0.03206655883789063, 0.0322529296875, 0.032080894470214845, 0.03176345634460449, 0.031336448669433595, 0.03194572830200195, 0.032702465057373044, 0.030904319763183592, 0.03104256057739258, 0.031558656692504884, 0.03231948852539063, 0.03178188705444336, 0.03139891242980957, 0.030899200439453125, 0.0313384952545166, 0.03118694305419922, 0.030699520111083983, 0.031072256088256835, 0.031322111129760744, 0.03120947265625, 0.03115110397338867, 0.031122432708740235, 0.031308799743652346, 0.03125043106079101, 0.031197183609008788, 0.03196723175048828, 0.031143936157226562, 0.0314839038848877, 0.030688255310058594, 0.030906368255615234, 0.03121971130371094, 0.032309249877929686, 0.032110591888427735, 0.0318525447845459, 0.03139583969116211, 0.03131699180603027, 0.03097907257080078, 0.03117465591430664, 0.0317071361541748, 0.03214950561523437, 0.031525888442993165, 0.031109119415283205, 0.03127705574035645, 0.030880767822265624, 0.03103027153015137, 0.031083520889282228, 0.031212543487548827, 0.03105996894836426, 0.030861312866210938, 0.0311910400390625, 0.031131647109985353, 0.031222784042358398, 0.031113216400146484, 0.03101286315917969, 0.031068159103393556, 0.03126067161560059, 0.03136716842651367, 0.032331775665283204, 0.03219046401977539, 0.03139686393737793, 0.032369663238525394, 0.031111167907714843, 0.031119359970092773, 0.03232665634155273, 0.034541568756103515, 0.03147776031494141, 0.0325591049194336, 0.03203379058837891, 0.03159244728088379, 0.031226879119873048, 0.03182489585876465, 0.0323061752319336, 0.032099327087402346, 0.03250175857543945, 0.03320217514038086, 0.03259910583496094, 0.032092159271240234, 0.03093708801269531, 0.030842880249023437, 0.03167231941223145, 0.03098214340209961, 0.03075584030151367, 0.031510528564453126, 0.03127807998657227, 0.031611904144287106, 0.031716352462768556, 0.03225600051879883, 0.032745471954345705, 0.031337472915649416, 0.03228876876831055, 0.032517120361328124, 0.031748096466064454, 0.032215038299560544, 0.03213107299804688, 0.03189248085021973, 0.03206758499145508, 0.03207680130004883, 0.031107072830200196, 0.03223961639404297, 0.0319498233795166, 0.03175833511352539, 0.03142758369445801, 0.031169536590576172, 0.031884288787841795, 0.031784959793090824, 0.0321638412475586, 0.032368640899658206, 0.032217086791992186, 0.032121856689453124, 0.03200204849243164, 0.031161344528198243, 0.03158220863342285, 0.031307775497436525, 0.031680511474609374, 0.03241676712036133, 0.03316633605957031, 0.03117568016052246, 0.03118387222290039, 0.031185920715332032, 0.03125862312316895, 0.030835712432861328, 0.03119513511657715, 0.032748542785644534, 0.032094207763671875, 0.03153715133666992, 0.03196620750427246, 0.0307640323638916, 0.031056896209716797, 0.03230003356933594, 0.03258777618408203, 0.031352832794189454, 0.03381964874267578, 0.0322979850769043, 0.03206758499145508, 0.032145408630371096, 0.03118796730041504, 0.0313436164855957, 0.030849023818969725, 0.031139839172363282, 0.031286272048950195, 0.031083520889282228, 0.030842880249023437, 0.030904319763183592, 0.03230822372436523, 0.03214233779907227, 0.03114188766479492, 0.031032320022583007, 0.03143168067932129, 0.031136768341064453, 0.031124479293823244, 0.03122585678100586, 0.031128576278686523, 0.03115724754333496, 0.03132825660705566, 0.03117568016052246, 0.03136511993408203, 0.03128832054138184, 0.03250175857543945, 0.031101951599121092, 0.031121408462524414, 0.03117977523803711, 0.03214233779907227, 0.03243315124511719, 0.0307957763671875, 0.031087615966796874, 0.031081504821777343, 0.03124732780456543, 0.030732288360595703, 0.030706687927246092, 0.03112652778625488, 0.031230976104736328, 0.03231027221679687, 0.03232153701782227, 0.03190169525146484, 0.03113471984863281, 0.031409151077270506, 0.03128319931030273, 0.031041536331176758, 0.031128576278686523, 0.031373311996459964, 0.03119308853149414, 0.03119615936279297, 0.03059404754638672, 0.03135078430175781, 0.031120384216308594, 0.030917631149291993, 0.030922752380371094, 0.03137945556640625, 0.03122790336608887, 0.03159040069580078, 0.03182387161254883, 0.031069183349609376, 0.031236095428466795, 0.0311592960357666, 0.03138764762878418, 0.03120025634765625, 0.031281152725219724, 0.031066112518310547, 0.031093759536743162, 0.031047679901123046, 0.031038463592529295, 0.031264768600463864, 0.03127603149414063, 0.03283865737915039, 0.03155148887634277, 0.030871551513671876, 0.031140863418579103, 0.03080294418334961, 0.03083776092529297, 0.031344671249389645, 0.031127519607543945, 0.03102720069885254, 0.03143168067932129, 0.030855167388916017, 0.031075328826904298, 0.03124019241333008, 0.03170816040039062, 0.03116646385192871, 0.03147776031494141, 0.03158016014099121, 0.03127705574035645, 0.0309616641998291, 0.03121664047241211, 0.03149619293212891, 0.03115519905090332, 0.031129600524902344, 0.031254528045654296, 0.03098214340209961, 0.03115724754333496, 0.031021055221557618, 0.03125555229187012, 0.03099443244934082, 0.03115110397338867, 0.03268710327148437, 0.031269887924194335, 0.031156223297119142, 0.03122380828857422, 0.030963712692260743, 0.03096883201599121, 0.030924800872802735, 0.03079680061340332, 0.031106048583984375, 0.03119206428527832, 0.031488000869750975, 0.031197183609008788, 0.031169536590576172, 0.03114188766479492, 0.030920703887939452, 0.03140812873840332, 0.031007743835449218, 0.030834688186645507, 0.031682559967041016, 0.03179212760925293, 0.031158304214477538, 0.0310732479095459, 0.031094783782958983, 0.03121151924133301, 0.033040382385253905, 0.032282623291015625, 0.03127705574035645, 0.031562751770019534, 0.031143936157226562, 0.030887935638427736, 0.03142860794067383, 0.032347137451171876, 0.03133235168457031, 0.03106412887573242, 0.031394752502441406, 0.03115827178955078, 0.03120742416381836, 0.031131647109985353, 0.031024127960205077, 0.03099238395690918, 0.03118387222290039, 0.03116851234436035, 0.031088640213012695, 0.031154176712036134, 0.031323135375976564, 0.03080703926086426, 0.031373311996459964, 0.0319180793762207, 0.03148595237731933, 0.031074304580688477, 0.031486976623535154, 0.032996353149414064, 0.031854591369628905, 0.03224883270263672, 0.03206860733032227, 0.03120742416381836, 0.031355903625488284, 0.031324159622192385, 0.03115519905090332, 0.031222784042358398, 0.03146751976013184, 0.03122380828857422, 0.031154176712036134, 0.031120384216308594, 0.030884864807128907, 0.03116339111328125, 0.03165286445617676, 0.03172761535644531, 0.031205375671386718, 0.03129651260375976, 0.030717952728271485, 0.030822399139404297, 0.03130572891235352, 0.030661632537841797, 0.030886911392211915, 0.0316753921508789, 0.03121561622619629, 0.031025152206420898, 0.031071231842041015, 0.032215038299560544, 0.03186380767822266, 0.031132671356201173, 0.031137792587280274, 0.0314839038848877, 0.03143270492553711, 0.030855167388916017, 0.030918655395507814, 0.03099545669555664, 0.031457279205322264, 0.031099903106689454, 0.031115264892578126, 0.03100057601928711, 0.031340543746948245, 0.03189657592773437, 0.032350208282470705, 0.03128422355651855, 0.031335424423217774, 0.031230976104736328, 0.03118796730041504, 0.03135385513305664, 0.030897151947021483, 0.030895103454589845, 0.03118387222290039, 0.03183923149108887, 0.03139379119873047, 0.03122790336608887, 0.031212543487548827, 0.03134771156311035, 0.03115007972717285, 0.031899648666381834, 0.03150028800964356, 0.031046655654907225, 0.031226879119873048, 0.03127807998657227, 0.031036415100097657, 0.03137945556640625, 0.031286272048950195, 0.03103539276123047, 0.030917631149291993, 0.03127603149414063, 0.03183718490600586, 0.03124940872192383, 0.031090688705444337, 0.03120742416381836, 0.03129651260375976, 0.03116543960571289, 0.031119359970092773, 0.031406080245971676, 0.03270348739624023, 0.03571712112426758, 0.0332564468383789, 0.03268915176391601, 0.030868480682373047, 0.031094783782958983, 0.031476736068725586, 0.03254272079467774, 0.030916608810424805, 0.031286272048950195, 0.033732608795166014, 0.0323389778137207, 0.031161312103271485, 0.03204915237426758, 0.031942655563354495, 0.03132723236083984, 0.03136614418029785, 0.03139583969116211, 0.0312412166595459, 0.031291391372680666, 0.031230976104736328, 0.031453184127807614, 0.03119923210144043, 0.031237119674682616, 0.031869951248168944, 0.03163545608520508, 0.032966655731201173, 0.03235942459106445, 0.031297536849975584, 0.031716352462768556]",tokens/s,31.68828729751602,,,,,,gemma,MB,4750.635008,6422.003712,0.0,5792.333824,5786.358272,s,1,11.19784375,11.19784375,0.0,11.19784375,11.19784375,11.19784375,11.19784375,[11.19784375],,kWh,5.108825101388499e-05,2.7984937504828067e-05,7.323616970000141e-05,0.00015230935821871445, @@ -1344,48 +1326,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,4118.462464,5287.44448,0.0,4640.997376,4330.668544,s,10,3.198638458251953,0.31986384582519534,0.0010414384981667266,0.31984342956542966,0.32109922180175776,0.32110626678466797,0.3211119027709961,"[0.31871585083007814, 0.32094384765625, 0.31872210693359376, 0.31898590087890627, 0.3198691711425781, 0.3184337158203125, 0.31981768798828125, 0.320939208984375, 0.3211133117675781, 0.32109765625]",tokens/s,800.3405303264666,kWh,3.758956124592159e-06,2.0597393240023246e-06,1.869149064762099e-05,2.451018609621547e-05,tokens/kWh,10444637.139639182,MB,4118.462464,5287.44448,0.0,4640.997376,4515.667456,s,10,175.24420703125003,17.524420703125003,0.0015038867693539047,17.524218750000003,17.5263189453125,17.526761035156248,17.52711470703125,"[17.526220703125, 17.5231015625, 17.527203125, 17.522798828125, 17.52478515625, 17.52365234375, 17.52528515625, 17.52256640625, 17.525375, 17.52321875]",tokens/s,3.5949833131297555,kWh,0.00020693968945389822,0.00011342005025165543,0.0010026636597691928,0.0013230233994747467,tokens/kWh,47618.20541119047,,s,629,177.84587469482406,0.28274383894248684,0.03805270523867711,0.27815731811523436,0.27831808471679687,0.27843297729492184,0.5974869262695313,"[0.2782863464355469, 0.2779576416015625, 0.278139892578125, 0.2781542358398438, 0.27818701171875, 0.2780078125, 0.2780887145996094, 0.27825048828125, 0.2781737060546875, 0.27795660400390626, 0.27808255004882815, 0.2781337585449219, 0.2782597045898437, 0.2780682373046875, 0.278076416015625, 0.27815936279296877, 0.27807949829101564, 0.27803750610351563, 0.27818496704101564, 0.27820440673828123, 0.27828939819335935, 0.27808154296875, 0.2783221740722656, 0.2781972351074219, 0.27828939819335935, 0.2780426330566406, 0.2782474365234375, 0.27815115356445314, 0.2782443542480469, 0.27812966918945314, 0.27824844360351564, 0.27823410034179685, 0.278192138671875, 0.27812353515625, 0.2781757507324219, 0.2781767578125, 0.27821466064453126, 0.2781388854980469, 0.2782740478515625, 0.2782474365234375, 0.2783907775878906, 0.2781747131347656, 0.278213623046875, 0.27813479614257813, 0.2781890563964844, 0.2782054443359375, 0.27834060668945315, 0.2782494812011719, 0.27826791381835936, 0.2781634521484375, 0.2782423095703125, 0.2781941833496094, 0.27816754150390627, 0.2781051025390625, 0.27813784790039064, 0.2783365173339844, 0.2782494812011719, 0.27823007202148436, 0.278236083984375, 0.27824639892578124, 0.2782371826171875, 0.27838668823242185, 0.599498779296875, 0.2780405883789063, 0.27799038696289063, 0.27804672241210937, 0.27807537841796875, 0.278002685546875, 0.27792691040039064, 0.2779658203125, 0.27797708129882814, 0.27801089477539065, 0.27793304443359373, 0.278181884765625, 0.2780989379882812, 0.2781317138671875, 0.2779873352050781, 0.2780405883789063, 0.27799038696289063, 0.27803750610351563, 0.27818496704101564, 0.2782443542480469, 0.2780999755859375, 0.2780477294921875, 0.2779596862792969, 0.2780641174316406, 0.27814093017578123, 0.27813479614257813, 0.2779473876953125, 0.2779862976074219, 0.27849215698242186, 0.2783109130859375, 0.2780006408691406, 0.2781685791015625, 0.2780159912109375, 0.27819622802734373, 0.27800473022460936, 0.2781419372558594, 0.2782095336914063, 0.278139892578125, 0.2781808776855469, 0.2781440124511719, 0.2782791748046875, 0.27829144287109375, 0.27804364013671873, 0.2783303527832031, 0.27811431884765625, 0.27809588623046877, 0.27827301025390627, 0.27833447265625, 0.27831808471679687, 0.2782095336914063, 0.2783221740722656, 0.27832525634765626, 0.2782023620605469, 0.2782791748046875, 0.2780692443847656, 0.278223876953125, 0.2782822265625, 0.2782843017578125, 0.278086669921875, 0.27830169677734373, 0.27828326416015625, 0.2781757507324219, 0.2781951904296875, 0.5972797241210938, 0.278181884765625, 0.27797503662109374, 0.27807437133789065, 0.2781481018066406, 0.2785935363769531, 0.27815115356445314, 0.2780415954589844, 0.2780426330566406, 0.27796990966796875, 0.27794638061523436, 0.2780631103515625, 0.27804568481445313, 0.2781419372558594, 0.2779852905273438, 0.2780979309082031, 0.27797708129882814, 0.2782310485839844, 0.27858432006835937, 0.27852801513671877, 0.27819314575195314, 0.2780692443847656, 0.27805902099609375, 0.2782637939453125, 0.2782218322753906, 0.278223876953125, 0.27796173095703125, 0.2782095336914063, 0.2780631103515625, 0.2781521911621094, 0.27807333374023435, 0.27829144287109375, 0.2782689208984375, 0.27826483154296877, 0.2780702819824219, 0.27833343505859376, 0.2782156677246094, 0.27824127197265625, 0.2781204833984375, 0.2782197570800781, 0.27825665283203127, 0.27865805053710935, 0.278181884765625, 0.27834368896484374, 0.2782474365234375, 0.2781951904296875, 0.27816754150390627, 0.27826174926757813, 0.2783006591796875, 0.2784194641113281, 0.2780559387207031, 0.2782791748046875, 0.27817062377929686, 0.27826791381835936, 0.2785535888671875, 0.27827609252929686, 0.278213623046875, 0.278287353515625, 0.278139892578125, 0.27827096557617187, 0.27832421875, 0.27824026489257814, 0.278302734375, 0.5975675048828125, 0.27800473022460936, 0.27799038696289063, 0.2780426330566406, 0.2780979309082031, 0.2780487670898438, 0.278002685546875, 0.27813580322265624, 0.27803851318359374, 0.27799655151367186, 0.27796890258789064, 0.27797708129882814, 0.27799655151367186, 0.27817266845703126, 0.27788595581054687, 0.27818292236328124, 0.27807949829101564, 0.2780968933105469, 0.2778961791992188, 0.2782248840332031, 0.2780928039550781, 0.2779934692382813, 0.27804672241210937, 0.27817166137695315, 0.27844198608398435, 0.27823309326171874, 0.2780631103515625, 0.27873587036132813, 0.278139892578125, 0.278181884765625, 0.27800473022460936, 0.2782525329589844, 0.27819110107421874, 0.2782320556640625, 0.2780415954589844, 0.2780907592773437, 0.2780498046875, 0.2780948486328125, 0.2780702819824219, 0.27808563232421873, 0.2780579833984375, 0.27810406494140627, 0.27798220825195313, 0.2784460754394531, 0.2780282897949219, 0.27806719970703125, 0.2782033996582031, 0.2781634521484375, 0.27819931030273437, 0.278265869140625, 0.2780999755859375, 0.27819931030273437, 0.2781808776855469, 0.27829351806640623, 0.27804568481445313, 0.27819314575195314, 0.2782054443359375, 0.2781214599609375, 0.2781460571289063, 0.27821978759765625, 0.2782689208984375, 0.27820748901367187, 0.27839590454101565, 0.5986314086914063, 0.2779955139160156, 0.2779934692382813, 0.2780477294921875, 0.27818496704101564, 0.2780887145996094, 0.2780159912109375, 0.278086669921875, 0.27812863159179685, 0.2779811706542969, 0.277939208984375, 0.27809588623046877, 0.2780610656738281, 0.27805490112304687, 0.2779852905273438, 0.27809381103515624, 0.278002685546875, 0.2780948486328125, 0.27802420043945314, 0.27817062377929686, 0.27817984008789065, 0.2781481018066406, 0.27805184936523436, 0.2781603698730469, 0.27810406494140627, 0.27812966918945314, 0.2779504699707031, 0.27823614501953126, 0.27807537841796875, 0.27818496704101564, 0.27801089477539065, 0.27824639892578124, 0.27826174926757813, 0.27823309326171874, 0.2781890563964844, 0.2786130065917969, 0.27832012939453127, 0.2780692443847656, 0.27823513793945315, 0.2781481018066406, 0.27836007690429687, 0.278192138671875, 0.2781440124511719, 0.2783109130859375, 0.27816549682617187, 0.27817880249023436, 0.2781562805175781, 0.27825357055664063, 0.2782853088378906, 0.2783088684082031, 0.27805490112304687, 0.27821774291992185, 0.2783375244140625, 0.27827508544921875, 0.2781112365722656, 0.27818597412109375, 0.2782791748046875, 0.27834878540039065, 0.2782515258789062, 0.2782843017578125, 0.27832525634765626, 0.27826995849609376, 0.2782822265625, 0.596843505859375, 0.27809588623046877, 0.27798220825195313, 0.27794638061523436, 0.27807333374023435, 0.2780569458007813, 0.2779648132324219, 0.27848193359375, 0.27803237915039064, 0.2779914245605469, 0.27780914306640625, 0.27796685791015624, 0.27810098266601563, 0.2780487670898438, 0.2779525146484375, 0.2781634521484375, 0.27802725219726565, 0.27810098266601563, 0.27794842529296876, 0.2782443542480469, 0.2781112365722656, 0.27802420043945314, 0.2780006408691406, 0.27815936279296877, 0.27825457763671874, 0.27819009399414063, 0.27797607421875, 0.2783078308105469, 0.27816754150390627, 0.27826174926757813, 0.27806002807617186, 0.27823410034179685, 0.2781603698730469, 0.278139892578125, 0.278066162109375, 0.2782218322753906, 0.27822796630859375, 0.2780313720703125, 0.2780968933105469, 0.27828326416015625, 0.27814093017578123, 0.27810714721679686, 0.2780415954589844, 0.27877169799804685, 0.27826483154296877, 0.27813784790039064, 0.27809381103515624, 0.2782853088378906, 0.27819931030273437, 0.2782525329589844, 0.27820440673828123, 0.2781839294433594, 0.2781767578125, 0.2782003173828125, 0.27814910888671873, 0.2783088684082031, 0.2781317138671875, 0.2781327209472656, 0.27802215576171874, 0.27833139038085936, 0.2786385803222656, 0.27829144287109375, 0.2782320556640625, 0.598614013671875, 0.2782054443359375, 0.27787774658203124, 0.27800985717773435, 0.2781419372558594, 0.27807949829101564, 0.2779596862792969, 0.27814501953125, 0.27807949829101564, 0.2780159912109375, 0.27796786499023435, 0.2780712890625, 0.2780405883789063, 0.2782515258789062, 0.27796685791015624, 0.278582275390625, 0.27811224365234377, 0.2780682373046875, 0.27798016357421873, 0.27817779541015625, 0.2780948486328125, 0.2780682373046875, 0.2780569458007813, 0.27831808471679687, 0.2780477294921875, 0.2781747131347656, 0.27801907348632815, 0.27813580322265624, 0.2781481018066406, 0.27813067626953125, 0.27826995849609376, 0.27831195068359377, 0.27821261596679686, 0.2783068237304688, 0.278066162109375, 0.27801907348632815, 0.27857614135742187, 0.27815115356445314, 0.2780702819824219, 0.2782105712890625, 0.27823822021484373, 0.278213623046875, 0.2780057678222656, 0.2782863464355469, 0.27815731811523436, 0.27815731811523436, 0.27811019897460937, 0.2782525329589844, 0.27829248046875, 0.27865292358398436, 0.2781532287597656, 0.27865496826171876, 0.27821261596679686, 0.2782494812011719, 0.2780498046875, 0.27819622802734373, 0.2781327209472656, 0.27826483154296877, 0.2780559387207031, 0.2783354797363281, 0.27823410034179685, 0.27827609252929686, 0.27823410034179685, 0.5994219360351563, 0.27807333374023435, 0.27835086059570313, 0.2780334167480469, 0.2780498046875, 0.27802420043945314, 0.2778890380859375, 0.2780139465332031, 0.27811224365234377, 0.27798220825195313, 0.27774362182617185, 0.2780610656738281, 0.2780088195800781, 0.27805184936523436, 0.27796173095703125, 0.27803750610351563, 0.2779873352050781, 0.27811328125, 0.2778306579589844, 0.2781440124511719, 0.278066162109375, 0.2780692443847656, 0.2779381713867187, 0.2784471130371094, 0.27798016357421873, 0.2783211669921875, 0.27794638061523436, 0.27829248046875, 0.27817880249023436, 0.2781665344238281, 0.278002685546875, 0.2782085266113281, 0.27816549682617187, 0.2781583251953125, 0.2779443054199219, 0.27813784790039064, 0.2780887145996094, 0.27800167846679685, 0.2784716796875, 0.27815115356445314, 0.2781337585449219, 0.27812966918945314, 0.2782371826171875, 0.27828121948242185, 0.27816754150390627, 0.27813784790039064, 0.27805184936523436, 0.27825457763671874, 0.2781808776855469, 0.27884747314453123, 0.27812863159179685, 0.278223876953125, 0.27812966918945314, 0.27817266845703126, 0.27808050537109374, 0.27827609252929686, 0.2781603698730469, 0.2782689208984375, 0.27807537841796875, 0.2783631286621094, 0.27827609252929686, 0.2781470642089844, 0.27817266845703126, 0.59964208984375, 0.2780498046875, 0.27801190185546876, 0.278118408203125, 0.2781603698730469, 0.2780948486328125, 0.2779371643066406, 0.2780354614257812, 0.27807437133789065, 0.2778419189453125, 0.2779832458496094, 0.2781214599609375, 0.27814093017578123, 0.27803851318359374, 0.27801702880859375, 0.2781890563964844, 0.27814297485351563, 0.27809588623046877, 0.27804364013671873, 0.2782105712890625, 0.27819110107421874, 0.27810098266601563, 0.2780559387207031, 0.27815731811523436, 0.27818292236328124, 0.27819827270507813, 0.27798834228515623, 0.2782371826171875, 0.2781388854980469, 0.2783160400390625, 0.27808563232421873, 0.278508544921875, 0.278265869140625, 0.278192138671875, 0.27808154296875, 0.2781665344238281, 0.2782167053222656, 0.2780569458007813, 0.27818292236328124, 0.2782689208984375, 0.27824331665039065, 0.27820748901367187, 0.27803033447265624, 0.2783057861328125, 0.27820440673828123, 0.2782423095703125, 0.2782597045898437, 0.27829861450195315, 0.278245361328125, 0.27833343505859376, 0.27811019897460937, 0.27825357055664063, 0.2783078308105469, 0.27828939819335935, 0.2780917663574219, 0.27822283935546877, 0.27825457763671874, 0.27847885131835937, 0.2783088684082031, 0.2783139953613281, 0.2784849853515625, 0.2782597045898437, 0.27825357055664063, 0.5996759033203125, 0.27808563232421873, 0.27793408203125, 0.27804568481445313, 0.2780579833984375, 0.27797607421875, 0.27798220825195313, 0.27805902099609375, 0.278086669921875, 0.2780498046875, 0.2779084777832031, 0.278076416015625, 0.27807437133789065, 0.27802520751953125, 0.2778480529785156, 0.2780989379882812, 0.2781051025390625, 0.2780712890625, 0.278044677734375, 0.27825665283203127, 0.2782033996582031, 0.27797607421875, 0.2780569458007813, 0.27826791381835936, 0.2781317138671875, 0.27831808471679687, 0.27803750610351563, 0.27817062377929686, 0.2781214599609375, 0.27823513793945315, 0.2780989379882812, 0.2782003173828125, 0.27824844360351564, 0.2781337585449219, 0.27802215576171874, 0.27807437133789065, 0.27814501953125, 0.27820645141601563, 0.2780979309082031, 0.27818597412109375, 0.2781695861816406, 0.27815731811523436, 0.27800372314453126, 0.27832525634765626, 0.27814093017578123, 0.27812762451171874, 0.2781757507324219, 0.27833242797851565, 0.27819110107421874, 0.2783999938964844, 0.27811431884765625, 0.2782023620605469, 0.2781941833496094, 0.27828839111328124, 0.2781317138671875, 0.27836517333984373, 0.27819622802734373, 0.27817984008789065, 0.27813784790039064, 0.27834573364257814, 0.27821978759765625, 0.2783170471191406, 0.278245361328125]",tokens/s,3.5367702572766255,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch @@ -2451,6 +2418,33 @@ ChildProcessError: Traceback (most recent call last): OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,MB,2108.841984,5971.116032,0.0,5324.668928,5080.87552,s,10,3.166915313720703,0.31669153137207035,0.0006471042681215324,0.3165300140380859,0.317839404296875,0.3178718963623047,0.31789789001464847,"[0.3165837707519531, 0.3160324401855469, 0.316489501953125, 0.3163083801269531, 0.31657052612304687, 0.3162070617675781, 0.31696994018554686, 0.3160171203613281, 0.31783218383789064, 0.3179043884277344]",tokens/s,808.3575802954898,kWh,3.734563280724817e-06,2.0463760689267472e-06,1.9076438872250856e-05,2.485737822190242e-05,tokens/kWh,10298753.058938146,MB,2108.841984,5971.116032,0.0,5324.668928,5264.397312,s,10,173.206740234375,17.320674023437498,0.0018067612818664913,17.32021484375,17.322722851562503,17.32348837890625,17.32410080078125,"[17.319333984375, 17.32053515625, 17.31989453125, 17.318197265625, 17.31883203125, 17.319619140625, 17.322552734375, 17.321099609375, 17.32425390625, 17.322421875]",tokens/s,3.6372718472012955,kWh,0.00020451679237393867,0.00011209220585674529,0.0010427547717031527,0.0013593637699338367,tokens/kWh,46345.210453171305,,s,629,175.7822831726073,0.2794630893046223,0.03762287358215295,0.2749214782714844,0.2751193115234375,0.27523646240234373,0.5911559643554688,"[0.2748784790039063, 0.27467367553710936, 0.2747135925292969, 0.2749491271972656, 0.27461016845703123, 0.2748641357421875, 0.27486514282226565, 0.27473715209960936, 0.274845703125, 0.27484262084960936, 0.2748590087890625, 0.27469415283203125, 0.2747821960449219, 0.2748784790039063, 0.27476275634765623, 0.27482623291015623, 0.2747105407714844, 0.274850830078125, 0.2747760620117187, 0.2747586669921875, 0.2748231811523438, 0.2747474060058594, 0.27487744140625, 0.2750013427734375, 0.2748467102050781, 0.27495526123046876, 0.27497268676757813, 0.27526861572265626, 0.2748641357421875, 0.2748160095214844, 0.27477297973632814, 0.2750904235839844, 0.275009521484375, 0.27487335205078123, 0.27486004638671874, 0.27478118896484377, 0.2747125854492187, 0.274914306640625, 0.2747955322265625, 0.27487335205078123, 0.2750126037597656, 0.27509963989257813, 0.2748374938964844, 0.2749931640625, 0.2751231994628906, 0.2749931640625, 0.27523275756835935, 0.2749440002441406, 0.27532492065429687, 0.27521331787109377, 0.274904052734375, 0.27517132568359376, 0.2750013427734375, 0.2749542541503906, 0.27493991088867187, 0.27488153076171873, 0.27494808959960937, 0.27505050659179686, 0.2748374938964844, 0.2749859924316406, 0.275051513671875, 0.27498291015625, 0.5910169677734375, 0.2748876647949219, 0.27511602783203126, 0.27487539672851563, 0.2748661804199219, 0.27482009887695313, 0.27477197265625, 0.2747750549316406, 0.27479244995117186, 0.2749716491699219, 0.27480062866210936, 0.27487026977539064, 0.274808837890625, 0.27494808959960937, 0.2749910888671875, 0.2747852783203125, 0.2747361145019531, 0.27484262084960936, 0.2748467102050781, 0.27510784912109376, 0.2748958740234375, 0.27481396484375, 0.2749491271972656, 0.27508120727539065, 0.27473919677734376, 0.27484774780273435, 0.27491226196289065, 0.27473715209960936, 0.27471463012695313, 0.27496652221679685, 0.27483544921875, 0.2747064208984375, 0.2750853271484375, 0.2747002868652344, 0.2747832336425781, 0.275125244140625, 0.27492556762695314, 0.2749081726074219, 0.2749634704589844, 0.27483648681640627, 0.2749020080566406, 0.2750320739746094, 0.2749306945800781, 0.2751682434082031, 0.274946044921875, 0.27498086547851563, 0.274951171875, 0.2749306945800781, 0.2748385314941406, 0.27479449462890626, 0.2751457214355469, 0.27537203979492186, 0.27509451293945314, 0.27496139526367186, 0.2750064697265625, 0.27482009887695313, 0.27509356689453124, 0.2749070739746094, 0.27508642578125, 0.2751886291503906, 0.2749306945800781, 0.27533108520507815, 0.2750474243164063, 0.59137841796875, 0.2748907470703125, 0.2747914123535156, 0.2748313598632813, 0.2749941711425781, 0.27480474853515624, 0.27486309814453125, 0.2748610534667969, 0.27473202514648437, 0.27523583984375, 0.27483544921875, 0.2749849548339844, 0.2746798095703125, 0.275420166015625, 0.27487539672851563, 0.27475762939453124, 0.27493170166015624, 0.27479244995117186, 0.2748170166015625, 0.27492044067382815, 0.27487335205078123, 0.27472280883789063, 0.2748989562988281, 0.2749245300292969, 0.2748661804199219, 0.27478732299804687, 0.27480062866210936, 0.2748661804199219, 0.27485592651367186, 0.2748968811035156, 0.2749091796875, 0.2750064697265625, 0.27490609741210936, 0.27480062866210936, 0.2749798278808594, 0.2747555847167969, 0.2752122802734375, 0.27495834350585935, 0.2748692626953125, 0.2747965393066406, 0.27490609741210936, 0.2749306945800781, 0.274808837890625, 0.2750402526855469, 0.2749696960449219, 0.27509442138671875, 0.2749982604980469, 0.2749347839355469, 0.2748436584472656, 0.2750433349609375, 0.27503411865234373, 0.2749245300292969, 0.2749644775390625, 0.2749819030761719, 0.2748231811523438, 0.27495834350585935, 0.2749214782714844, 0.2749378662109375, 0.27481497192382814, 0.27504537963867187, 0.27529522705078124, 0.2750320739746094, 0.27503411865234373, 0.5913456420898437, 0.2747432861328125, 0.27471463012695313, 0.274808837890625, 0.2747432861328125, 0.27475149536132815, 0.2747965393066406, 0.2746982421875, 0.274798583984375, 0.27487744140625, 0.27507611083984373, 0.27470132446289064, 0.27465216064453124, 0.27487335205078123, 0.274830322265625, 0.2748713073730469, 0.274840576171875, 0.2747258911132813, 0.27484262084960936, 0.2746705932617188, 0.2747340698242188, 0.27486004638671874, 0.27495834350585935, 0.2748948364257813, 0.2748671875, 0.27479244995117186, 0.27482110595703124, 0.2748160095214844, 0.27477197265625, 0.2747535400390625, 0.2749010009765625, 0.27479449462890626, 0.2751201171875, 0.27503103637695314, 0.2749521789550781, 0.275061767578125, 0.2749706115722656, 0.2748713073730469, 0.2749276123046875, 0.2748272705078125, 0.2749644775390625, 0.2747965393066406, 0.2749010009765625, 0.2749573059082031, 0.27494195556640627, 0.27487539672851563, 0.27480679321289064, 0.27482009887695313, 0.275019775390625, 0.2750074768066406, 0.27485498046875, 0.2748620300292969, 0.27502899169921874, 0.2749163513183594, 0.2753208312988281, 0.27496038818359375, 0.27484466552734377, 0.27511602783203126, 0.2754334716796875, 0.274951171875, 0.2750689392089844, 0.27506689453125, 0.2749266052246094, 0.5915023193359376, 0.27469003295898436, 0.27464703369140625, 0.27473919677734376, 0.27470950317382814, 0.2747801513671875, 0.2747607116699219, 0.27479757690429685, 0.2747105407714844, 0.2748528747558594, 0.27475967407226565, 0.2747330627441406, 0.2746777648925781, 0.2749992980957031, 0.2749941711425781, 0.27480780029296875, 0.2747218017578125, 0.2747658386230469, 0.2748866577148438, 0.27479244995117186, 0.2747268981933594, 0.27486004638671874, 0.2748671875, 0.27482623291015623, 0.2747914123535156, 0.27491226196289065, 0.27511911010742185, 0.274845703125, 0.27484262084960936, 0.2749163513183594, 0.2753403015136719, 0.27483544921875, 0.27476275634765623, 0.27486004638671874, 0.27502899169921874, 0.274724853515625, 0.27497882080078123, 0.27488153076171873, 0.27493887329101563, 0.2747637634277344, 0.27508837890625, 0.27523687744140624, 0.2749593505859375, 0.27493374633789064, 0.274850830078125, 0.2753116149902344, 0.27480780029296875, 0.2749286499023437, 0.27506381225585935, 0.2749368286132812, 0.27501055908203126, 0.2748190612792969, 0.2750545959472656, 0.27506689453125, 0.2750689392089844, 0.27508120727539065, 0.27498291015625, 0.2749368286132812, 0.2751037292480469, 0.2750126037597656, 0.27486004638671874, 0.2750802001953125, 0.2750535583496094, 0.591194091796875, 0.2748436584472656, 0.2747197570800781, 0.27485696411132815, 0.27480780029296875, 0.27491326904296876, 0.27488357543945313, 0.2747750549316406, 0.27480575561523435, 0.2748436584472656, 0.2746849365234375, 0.2747965393066406, 0.274766845703125, 0.2749091796875, 0.2749153137207031, 0.27480679321289064, 0.274798583984375, 0.2750392456054688, 0.27481195068359376, 0.27466339111328125, 0.27474432373046875, 0.2748590087890625, 0.2747484130859375, 0.27513446044921874, 0.2747361145019531, 0.27480267333984376, 0.27480267333984376, 0.27488461303710937, 0.2747166748046875, 0.27480780029296875, 0.27510067749023437, 0.274914306640625, 0.274840576171875, 0.27487744140625, 0.27528192138671875, 0.2748907470703125, 0.2748794860839844, 0.2748118896484375, 0.2750136413574219, 0.2750320739746094, 0.2749593505859375, 0.2749214782714844, 0.27498086547851563, 0.274977783203125, 0.27479449462890626, 0.2750894165039062, 0.27514776611328123, 0.27490304565429685, 0.2749347839355469, 0.27515911865234377, 0.27516000366210935, 0.2748241882324219, 0.27503616333007813, 0.2748866577148438, 0.27518875122070313, 0.27508837890625, 0.2749327392578125, 0.2751457214355469, 0.27504229736328123, 0.27509451293945314, 0.2749429626464844, 0.274988037109375, 0.27498291015625, 0.5922621459960937, 0.2748334045410156, 0.27478936767578127, 0.274671630859375, 0.274904052734375, 0.27486209106445314, 0.27468798828125, 0.27506689453125, 0.2748272705078125, 0.2748641357421875, 0.2748590087890625, 0.27523175048828125, 0.27472998046875, 0.2749521789550781, 0.2749706115722656, 0.2748108825683594, 0.27493887329101563, 0.2748170166015625, 0.2746921081542969, 0.27491326904296876, 0.27478936767578127, 0.27473919677734376, 0.27510885620117187, 0.2749378662109375, 0.2750115966796875, 0.27491943359375, 0.27507611083984373, 0.27492556762695314, 0.27495526123046876, 0.27483544921875, 0.27498086547851563, 0.2751180725097656, 0.27493991088867187, 0.27493170166015624, 0.2750433349609375, 0.2749429626464844, 0.2749429626464844, 0.27481396484375, 0.2750146484375, 0.27511602783203126, 0.27493374633789064, 0.27495834350585935, 0.2749859924316406, 0.2748938293457031, 0.27500238037109376, 0.2749153137207031, 0.27501773071289065, 0.27500851440429686, 0.27495834350585935, 0.2749931640625, 0.2754150390625, 0.27491021728515624, 0.27495526123046876, 0.27501773071289065, 0.27503103637695314, 0.27497369384765624, 0.2750279541015625, 0.27503717041015624, 0.2749358215332031, 0.2750136413574219, 0.2751170654296875, 0.2751037292480469, 0.27543038940429687, 0.5910579223632813, 0.2748610534667969, 0.27487335205078123, 0.27465625, 0.2748170166015625, 0.2747187194824219, 0.274766845703125, 0.27465933227539063, 0.27502386474609375, 0.2747832336425781, 0.27492044067382815, 0.2748713073730469, 0.2748293151855469, 0.27487640380859374, 0.2746726379394531, 0.27485797119140626, 0.27496652221679685, 0.2746132507324219, 0.2750822448730469, 0.2748190612792969, 0.27481497192382814, 0.2748252258300781, 0.274798583984375, 0.2749747314453125, 0.2747555847167969, 0.2747381896972656, 0.2749081726074219, 0.2749941711425781, 0.27478427124023436, 0.27480679321289064, 0.27512115478515625, 0.27513037109375, 0.27485491943359375, 0.27482623291015623, 0.2749849548339844, 0.27486309814453125, 0.27489791870117186, 0.2749706115722656, 0.27500543212890627, 0.2748241882324219, 0.27498904418945314, 0.27505767822265625, 0.274977783203125, 0.274977783203125, 0.2749982604980469, 0.2750975952148437, 0.27501568603515625, 0.27494195556640627, 0.27497882080078123, 0.27503616333007813, 0.27515802001953127, 0.27503616333007813, 0.2753914794921875, 0.27505050659179686, 0.27500338745117187, 0.2749440002441406, 0.2753894348144531, 0.274946044921875, 0.27488461303710937, 0.27529010009765625, 0.2749910888671875, 0.275162109375, 0.27531781005859374, 0.5928375854492187, 0.27484979248046876, 0.2748805236816406, 0.2747832336425781, 0.27487026977539064, 0.27483544921875, 0.27488153076171873, 0.27486004638671874, 0.27469720458984376, 0.27491021728515624, 0.27487136840820314, 0.2749767150878906, 0.27499520874023436, 0.2750126037597656, 0.27480575561523435, 0.2747801513671875, 0.2749941711425781, 0.2747760620117187, 0.2748794860839844, 0.2749521789550781, 0.2749286499023437, 0.27487026977539064, 0.2749327392578125, 0.2751047668457031, 0.2753587341308594, 0.2748385314941406, 0.274914306640625, 0.2748487548828125, 0.2753208312988281, 0.27483544921875, 0.27501055908203126, 0.2749910888671875, 0.27491326904296876, 0.2748713073730469, 0.2750392456054688, 0.2749214782714844, 0.27502590942382815, 0.27495526123046876, 0.2749573059082031, 0.27490304565429685, 0.27542733764648436, 0.27513446044921874, 0.2750699462890625, 0.27546929931640624, 0.2749900817871094, 0.2750146484375, 0.27492352294921873, 0.2749347839355469, 0.2750044250488281, 0.27507403564453126, 0.27504434204101563, 0.2749593505859375, 0.274988037109375, 0.2751805419921875, 0.2751549377441406, 0.27508837890625, 0.2750044250488281, 0.27502285766601564, 0.275135498046875, 0.2751109008789063, 0.2751518859863281, 0.2752655334472656, 0.2749982604980469, 0.5929502563476563, 0.2749910888671875, 0.2748016662597656, 0.2748037109375, 0.27479757690429685, 0.2747115478515625, 0.274872314453125, 0.2748487548828125, 0.2747688903808594, 0.27480267333984376, 0.27472998046875, 0.2750965881347656, 0.2751508483886719, 0.27519180297851564, 0.27491839599609375, 0.27517236328125, 0.274777099609375, 0.27474533081054686, 0.27501055908203126, 0.274914306640625, 0.27491738891601564, 0.2747781066894531, 0.27482623291015623, 0.27508633422851564, 0.27482827758789063, 0.2750146484375, 0.2749716491699219, 0.27486209106445314, 0.275083251953125, 0.2749542541503906, 0.27511911010742185, 0.2748805236816406, 0.274946044921875, 0.2749747314453125, 0.2749982604980469, 0.27494091796875, 0.2749368286132812, 0.2750382080078125, 0.2749624328613281, 0.27491738891601564, 0.27494195556640627, 0.2748528747558594, 0.27500543212890627, 0.27509246826171874, 0.27494091796875, 0.27498291015625, 0.27497268676757813, 0.27491943359375, 0.2749921264648437, 0.2751109008789063, 0.27503103637695314, 0.27493170166015624, 0.2751805419921875, 0.27501669311523436, 0.275051513671875, 0.27502590942382815, 0.2750494689941406, 0.27510580444335936, 0.27505050659179686, 0.2749716491699219, 0.2749798278808594, 0.2751068115234375, 0.27494503784179686]",tokens/s,3.578290079338431,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -2749,48 +2743,33 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 86.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -2861,39 +2840,33 @@ OSError: 1 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained - model = cls(config, *model_args, **model_kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm-20b/80729bcf52fbc4553d965926b27304ac5e156d98/modeling_internlm.py"", line 906, in __init__ - self.model = InternLMModel(config) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm-20b/80729bcf52fbc4553d965926b27304ac5e156d98/modeling_internlm.py"", line 729, in __init__ - self.layers = nn.ModuleList([InternLMDecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm-20b/80729bcf52fbc4553d965926b27304ac5e156d98/modeling_internlm.py"", line 729, in - self.layers = nn.ModuleList([InternLMDecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm-20b/80729bcf52fbc4553d965926b27304ac5e156d98/modeling_internlm.py"", line 545, in __init__ - self.self_attn = INTERNLM_ATTENTION_CLASSES[config.attn_implementation](config=config) -KeyError: 'sdpa' + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,internlm,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,,MB,3095.433216,13793.492992,0.0,13147.045888,12781.924864,s,10,8.988890075683594,0.8988890075683594,0.001423213612699,0.8981753845214844,0.9012999267578126,0.9014743347167968,0.9016138610839843,"[0.8981488037109375, 0.8976058959960938, 0.8978265991210937, 0.8977764282226562, 0.8990162353515625, 0.8982019653320312, 0.899675048828125, 0.9012611694335938, 0.9016487426757812, 0.8977291870117188]",tokens/s,284.79600689802794,kWh,1.0606916911072201e-05,5.812208237193772e-06,5.517099321083324e-05,7.159011835909921e-05,tokens/kWh,3575912.5123371445,MB,3096.809472,13793.492992,0.0,13147.045888,12906.706432,s,10,496.00179296875,49.600179296875,0.0025068761942268125,49.599537109375,49.60291796875,49.60425390625,49.60532265625,"[49.60262109375, 49.60017578125, 49.599390625, 49.5969921875, 49.60200390625, 49.59917578125, 49.59676953125, 49.5996796875, 49.60558984375, 49.59939453125]",tokens/s,1.2701566988885713,kWh,0.0005855437311530114,0.00032092896961866243,0.0030658153091317657,0.003972288009903439,tokens/kWh,15859.877189904828,,s,629,503.3055490112303,0.8001678044693649,0.10686695302898579,0.78727783203125,0.7875329833984375,0.7876315185546875,1.6862559765625,"[0.7872880859375, 0.7870259399414062, 0.7873024291992188, 0.787430419921875, 0.7872225341796875, 0.7870167236328125, 0.7872471313476562, 0.787135498046875, 0.7874212036132813, 0.787646484375, 0.7871190795898437, 0.787198974609375, 0.78721533203125, 0.7873003540039063, 0.7873402709960937, 0.7873802490234375, 0.7873699951171875, 0.7872757568359375, 0.7873751220703125, 0.7871262817382813, 0.7870422973632812, 0.78744677734375, 0.7872112426757812, 0.7874457397460938, 0.7872388916015625, 0.7871692504882812, 0.7872634887695312, 0.7872112426757812, 0.7875809326171875, 0.7871897583007812, 0.7871918334960938, 0.7872716674804687, 0.7872532348632812, 0.7870771484375, 0.787178466796875, 0.7871795043945312, 0.7873187866210938, 0.7872747802734374, 0.7873935546875, 0.7871959228515625, 0.787209228515625, 0.7876884765625, 0.7873280029296875, 0.7873853149414063, 0.7873720092773437, 0.787345458984375, 0.7874744262695312, 0.787357666015625, 0.7874508666992187, 0.7872317504882812, 0.78719384765625, 0.7873925170898437, 0.7873443603515625, 0.787577880859375, 0.7875277099609375, 0.7875594482421875, 0.7873863525390625, 0.7875327758789062, 0.7873659057617187, 0.7875850219726562, 0.7875235595703125, 0.787483642578125, 1.68622802734375, 0.7872767944335938, 0.7871539916992187, 0.7875859985351562, 0.7870955810546875, 0.787146728515625, 0.7872962646484375, 0.787230712890625, 0.7869522094726562, 0.7872296752929687, 0.7870853271484375, 0.7870474243164063, 0.7871734008789063, 0.7869942016601562, 0.787184814453125, 0.7870564575195312, 0.7873433837890625, 0.7872163696289063, 0.7871109008789062, 0.7871221923828124, 0.7873392944335937, 0.7873372192382813, 0.7872645263671875, 0.7872798461914062, 0.7871262817382813, 0.7875082397460937, 0.7871580200195313, 0.7872501831054688, 0.7874375610351563, 0.787314697265625, 0.7873116455078125, 0.7874232177734375, 0.78718359375, 0.7870259399414062, 0.7873351440429688, 0.7873904418945312, 0.78729931640625, 0.7872317504882812, 0.7870996704101563, 0.7874580688476562, 0.7870596923828125, 0.7873372192382813, 0.78745703125, 0.7872686157226563, 0.7873843383789062, 0.787409912109375, 0.7872471313476562, 0.7873372192382813, 0.787252197265625, 0.787504150390625, 0.7873597412109375, 0.787330078125, 0.7872645263671875, 0.7876270141601562, 0.78727783203125, 0.7873699951171875, 0.78739453125, 0.7873607788085938, 0.7875860595703125, 0.7873536376953125, 0.78723681640625, 0.7874774780273438, 0.7876034545898437, 1.686560791015625, 0.7868671875, 0.7871846313476563, 0.7875133666992188, 0.78708837890625, 0.78710986328125, 0.7871549682617187, 0.7870218505859375, 0.7871795043945312, 0.7870945434570312, 0.787240966796875, 0.7870156860351563, 0.7869183959960937, 0.7871447143554687, 0.78708837890625, 0.7876351928710937, 0.78715185546875, 0.7872379150390625, 0.7870535888671875, 0.7870084838867187, 0.7871365356445312, 0.78715185546875, 0.7871734619140625, 0.7874733276367187, 0.7874068603515625, 0.7871559448242188, 0.787051513671875, 0.7870802001953126, 0.7871375122070312, 0.7869869995117188, 0.7874959106445313, 0.7874119873046875, 0.7872788696289063, 0.7873699951171875, 0.7870177001953125, 0.7873443603515625, 0.7871549682617187, 0.7873157348632812, 0.7871876831054687, 0.78721435546875, 0.7873106079101563, 0.7871416015625, 0.787336181640625, 0.7874345092773437, 0.7872593994140625, 0.7873280029296875, 0.7874805908203125, 0.787968017578125, 0.7871365356445312, 0.7875297241210938, 0.7873392944335937, 0.7876546630859375, 0.7874345092773437, 0.7874006958007812, 0.787794921875, 0.7873331298828125, 0.7873351440429688, 0.7873392944335937, 0.7871876831054687, 0.7873443603515625, 0.78742529296875, 0.7876710205078125, 0.7873526000976563, 1.6864215087890626, 0.787156982421875, 0.7871692504882812, 0.7870320434570313, 0.7871047973632812, 0.78719384765625, 0.787282958984375, 0.7871682739257813, 0.7870494995117188, 0.7871918334960938, 0.7872501831054688, 0.7871016845703125, 0.7874375610351563, 0.7870842895507812, 0.7874447631835938, 0.7872757568359375, 0.7871201171875, 0.787314697265625, 0.78710888671875, 0.787078125, 0.7875952758789062, 0.7874120483398438, 0.7871282348632812, 0.7870576782226563, 0.7873341674804688, 0.7872553100585937, 0.7871815795898438, 0.7870156860351563, 0.7871897583007812, 0.7872532348632812, 0.7870975952148438, 0.7872348022460938, 0.7872747802734374, 0.7872174072265625, 0.7872041015625, 0.7873034057617188, 0.7870156860351563, 0.787197021484375, 0.7872992553710938, 0.7872368774414062, 0.787378173828125, 0.7874457397460938, 0.787209228515625, 0.7874713745117188, 0.7871426391601563, 0.7871395874023438, 0.7870279541015625, 0.7873341674804688, 0.7872757568359375, 0.7874959106445313, 0.7871610717773437, 0.7873177490234375, 0.7875563354492188, 0.7873013916015625, 0.787205078125, 0.7873659057617187, 0.7874191284179688, 0.7873065185546875, 0.7872808837890625, 0.7878389892578125, 0.7872348022460938, 0.78746826171875, 0.787251220703125, 1.686266845703125, 0.787282958984375, 0.7871907958984375, 0.7871898193359375, 0.7868651123046875, 0.787078125, 0.7870648193359375, 0.7873024291992188, 0.7872276611328125, 0.78713037109375, 0.7872348022460938, 0.7871282958984375, 0.7874324340820312, 0.7871651611328125, 0.7873106079101563, 0.787651611328125, 0.7873771362304688, 0.7877498779296875, 0.7874212036132813, 0.7871027221679687, 0.7872348022460938, 0.7873812255859375, 0.7873751220703125, 0.7872491455078126, 0.787462158203125, 0.7871181030273438, 0.7874293823242188, 0.7872245483398438, 0.7871190795898437, 0.7873648681640625, 0.7871498413085938, 0.7873106079101563, 0.78740478515625, 0.787589111328125, 0.787209228515625, 0.7874078979492187, 0.7869603881835937, 0.787103759765625, 0.7870873413085937, 0.7873751220703125, 0.7874447631835938, 0.787420166015625, 0.787315673828125, 0.7874508666992187, 0.787472412109375, 0.7872890625, 0.7874396362304688, 0.787357666015625, 0.787262451171875, 0.7875082397460937, 0.7873976440429687, 0.7874017333984376, 0.7876085815429688, 0.787541015625, 0.7874969482421875, 0.787430419921875, 0.7876658935546875, 0.7873812255859375, 0.7873597412109375, 0.7874191284179688, 0.7873771362304688, 0.7876802368164062, 0.7875235595703125, 1.6858050537109375, 0.7872235717773437, 0.7872061157226562, 0.7873433837890625, 0.7870330810546875, 0.7872696533203125, 0.787198974609375, 0.787314697265625, 0.7871016845703125, 0.7874959106445313, 0.7869450073242188, 0.787140625, 0.7872020263671875, 0.7870494995117188, 0.7871918334960938, 0.787009521484375, 0.7873853149414063, 0.7873126220703125, 0.7870853271484375, 0.7873423461914062, 0.7871283569335937, 0.7871856079101562, 0.787072021484375, 0.7873239135742187, 0.7870382080078125, 0.7873341674804688, 0.7877058715820312, 0.7871836547851563, 0.7872327270507813, 0.787140625, 0.7872317504882812, 0.78727783203125, 0.7871293334960937, 0.7871600341796875, 0.7873535766601563, 0.7871764526367188, 0.7871395874023438, 0.7871068115234375, 0.7872532348632812, 0.787435546875, 0.7875829467773438, 0.7872645263671875, 0.7874283447265625, 0.7873167114257813, 0.7873587036132812, 0.7876290283203125, 0.7873638305664062, 0.787304443359375, 0.7874037475585938, 0.787282958984375, 0.7874293823242188, 0.7872808837890625, 0.7875972900390625, 0.7874498291015625, 0.7873607788085938, 0.7873402709960937, 0.7875440673828125, 0.7874150390625, 0.7875194702148437, 0.7875552978515625, 0.7872030639648437, 0.7873720092773437, 0.7872184448242188, 1.6870113525390624, 0.7871713256835937, 0.7869183959960937, 0.7871068115234375, 0.7869962158203125, 0.7871795043945312, 0.7872501831054688, 0.7877222290039062, 0.7870320434570313, 0.78708837890625, 0.7871897583007812, 0.7871375122070312, 0.7872225341796875, 0.7870443725585937, 0.7869265747070312, 0.7872819213867187, 0.7871897583007812, 0.7873556518554687, 0.78718359375, 0.7872952270507813, 0.787251220703125, 0.7871815795898438, 0.7872481079101562, 0.7872788696289063, 0.7871242065429688, 0.7873423461914062, 0.7870812377929688, 0.7873597412109375, 0.7870422973632812, 0.7870965576171876, 0.7872737426757812, 0.7871129760742187, 0.7872225341796875, 0.78717236328125, 0.7872665405273438, 0.787083251953125, 0.7871498413085938, 0.7871826171875, 0.7873556518554687, 0.7874365234375, 0.7871846313476563, 0.7872051391601562, 0.787178466796875, 0.7873925170898437, 0.7873955688476563, 0.78745703125, 0.787177490234375, 0.7872665405273438, 0.7871211547851562, 0.78724609375, 0.7872604370117188, 0.78769970703125, 0.7871692504882812, 0.787398681640625, 0.7871006469726562, 0.7874293823242188, 0.787251220703125, 0.7878092651367188, 0.78738330078125, 0.78727783203125, 0.7872849731445313, 0.787525634765625, 0.7875266723632812, 1.688580078125, 0.7875010375976562, 0.78710888671875, 0.787219482421875, 0.78703515625, 0.7871795043945312, 0.7870013427734375, 0.787435546875, 0.7873505249023437, 0.7871016845703125, 0.78717236328125, 0.7871641845703125, 0.7875369262695312, 0.7873310546875, 0.78725732421875, 0.7872532348632812, 0.7872379150390625, 0.7872910766601563, 0.7870904541015625, 0.7873751220703125, 0.7871580810546875, 0.7873095092773438, 0.787483642578125, 0.7870177001953125, 0.7870822143554688, 0.7873853149414063, 0.787145751953125, 0.7872327880859376, 0.78710888671875, 0.7874580688476562, 0.7870371704101562, 0.7873546142578125, 0.7873228759765625, 0.787219482421875, 0.7871498413085938, 0.7874293823242188, 0.7872030639648437, 0.7873638305664062, 0.7872952270507813, 0.7876177978515625, 0.787346435546875, 0.78724609375, 0.787198974609375, 0.7873720092773437, 0.7875277099609375, 0.7873535766601563, 0.7872420043945313, 0.7872880859375, 0.7872399291992187, 0.7871764526367188, 0.787567626953125, 0.787324951171875, 0.7877447509765625, 0.78729931640625, 0.7872399291992187, 0.7876454467773437, 0.787399658203125, 0.7873812255859375, 0.787399658203125, 0.7873167114257813, 0.7872716674804687, 0.7875164184570312, 0.7873894653320312, 1.6889835205078125, 0.78731982421875, 0.7872686157226563, 0.78731982421875, 0.7867996215820312, 0.7874959106445313, 0.7870904541015625, 0.78729931640625, 0.7871846313476563, 0.78740478515625, 0.7873280029296875, 0.7871190795898437, 0.787230712890625, 0.7871426391601563, 0.7872501831054688, 0.7871344604492188, 0.787125244140625, 0.7872696533203125, 0.7874662475585937, 0.78742626953125, 0.787325927734375, 0.7875921630859375, 0.7871027221679687, 0.787114990234375, 0.7877273559570312, 0.7872798461914062, 0.7875399780273438, 0.7872379150390625, 0.7872163696289063, 0.7874652099609375, 0.7871846313476563, 0.787125244140625, 0.7873515625, 0.7874293823242188, 0.7873546142578125, 0.7871641845703125, 0.7871826171875, 0.787314697265625, 0.7872696533203125, 0.7872747802734374, 0.78727783203125, 0.7873239135742187, 0.78721533203125, 0.787272705078125, 0.7872870483398438, 0.7874498291015625, 0.7874037475585938, 0.7873505249023437, 0.7871979370117187, 0.7875717163085938, 0.7872645263671875, 0.7875338134765625, 0.7875123291015625, 0.7872870483398438, 0.787283935546875, 0.7875277099609375, 0.7874907836914062, 0.7879116821289063, 0.7873699951171875, 0.7874068603515625, 0.7874037475585938, 0.7875430297851562, 0.7881441040039062, 1.688532958984375, 0.7873351440429688, 0.78715185546875, 0.7871795043945312, 0.7868784790039063, 0.7871682739257813, 0.7871661987304688, 0.787072998046875, 0.7870259399414062, 0.7871641845703125, 0.7871897583007812, 0.7870986328125, 0.78719384765625, 0.7872757568359375, 0.7872184448242188, 0.7872747802734374, 0.787252197265625, 0.7876331787109375, 0.7872133178710937, 0.7871109008789062, 0.7872471313476562, 0.7872225341796875, 0.7870023803710937, 0.7875215454101563, 0.7872635498046875, 0.7872357788085937, 0.7873863525390625, 0.7871764526367188, 0.7871734008789063, 0.7871651611328125, 0.7871242065429688, 0.787431396484375, 0.7871795043945312, 0.7872102661132813, 0.787019775390625, 0.7872379150390625, 0.7876024169921875, 0.7872655639648437, 0.7871641845703125, 0.7872716674804687, 0.7873771362304688, 0.787409912109375, 0.7874385986328125, 0.7872849731445313, 0.7871764526367188, 0.7874283447265625, 0.7873208618164063, 0.7872880859375, 0.78717236328125, 0.7874662475585937, 0.7873935546875, 0.7875758056640625, 0.78751025390625, 0.7873925170898437, 0.7873597412109375, 0.7873290405273438, 0.7875635375976563, 0.7874447631835938, 0.7875901489257813, 0.7875963134765624, 0.7870873413085937, 0.7879024658203125, 0.7874406127929687]",tokens/s,1.2497378605018417,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-30b,facebook/opt-30b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -2949,48 +2922,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2_moe,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,mixtral,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,MB,1336.54528,1292.36992,0.0,645.922816,583.517696,s,10,0.3397931861877441,0.033979318618774415,0.001934910071424721,0.032971408843994146,0.03566902961730957,0.03745379467010498,0.03888160671234131,"[0.03923855972290039, 0.032909088134765625, 0.03527241516113281, 0.0328919677734375, 0.03299142456054688, 0.03283273696899414, 0.03469622421264648, 0.03295139312744141, 0.03290889739990235, 0.033100479125976565]",tokens/s,7533.99451213697,kWh,3.979384447589066e-07,2.1805127691752405e-07,1.234731627515188e-06,1.8507213491916185e-06,tokens/kWh,138324443.12149903,MB,1336.54528,1292.36992,0.0,645.922816,617.915392,s,10,21.585919433593748,2.158591943359375,0.032699156970019405,2.151986328125,2.2039208740234377,2.2094919311523435,2.2139487768554686,"[2.21506298828125, 2.185614990234375, 2.202682861328125, 2.14994970703125, 2.1486455078125, 2.100208251953125, 2.15894091796875, 2.15402294921875, 2.126743896484375, 2.14404736328125]",tokens/s,29.185692179483592,kWh,2.5160195789854937e-05,1.378844924343664e-05,7.117124296388444e-05,0.00011011988799717602,tokens/kWh,572103.7420744164,,s,629,21.848751106262192,0.03473569333269031,0.003915216842516604,0.03429580688476563,0.03490857086181641,0.03530035171508789,0.06544596710205079,"[0.03697049713134765, 0.036383743286132815, 0.03609395217895508, 0.036050945281982424, 0.03664492797851562, 0.03628230285644531, 0.03609292984008789, 0.035920894622802735, 0.03645132827758789, 0.03612364959716797, 0.036528129577636716, 0.03566899108886719, 0.03584204864501953, 0.035465217590332034, 0.03479449462890625, 0.03489279937744141, 0.035095550537109374, 0.034969600677490234, 0.03465318298339844, 0.034769920349121096, 0.03466649627685547, 0.03486105728149414, 0.03489791870117188, 0.0348671989440918, 0.0348221435546875, 0.03500032043457031, 0.03470848083496094, 0.0350115852355957, 0.03482419204711914, 0.03486207962036133, 0.03474534225463867, 0.03450265502929688, 0.03512627029418945, 0.03491020965576172, 0.03476172637939453, 0.034699264526367186, 0.03464908981323242, 0.03466239929199219, 0.03463987350463867, 0.03472895812988281, 0.03482419204711914, 0.03482931137084961, 0.03481190490722656, 0.03476275253295898, 0.03470336151123047, 0.034677761077880856, 0.034764801025390625, 0.03464089584350586, 0.03486617660522461, 0.03481190490722656, 0.034887680053710936, 0.03473100662231445, 0.034721790313720705, 0.03487744140625, 0.034857982635498046, 0.03471769714355469, 0.0349224967956543, 0.034756607055664065, 0.03473612976074219, 0.03489996719360351, 0.03482009506225586, 0.03475251388549805, 0.06539878082275391, 0.03443711853027344, 0.0344453125, 0.03467264175415039, 0.034819072723388675, 0.03465523147583008, 0.03469619369506836, 0.03466239929199219, 0.03475763320922851, 0.03470131301879883, 0.03472076797485352, 0.03474739074707031, 0.03471155166625976, 0.034648063659667966, 0.03486207962036133, 0.03484979248046875, 0.03461324691772461, 0.03467468643188477, 0.03463987350463867, 0.034816001892089846, 0.03467161560058594, 0.034718719482421875, 0.03478015899658203, 0.034561023712158204, 0.03474431991577148, 0.03470848083496094, 0.03480678558349609, 0.034680831909179685, 0.034576385498046876, 0.03468288040161133, 0.03468185424804687, 0.034614273071289066, 0.03450982284545898, 0.03474227142333984, 0.03460812759399414, 0.03462656021118164, 0.03469311904907227, 0.03471257781982422, 0.03472895812988281, 0.03481292724609375, 0.03467161560058594, 0.03452620697021484, 0.034835456848144535, 0.034985984802246094, 0.03463679885864258, 0.0347883529663086, 0.03482931137084961, 0.03485696029663086, 0.0349194221496582, 0.034915359497070315, 0.03500336074829102, 0.03472588729858399, 0.03476684951782227, 0.034735103607177735, 0.034678783416748044, 0.03481702423095703, 0.03480678558349609, 0.034797569274902344, 0.03487334442138672, 0.03478732681274414, 0.03483852767944336, 0.03467673492431641, 0.03490816116333008, 0.07052390289306641, 0.03481190490722656, 0.0346163215637207, 0.03483443069458008, 0.03470745468139649, 0.035148799896240236, 0.034685951232910156, 0.03495731353759766, 0.03478937530517578, 0.03479040145874023, 0.03482726287841797, 0.03472076797485352, 0.03478732681274414, 0.03483955383300781, 0.03490611267089844, 0.03496857452392578, 0.034857982635498046, 0.0347586555480957, 0.03474534225463867, 0.03484262466430664, 0.0348037109375, 0.03506687927246094, 0.03475763320922851, 0.036310016632080076, 0.03513651275634765, 0.0348671989440918, 0.03481292724609375, 0.03470336151123047, 0.03473920059204102, 0.03484672164916992, 0.03568230438232422, 0.035320831298828126, 0.034955265045166016, 0.03488358306884766, 0.03475046539306641, 0.035471359252929685, 0.035124225616455076, 0.034841598510742186, 0.034840576171875, 0.03480473709106445, 0.034800640106201174, 0.034871295928955076, 0.03530854415893555, 0.03491839981079101, 0.034948097229003904, 0.03496550369262695, 0.03504435348510742, 0.03487641525268555, 0.03488665771484375, 0.03473305511474609, 0.03489382553100586, 0.03489484786987305, 0.034705406188964845, 0.03483340835571289, 0.03720908737182617, 0.03559936141967773, 0.03501772689819336, 0.03483852767944336, 0.03479449462890625, 0.03467059326171875, 0.034700286865234374, 0.034802688598632815, 0.03480883026123047, 0.06551551818847656, 0.03234918212890625, 0.03228057479858398, 0.0322242546081543, 0.03220787048339844, 0.032233470916748046, 0.032456703186035156, 0.032304126739501955, 0.032252960205078125, 0.032616416931152345, 0.03248025512695313, 0.03226828765869141, 0.032295936584472655, 0.03215769577026367, 0.0321607666015625, 0.03411763381958008, 0.03482624053955078, 0.03470438385009766, 0.0346429443359375, 0.03479040145874023, 0.0345682258605957, 0.03465008163452148, 0.03458047866821289, 0.03469823837280273, 0.03468803024291992, 0.034671585083007814, 0.034732032775878906, 0.03471462249755859, 0.034601982116699216, 0.034669567108154296, 0.03480166244506836, 0.03470950317382813, 0.03479449462890625, 0.034857982635498046, 0.03468492889404297, 0.03466342544555664, 0.03495935821533203, 0.03487948989868164, 0.0349378547668457, 0.03482624053955078, 0.03474227142333984, 0.034748416900634765, 0.03469209671020508, 0.03465932846069336, 0.03486412811279297, 0.03466547012329101, 0.03473920059204102, 0.03483443069458008, 0.034726913452148435, 0.03474534225463867, 0.03469004821777344, 0.03482828903198242, 0.03479449462890625, 0.03474227142333984, 0.03463679885864258, 0.03486822509765625, 0.03467264175415039, 0.03469823837280273, 0.034855934143066404, 0.034249729156494144, 0.034233345031738284, 0.03444224166870117, 0.03420774459838867, 0.06550527954101562, 0.03235737609863281, 0.03217715072631836, 0.03259392166137695, 0.03417804718017578, 0.034097152709960936, 0.034195457458496094, 0.03415244674682617, 0.03408588790893555, 0.0343633918762207, 0.03422208023071289, 0.03404083251953125, 0.0341319694519043, 0.034049022674560545, 0.0341104621887207, 0.03403059387207031, 0.03411763381958008, 0.03412582397460937, 0.03405619049072266, 0.03418521499633789, 0.03429683303833008, 0.034323455810546875, 0.034119678497314454, 0.03416985702514649, 0.03429171371459961, 0.03431423950195313, 0.0343296012878418, 0.0341473274230957, 0.03392716979980469, 0.0340684814453125, 0.034277374267578126, 0.034459648132324217, 0.03410636901855469, 0.0342927360534668, 0.03419340896606445, 0.03426201629638672, 0.03416883087158203, 0.03408281707763672, 0.03406438446044922, 0.034609153747558595, 0.034181121826171876, 0.03412684631347656, 0.03419033432006836, 0.03419852828979492, 0.03528806304931641, 0.03522867202758789, 0.034601982116699216, 0.03428966522216797, 0.03416064071655273, 0.034098175048828124, 0.03407155227661133, 0.03415552139282227, 0.03415961456298828, 0.03417804718017578, 0.03420876693725586, 0.034206718444824216, 0.034146305084228515, 0.03426713562011719, 0.0336363525390625, 0.034272254943847655, 0.03419443130493164, 0.03422208023071289, 0.034298881530761716, 0.06534349060058593, 0.03220479965209961, 0.03225600051879883, 0.03226521682739258, 0.03239321517944336, 0.03222220611572266, 0.032230400085449216, 0.03233894348144531, 0.03224371337890625, 0.03223244857788086, 0.032254974365234376, 0.03231846237182617, 0.03247923278808594, 0.032366592407226565, 0.03223961639404297, 0.03237171173095703, 0.03228876876831055, 0.03222118377685547, 0.03220889663696289, 0.03232767868041992, 0.032510974884033206, 0.03239116668701172, 0.03240857696533203, 0.03242905426025391, 0.032266239166259765, 0.032274463653564456, 0.03230819320678711, 0.03278335952758789, 0.03282124710083008, 0.03420876693725586, 0.03411763381958008, 0.03430809783935547, 0.034135040283203126, 0.03419033432006836, 0.03405721664428711, 0.034111488342285154, 0.03420159912109375, 0.03409203338623047, 0.03424051284790039, 0.0341923828125, 0.034229248046875, 0.0341401596069336, 0.03411763381958008, 0.03416371154785156, 0.03425791931152344, 0.03424358367919922, 0.034154495239257815, 0.034111488342285154, 0.03416883087158203, 0.03417292785644531, 0.03435520172119141, 0.03417804718017578, 0.03422208023071289, 0.03403059387207031, 0.03409203338623047, 0.034269184112548826, 0.03425996780395508, 0.03407974243164062, 0.03420979309082031, 0.034148353576660156, 0.034111488342285154, 0.03418214416503906, 0.0343296012878418, 0.06927974700927735, 0.034157569885253904, 0.03397119903564453, 0.03403673553466797, 0.034080768585205076, 0.034154495239257815, 0.03405516815185547, 0.03415859222412109, 0.03419647979736328, 0.03408998489379883, 0.03414220809936523, 0.034141185760498044, 0.034083839416503905, 0.033983486175537106, 0.03414220809936523, 0.034255870819091795, 0.034239486694335936, 0.03566796875, 0.0341923828125, 0.034138111114501955, 0.03414425659179687, 0.034354209899902344, 0.03432342529296875, 0.034239486694335936, 0.03420159912109375, 0.034202625274658206, 0.03419852828979492, 0.03434700775146484, 0.03419647979736328, 0.03412377548217774, 0.03443302536010742, 0.03422412872314453, 0.03428147125244141, 0.03423436737060547, 0.034070526123046875, 0.034184192657470705, 0.03410124969482422, 0.03435007858276367, 0.034184192657470705, 0.03421491241455078, 0.034121726989746096, 0.03418521499633789, 0.03455487823486328, 0.034127872467041014, 0.03487027359008789, 0.034462718963623046, 0.03412275314331055, 0.034272254943847655, 0.03421388626098633, 0.0342108154296875, 0.03415039825439453, 0.03427840042114258, 0.034323455810546875, 0.034253822326660154, 0.03424358367919922, 0.03418214416503906, 0.03430912017822266, 0.03426921463012695, 0.03443606567382813, 0.03424563217163086, 0.03502899169921875, 0.03460300827026367, 0.03441561508178711, 0.06546431732177735, 0.03232972717285156, 0.03248537445068359, 0.03227443313598633, 0.03256012725830078, 0.036160511016845705, 0.03467571258544922, 0.034255870819091795, 0.034164737701416016, 0.03431935882568359, 0.034587646484375, 0.03426611328125, 0.034237438201904294, 0.034298881530761716, 0.034149375915527344, 0.03421388626098633, 0.034784255981445314, 0.0343111686706543, 0.03438489532470703, 0.034552833557128904, 0.034288639068603514, 0.03420774459838867, 0.03413913726806641, 0.03439308929443359, 0.03422822570800781, 0.0342108154296875, 0.034097152709960936, 0.034334720611572264, 0.03488153457641602, 0.035192832946777344, 0.034315265655517575, 0.034336769104003906, 0.03435724639892578, 0.034326526641845705, 0.03412377548217774, 0.03410124969482422, 0.03414425659179687, 0.034124801635742184, 0.03409718322753906, 0.034401248931884766, 0.03448934555053711, 0.03424665451049805, 0.03424358367919922, 0.03410227203369141, 0.0342999038696289, 0.03432550430297852, 0.034288639068603514, 0.034274303436279296, 0.034100223541259765, 0.034260990142822266, 0.03431423950195313, 0.03435520172119141, 0.03426406478881836, 0.03450163269042969, 0.03422412872314453, 0.03429580688476563, 0.03424256134033203, 0.03424870300292969, 0.03429580688476563, 0.03424051284790039, 0.03419750213623047, 0.034272254943847655, 0.034162689208984375, 0.06556877136230468, 0.03236454391479492, 0.03220991897583008, 0.03225190353393555, 0.03250380706787109, 0.0322979850769043, 0.03236044692993164, 0.03242905426025391, 0.032307201385498044, 0.0321976318359375, 0.03225804901123047, 0.03367935943603516, 0.034154495239257815, 0.03419340896606445, 0.034170879364013675, 0.0341923828125, 0.03403878402709961, 0.034118656158447266, 0.03405414581298828, 0.03410124969482422, 0.034195457458496094, 0.03435417556762695, 0.03437363052368164, 0.034181121826171876, 0.03409100723266602, 0.03419750213623047, 0.03425075149536133, 0.03439308929443359, 0.03425484848022461, 0.03426611328125, 0.034012161254882815, 0.03453849411010742, 0.03414220809936523, 0.0342845458984375, 0.03453849411010742, 0.034440193176269535, 0.03425894546508789, 0.0341288948059082, 0.03415244674682617, 0.034249729156494144, 0.034405376434326174, 0.03426816177368164, 0.03425996780395508, 0.034219009399414066, 0.03417292785644531, 0.03422208023071289, 0.034216960906982424, 0.034342910766601564, 0.034377727508544925, 0.03428044891357422, 0.03420774459838867, 0.034187263488769534, 0.034367488861083983, 0.034165760040283204, 0.034167808532714845, 0.034157569885253904, 0.03420163345336914, 0.03417187118530273, 0.033160190582275394, 0.03238092803955078, 0.03235430526733398, 0.03234201431274414, 0.03231948852539063, 0.06578995513916015, 0.03231948852539063, 0.032159744262695314, 0.032145408630371096, 0.03244441604614258, 0.03218431854248047, 0.03354214477539062, 0.034202625274658206, 0.0341923828125, 0.03407257461547852, 0.0342476806640625, 0.034253822326660154, 0.034146305084228515, 0.03407462310791016, 0.034118656158447266, 0.03422515106201172, 0.03419443130493164, 0.03403878402709961, 0.034045951843261715, 0.03427328109741211, 0.03504947280883789, 0.034955265045166016, 0.03431731033325195, 0.03428147125244141, 0.03412070465087891, 0.0342108154296875, 0.03422515106201172, 0.034105342864990236, 0.03411455917358398, 0.03442585754394531, 0.034113536834716796, 0.034236415863037106, 0.03417292785644531, 0.034167808532714845, 0.03415859222412109, 0.03418931198120117, 0.034301952362060545, 0.03418931198120117, 0.03436032104492188, 0.03430912017822266, 0.034167808532714845, 0.03411251068115234, 0.03413708877563477, 0.0343296012878418, 0.03417804718017578, 0.03430092620849609, 0.03407155227661133, 0.03410227203369141, 0.034157569885253904, 0.034116607666015625, 0.03416678237915039, 0.03409203338623047, 0.034239486694335936, 0.03420569610595703, 0.03434188842773438, 0.034149375915527344, 0.03424358367919922, 0.034151424407958986, 0.03425689697265625, 0.03420774459838867, 0.034179073333740234, 0.034237438201904294, 0.034328575134277346]",tokens/s,28.788830855403834,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -3541,48 +3499,33 @@ ValueError: FalconForCausalLM does not support an attention implementation throu ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,MB,1274.847232,1017.643008,0.0,371.195904,315.552256,s,10,0.17563712120056152,0.01756371212005615,0.0005432831148220809,0.017329248428344726,0.01822780475616455,0.018613710689544677,0.01892243543624878,"[0.018999616622924806, 0.017280960083007814, 0.017199647903442382, 0.017273727416992186, 0.01746089553833008, 0.017245920181274414, 0.017359840393066407, 0.017298656463623045, 0.017375808715820312, 0.018142047882080077]",tokens/s,14575.506490320542,kWh,2.0559210333847194e-07,1.126542116357798e-07,4.899288943582825e-07,8.081752093325342e-07,tokens/kWh,316762995.25622475,MB,1274.847232,1017.643008,0.0,371.195904,315.554816,s,10,10.830288574218752,1.0830288574218752,0.018539693960631418,1.0757147216796876,1.1116017211914062,1.114869219970703,1.1174832189941406,"[1.07570361328125, 1.0686051025390626, 1.0688455810546875, 1.061297607421875, 1.0667755126953125, 1.075725830078125, 1.094371337890625, 1.1108756103515625, 1.08995166015625, 1.11813671875]",tokens/s,58.17019515986862,kWh,1.2846804902749363e-05,7.0396259675622064e-06,2.840001088423971e-05,4.8286441754551286e-05,tokens/kWh,1304714.0710893627,,s,629,10.9678919429779,0.017437030116022104,0.0021411286712125165,0.016991231918334963,0.01768878059387207,0.017790565872192382,0.034496674957275396,"[0.01860710334777832, 0.018717695236206054, 0.017124351501464845, 0.01680384063720703, 0.016792575836181642, 0.01681715202331543, 0.01680384063720703, 0.016913408279418944, 0.01680486488342285, 0.01680384063720703, 0.01679769515991211, 0.016866304397583007, 0.016916479110717773, 0.01737932777404785, 0.01800294494628906, 0.01755340766906738, 0.0168089599609375, 0.01680179214477539, 0.01683558464050293, 0.01681612777709961, 0.01742336082458496, 0.01781862449645996, 0.017550336837768556, 0.01725132751464844, 0.017077247619628907, 0.016944128036499022, 0.01679871940612793, 0.016742399215698242, 0.016852991104125976, 0.016850944519042968, 0.017503231048583985, 0.017689599990844726, 0.018131967544555663, 0.017141759872436522, 0.01681510353088379, 0.016827392578125, 0.01675775909423828, 0.01681715202331543, 0.016874496459960937, 0.016773120880126953, 0.016852991104125976, 0.016736255645751954, 0.016857088088989256, 0.016776191711425782, 0.016825344085693358, 0.01677824020385742, 0.016776191711425782, 0.016721920013427736, 0.01677516746520996, 0.01677004814147949, 0.016793600082397463, 0.01681817626953125, 0.01722060775756836, 0.01681305694580078, 0.01681817626953125, 0.016722944259643553, 0.0168407039642334, 0.0168407039642334, 0.017089536666870117, 0.016963584899902344, 0.016806943893432617, 0.017143775939941406, 0.034566143035888675, 0.0167956485748291, 0.016857088088989256, 0.01679871940612793, 0.016780288696289062, 0.01686016082763672, 0.01680076789855957, 0.01677926445007324, 0.01684787178039551, 0.01681612777709961, 0.01679871940612793, 0.01678950309753418, 0.016821247100830078, 0.01681612777709961, 0.01683865547180176, 0.016943103790283204, 0.016970752716064453, 0.016923648834228516, 0.016827392578125, 0.01684377670288086, 0.01683865547180176, 0.01682841682434082, 0.01678335952758789, 0.01679974365234375, 0.016858112335205077, 0.016772096633911132, 0.016925695419311524, 0.01684486389160156, 0.016763839721679687, 0.01679871940612793, 0.016750591278076172, 0.01697587203979492, 0.01685606384277344, 0.016828447341918944, 0.016663520812988282, 0.01679155158996582, 0.017949695587158202, 0.017526784896850587, 0.01736396789550781, 0.017327104568481445, 0.017329151153564454, 0.017271839141845702, 0.017315807342529296, 0.017354751586914064, 0.017337343215942384, 0.017311744689941407, 0.017309696197509765, 0.017297439575195313, 0.017380319595336913, 0.017270784378051757, 0.01696870422363281, 0.01684377670288086, 0.016852991104125976, 0.016885759353637696, 0.01680384063720703, 0.016925695419311524, 0.01678950309753418, 0.01683558464050293, 0.01681817626953125, 0.016743423461914063, 0.01678335952758789, 0.016958463668823243, 0.01683865547180176, 0.03440332794189453, 0.01681100845336914, 0.0168407039642334, 0.016842752456665038, 0.01679769515991211, 0.01683660888671875, 0.01682636833190918, 0.016821247100830078, 0.01680384063720703, 0.016781312942504883, 0.016878591537475587, 0.017201152801513672, 0.01739263916015625, 0.017334272384643554, 0.017307647705078123, 0.016936960220336913, 0.01698918342590332, 0.017287168502807617, 0.017297407150268555, 0.017319936752319336, 0.01720729637145996, 0.016792575836181642, 0.01679155158996582, 0.016869375228881836, 0.01682329559326172, 0.016876575469970703, 0.017285087585449218, 0.0169932804107666, 0.01679462432861328, 0.016881664276123046, 0.016785408020019533, 0.016825344085693358, 0.01679155158996582, 0.01680281639099121, 0.016743423461914063, 0.017043455123901367, 0.017625087738037108, 0.01767628860473633, 0.017468416213989257, 0.017336320877075196, 0.017184768676757813, 0.017017887115478515, 0.01681814384460449, 0.01722265625, 0.016845823287963867, 0.01677414321899414, 0.016776191711425782, 0.01681817626953125, 0.01681920051574707, 0.016917503356933594, 0.01682431983947754, 0.016764959335327147, 0.016802783966064452, 0.01682431983947754, 0.01677824020385742, 0.016935935974121095, 0.01680076789855957, 0.01677516746520996, 0.0167956485748291, 0.016914432525634765, 0.016858112335205077, 0.016871423721313478, 0.0168089599609375, 0.03443916702270508, 0.01684787178039551, 0.01681715202331543, 0.01681715202331543, 0.01681510353088379, 0.01680179214477539, 0.01680793571472168, 0.016739328384399413, 0.016772096633911132, 0.016776191711425782, 0.016777215957641603, 0.01682329559326172, 0.01679974365234375, 0.01681920051574707, 0.016857088088989256, 0.016777215957641603, 0.016827392578125, 0.016825344085693358, 0.016804927825927733, 0.016981952667236327, 0.01683865547180176, 0.016916479110717773, 0.01686016082763672, 0.016849920272827147, 0.01679155158996582, 0.016827392578125, 0.01678950309753418, 0.01681817626953125, 0.016781312942504883, 0.0168407039642334, 0.016705535888671876, 0.016703487396240235, 0.016707584381103514, 0.01676697540283203, 0.016788480758666992, 0.01679052734375, 0.016742399215698242, 0.016722944259643553, 0.01677926445007324, 0.016784383773803712, 0.01680384063720703, 0.016743423461914063, 0.016763904571533202, 0.016910335540771485, 0.016961536407470702, 0.01675468826293945, 0.016892927169799805, 0.016763904571533202, 0.01677107238769531, 0.0168222713470459, 0.016865280151367186, 0.016740352630615234, 0.01683558464050293, 0.01676803207397461, 0.016789472579956055, 0.0168407039642334, 0.01681715202331543, 0.017492992401123047, 0.017527807235717775, 0.017064960479736328, 0.01682431983947754, 0.016769023895263673, 0.017338367462158204, 0.034697216033935545, 0.016945152282714843, 0.01739263916015625, 0.01763430404663086, 0.017093631744384767, 0.016842752456665038, 0.016991231918334963, 0.017481760025024416, 0.017557472229003907, 0.01738751983642578, 0.017356800079345702, 0.01743974494934082, 0.01721343994140625, 0.016815168380737305, 0.016809919357299804, 0.016876544952392578, 0.017435647964477538, 0.017688575744628905, 0.016929792404174804, 0.016788480758666992, 0.01681510353088379, 0.016886783599853517, 0.017145856857299805, 0.016948223114013672, 0.01683251190185547, 0.016889856338500975, 0.016779327392578126, 0.01688057518005371, 0.016745471954345705, 0.01681100845336914, 0.01678643226623535, 0.016723968505859374, 0.016750591278076172, 0.016878591537475587, 0.016743423461914063, 0.016792575836181642, 0.016769023895263673, 0.01686425590515137, 0.016784383773803712, 0.016780288696289062, 0.01677004814147949, 0.01683967971801758, 0.016721920013427736, 0.016753664016723634, 0.016727039337158203, 0.016786464691162108, 0.016804832458496094, 0.016761856079101564, 0.016746496200561522, 0.01684787178039551, 0.01679155158996582, 0.016829439163208008, 0.01679974365234375, 0.016756736755371093, 0.017069055557250978, 0.0168724479675293, 0.016726015090942382, 0.01683865547180176, 0.016788480758666992, 0.01676288032531738, 0.016755712509155272, 0.016772096633911132, 0.016773120880126953, 0.034871295928955076, 0.017183744430541992, 0.01678950309753418, 0.016751615524291993, 0.01681100845336914, 0.01681510353088379, 0.016874496459960937, 0.016742464065551757, 0.016807872772216795, 0.016902143478393555, 0.01678335952758789, 0.017366016387939453, 0.01736396789550781, 0.01745408058166504, 0.01703731155395508, 0.016784383773803712, 0.01721958351135254, 0.017246208190917968, 0.017286144256591796, 0.017309696197509765, 0.017330175399780275, 0.017060863494873048, 0.01728102493286133, 0.016892927169799805, 0.01677107238769531, 0.017072128295898437, 0.016961536407470702, 0.01740595245361328, 0.01686835289001465, 0.016759807586669923, 0.01681510353088379, 0.01723187255859375, 0.01723187255859375, 0.01737932777404785, 0.01716633605957031, 0.016920576095581053, 0.01722060775756836, 0.01720217514038086, 0.016769023895263673, 0.016849920272827147, 0.016858112335205077, 0.016793600082397463, 0.016752639770507814, 0.01720319938659668, 0.017043455123901367, 0.017276927947998046, 0.017531904220581054, 0.017293312072753905, 0.017319936752319336, 0.01748684883117676, 0.01683660888671875, 0.01684787178039551, 0.016842752456665038, 0.017305599212646485, 0.017276927947998046, 0.01745305633544922, 0.017333248138427734, 0.017323007583618166, 0.017357824325561523, 0.01681817626953125, 0.016825344085693358, 0.01680179214477539, 0.016829439163208008, 0.03451903915405274, 0.016761856079101564, 0.016750591278076172, 0.01677824020385742, 0.01683865547180176, 0.016850944519042968, 0.01658470344543457, 0.016767999649047852, 0.01677004814147949, 0.016743423461914063, 0.01680076789855957, 0.016741376876831054, 0.016746496200561522, 0.016746496200561522, 0.016793600082397463, 0.016891904830932617, 0.016784383773803712, 0.01675468826293945, 0.016912384033203123, 0.017947647094726564, 0.017836032867431642, 0.017657855987548828, 0.017747968673706056, 0.017719295501708983, 0.017574911117553712, 0.01759231948852539, 0.017604608535766602, 0.0176312313079834, 0.017525760650634766, 0.01761075210571289, 0.0175994873046875, 0.0176312313079834, 0.017571840286254883, 0.017605632781982423, 0.017721343994140625, 0.01759846305847168, 0.01760358428955078, 0.017640447616577147, 0.017542144775390626, 0.017612800598144532, 0.017537023544311522, 0.01760767936706543, 0.017613824844360353, 0.017750015258789064, 0.01769267272949219, 0.01761894416809082, 0.017726463317871095, 0.01754627227783203, 0.017348575592041015, 0.017467391967773437, 0.017570816040039062, 0.01764761543273926, 0.01756159973144531, 0.017548288345336914, 0.017571840286254883, 0.017693695068359376, 0.01764454460144043, 0.01756876754760742, 0.017572864532470703, 0.017574911117553712, 0.01762816047668457, 0.017588224411010742, 0.017557504653930665, 0.03539763259887695, 0.01764556884765625, 0.017670143127441407, 0.01763737678527832, 0.017632255554199217, 0.01760051155090332, 0.017574911117553712, 0.017735679626464843, 0.01759231948852539, 0.01769267272949219, 0.01763532829284668, 0.017566719055175782, 0.017549312591552735, 0.01760767936706543, 0.017686527252197267, 0.01760972785949707, 0.017702911376953127, 0.017665023803710937, 0.01760767936706543, 0.01761484718322754, 0.017532928466796875, 0.017574911117553712, 0.0176680965423584, 0.017690624237060547, 0.017683456420898438, 0.017688575744628905, 0.01761075210571289, 0.017590272903442384, 0.017563648223876953, 0.017622016906738282, 0.017537023544311522, 0.01762713623046875, 0.01762099266052246, 0.01759231948852539, 0.017657855987548828, 0.017580032348632812, 0.017977344512939454, 0.01760256004333496, 0.017563648223876953, 0.01765478324890137, 0.01759334373474121, 0.017812480926513673, 0.017675264358520508, 0.01756979179382324, 0.017584127426147463, 0.01760767936706543, 0.01760665512084961, 0.01759436798095703, 0.017560575485229494, 0.017597440719604493, 0.01762303924560547, 0.01780838394165039, 0.017544191360473634, 0.017641471862792968, 0.017567743301391603, 0.01761075210571289, 0.017566719055175782, 0.017581056594848633, 0.01763942337036133, 0.017588224411010742, 0.017562623977661132, 0.017663999557495116, 0.017550336837768556, 0.03456409454345703, 0.016745471954345705, 0.01675775909423828, 0.016773151397705077, 0.016796640396118164, 0.01678335952758789, 0.016767999649047852, 0.01679974365234375, 0.01676697540283203, 0.016793600082397463, 0.016871423721313478, 0.01680179214477539, 0.01679871940612793, 0.016736255645751954, 0.0168222713470459, 0.01679155158996582, 0.016780288696289062, 0.01681203269958496, 0.016788480758666992, 0.016780288696289062, 0.01679974365234375, 0.016756736755371093, 0.01678643226623535, 0.01678950309753418, 0.016759807586669923, 0.016735231399536133, 0.017673215866088866, 0.01766912078857422, 0.01760870361328125, 0.017546239852905272, 0.01774284744262695, 0.017704959869384765, 0.01783296012878418, 0.01759539222717285, 0.017584127426147463, 0.01765376091003418, 0.017625087738037108, 0.01762918472290039, 0.017604608535766602, 0.01761996841430664, 0.01760153579711914, 0.01757900810241699, 0.01760665512084961, 0.017564672470092774, 0.017624063491821287, 0.01761587142944336, 0.017640447616577147, 0.0176363525390625, 0.017731584548950196, 0.017681407928466796, 0.01779097557067871, 0.01781350326538086, 0.01765478324890137, 0.017682432174682617, 0.01763020706176758, 0.017696767807006835, 0.01761894416809082, 0.017689599990844726, 0.017624063491821287, 0.017672191619873046, 0.0175861759185791, 0.017737728118896484, 0.0176312313079834, 0.03626393508911133, 0.017755136489868165, 0.019316736221313476, 0.017987583160400392, 0.017670143127441407, 0.017711103439331053, 0.017671167373657228, 0.017716224670410157, 0.01766092872619629, 0.017704959869384765, 0.017719295501708983, 0.01764761543273926, 0.017771520614624024, 0.017674240112304687, 0.017789951324462892, 0.017649663925170898, 0.017666048049926757, 0.017688575744628905, 0.01763839912414551, 0.017672191619873046, 0.017691648483276368, 0.017523712158203125, 0.017796127319335938, 0.01910780715942383, 0.018461696624755858, 0.017769472122192383, 0.017679359436035155, 0.017622016906738282, 0.018100223541259765, 0.01782476806640625, 0.017666048049926757, 0.017857568740844727, 0.01765679931640625, 0.0176629753112793, 0.017612800598144532, 0.01759231948852539, 0.017597440719604493, 0.01763020706176758, 0.017682432174682617, 0.01762918472290039, 0.01754521560668945, 0.017605632781982423, 0.017727487564086913, 0.01756876754760742, 0.017679359436035155, 0.01765068817138672, 0.017713151931762695, 0.01763430404663086, 0.017523712158203125, 0.017596416473388672, 0.017552383422851564, 0.017622016906738282, 0.017624063491821287, 0.017580032348632812, 0.01761996841430664, 0.017716224670410157, 0.017933311462402343, 0.017604608535766602, 0.017714176177978515, 0.017683456420898438, 0.01760358428955078, 0.017680383682250975, 0.01759846305847168]",tokens/s,57.34921562595369,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,,cuda,0,42,,,,,False,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,56387f0ed9acb6808de92d1e9f4c4b67cc3f2798,4.40.2,,0.30.0,,,,1.19.1,,,,0.10.0,,,MB,3991.42912,12218.53184,0.0,11572.084736,11067.6992,s,10,6.169018859863281,0.6169018859863281,0.0006143848204495043,0.6168476867675781,0.6177305908203126,0.6177555847167969,0.6177755798339845,"[0.6164817504882812, 0.6169854736328125, 0.6172991943359375, 0.6177250366210938, 0.6173790283203126, 0.6177805786132813, 0.6158299560546875, 0.6166307373046875, 0.6167098999023437, 0.6161972045898437]",tokens/s,414.97684772141145,kWh,7.279734644624921e-06,3.988978911205348e-06,3.927705756541092e-05,5.0545771121241186e-05,tokens/kWh,5064716.480157118,MB,3991.42912,12218.53184,0.0,11572.084736,11417.442816,s,10,337.63061328125,33.763061328125,0.00225096621642262,33.763652343749996,33.765573046875005,33.7657611328125,33.7659116015625,"[33.76119921875, 33.7635546875, 33.76400390625, 33.76501171875, 33.76553125, 33.76594921875, 33.76088671875, 33.76237109375, 33.76375, 33.75835546875]",tokens/s,1.8659445418096703,kWh,0.0003985123106837273,0.00021841917147055615,0.002184355276894591,0.0028012867590488746,tokens/kWh,22489.6647215762,,s,629,342.6465277709964,0.5447480568696281,0.0733593452296599,0.5358714599609375,0.5361645751953125,0.5362993408203125,1.153045400390625,"[0.5356687622070313, 0.535673828125, 0.5356615600585938, 0.5357639770507813, 0.5357711181640625, 0.5356759033203125, 0.535635986328125, 0.5359185791015625, 0.5357578125, 0.5356636352539063, 0.5356380004882813, 0.5357578125, 0.5359759521484375, 0.5356769409179688, 0.535857177734375, 0.5356062622070312, 0.53585302734375, 0.5361551513671875, 0.535804931640625, 0.5360128784179687, 0.5360413818359375, 0.5358387451171875, 0.535816162109375, 0.535973876953125, 0.535636962890625, 0.5359267578125, 0.5362216796875, 0.5361387329101562, 0.535815185546875, 0.5357230224609375, 0.5360496826171876, 0.5357578125, 0.53587353515625, 0.5360732421875, 0.5357168579101562, 0.5360292358398437, 0.5357178344726562, 0.5357557983398438, 0.5360343017578125, 0.535947265625, 0.5358602294921875, 0.535920654296875, 0.5357916259765625, 0.535920654296875, 0.5360834350585938, 0.5358305053710938, 0.53599951171875, 0.5358602294921875, 0.5360465698242187, 0.5363035888671875, 0.536195068359375, 0.5359042358398437, 0.5357721557617188, 0.5357567749023437, 0.5359011840820312, 0.5358878784179687, 0.5359196166992187, 0.536026123046875, 0.53610498046875, 0.5359933471679688, 0.5357721557617188, 0.5357793579101563, 1.1541483154296874, 0.5357281494140625, 0.5357219848632813, 0.5358991088867188, 0.535482421875, 0.5357772216796876, 0.5355274047851563, 0.535552001953125, 0.5355714721679687, 0.5356912841796875, 0.5359759521484375, 0.5358028564453124, 0.53585302734375, 0.5360025634765625, 0.535615478515625, 0.5357639770507813, 0.5356759033203125, 0.5356615600585938, 0.535815185546875, 0.5357291259765625, 0.5358991088867188, 0.5358521118164062, 0.5355857543945313, 0.5361275024414063, 0.5361817626953125, 0.5357117309570313, 0.535826416015625, 0.53583154296875, 0.536079345703125, 0.5360015258789063, 0.5357875366210938, 0.5358714599609375, 0.5357742309570312, 0.5356636352539063, 0.5358673706054687, 0.5356482543945312, 0.535783447265625, 0.5357783203125, 0.5358848266601562, 0.5359739379882813, 0.5360004272460938, 0.535836669921875, 0.536052734375, 0.5358960571289062, 0.5359277954101562, 0.5359247436523438, 0.5373921508789062, 0.5360332641601563, 0.5358356323242187, 0.53625341796875, 0.5360988159179687, 0.53592578125, 0.536173583984375, 0.5359595336914063, 0.5357752685546875, 0.5361581420898438, 0.536289306640625, 0.5364039916992187, 0.536131591796875, 0.5361654052734375, 0.5361387329101562, 0.5359298706054687, 0.5361428833007813, 1.1538216552734375, 0.5354352416992187, 0.5355479125976562, 0.5360906372070312, 0.5357619018554688, 0.5358991088867188, 0.5355653076171875, 0.5356021728515625, 0.5357342529296875, 0.5358622436523437, 0.5358602294921875, 0.5358090209960937, 0.5357588500976562, 0.5360332641601563, 0.535568359375, 0.5357342529296875, 0.535677978515625, 0.535667724609375, 0.5356574096679687, 0.5356902465820312, 0.5361408081054687, 0.5363302612304688, 0.5360311889648437, 0.5362800903320313, 0.5358776245117187, 0.5357772827148437, 0.535847900390625, 0.5357383422851563, 0.536015869140625, 0.535689208984375, 0.5360271606445313, 0.5359616088867187, 0.5356748657226562, 0.5358253784179687, 0.53612646484375, 0.5359862060546875, 0.5360189208984375, 0.535930908203125, 0.5358038940429688, 0.5359820556640625, 0.5359185791015625, 0.5363363647460937, 0.5360036010742187, 0.5360660400390626, 0.5358038940429688, 0.5359585571289063, 0.5359534301757812, 0.5361141967773437, 0.53621142578125, 0.5362186279296876, 0.5361489868164062, 0.5361234130859375, 0.5360465698242187, 0.53627392578125, 0.5361008911132813, 0.5359564819335938, 0.5358325805664063, 0.5361940307617188, 0.53595751953125, 0.535931884765625, 0.5361787109375, 0.536142822265625, 0.5360486450195312, 1.1542691650390624, 0.5355407104492188, 0.535552001953125, 0.5357998046875, 0.5357117309570313, 0.53576806640625, 0.5355653076171875, 0.5356492919921875, 0.5357056274414063, 0.5357301635742188, 0.5358960571289062, 0.5357343139648437, 0.5358621826171875, 0.5359462280273437, 0.5356953735351563, 0.53574755859375, 0.5355867919921875, 0.53578857421875, 0.5362288818359375, 0.5359042358398437, 0.5363026123046875, 0.5359984741210938, 0.5359585571289063, 0.5359851684570313, 0.5359820556640625, 0.5368453369140626, 0.5359144897460938, 0.5357189331054687, 0.5361551513671875, 0.5356165161132812, 0.5358305053710938, 0.5357547607421875, 0.535568359375, 0.535858154296875, 0.5361868896484375, 0.5357393798828125, 0.5358377075195313, 0.5361234130859375, 0.5355980834960937, 0.5359616088867187, 0.5359390869140624, 0.5359196166992187, 0.5360025634765625, 0.5359226684570313, 0.5374095458984375, 0.53625341796875, 0.5358622436523437, 0.5358540649414063, 0.5358479614257813, 0.5359902954101563, 0.5359892578125, 0.5361889038085937, 0.53610498046875, 0.5359605712890625, 0.5358724975585938, 0.5362933959960937, 0.5360650024414062, 0.5361162109375, 0.53609375, 0.5360209350585937, 0.5362667236328125, 0.5360220336914062, 0.5361018676757813, 1.1538472900390624, 0.5356400756835937, 0.5355745239257812, 0.5357875366210938, 0.53564208984375, 0.5358479614257813, 0.5355479125976562, 0.53600048828125, 0.5358909301757813, 0.5358223266601563, 0.5358356323242187, 0.535784423828125, 0.53595751953125, 0.5359810791015625, 0.535689208984375, 0.5359032592773437, 0.5357056274414063, 0.5373982543945313, 0.536026123046875, 0.53581005859375, 0.5361223754882812, 0.5358602294921875, 0.5356083374023437, 0.5358510131835937, 0.5360107421875, 0.5359534301757812, 0.5358714599609375, 0.535973876953125, 0.5361254272460938, 0.5358253784179687, 0.5358182373046875, 0.5358540649414063, 0.5357783203125, 0.53592578125, 0.53614697265625, 0.5357598876953125, 0.5358438110351562, 0.5359124755859375, 0.5355714721679687, 0.5358827514648438, 0.5358745727539063, 0.5369292602539063, 0.5359862060546875, 0.5359390869140624, 0.5358551025390625, 0.5360128173828125, 0.5357086791992187, 0.5359451904296875, 0.5359124755859375, 0.5360537719726562, 0.5361254272460938, 0.5364070434570313, 0.5360322265625, 0.5359759521484375, 0.5359114379882812, 0.5361408081054687, 0.5360117797851562, 0.5362329711914062, 0.5359042358398437, 0.536247314453125, 0.5362728881835938, 0.5360906372070312, 0.5360189208984375, 1.154509765625, 0.53538916015625, 0.535583740234375, 0.5358110961914062, 0.53564111328125, 0.5358591918945312, 0.535794677734375, 0.53629443359375, 0.5356840209960938, 0.536141845703125, 0.5361029052734375, 0.5356216430664062, 0.5357506713867187, 0.5357485961914062, 0.535677978515625, 0.535900146484375, 0.5358776245117187, 0.5359073486328125, 0.5359022216796875, 0.5357670288085937, 0.5359923095703125, 0.5357567749023437, 0.5358325805664063, 0.53583154296875, 0.5359749145507813, 0.5359882202148437, 0.5361643676757812, 0.5356267700195313, 0.5359032592773437, 0.5372846069335937, 0.5357987670898438, 0.5357557983398438, 0.53564111328125, 0.5357557983398438, 0.5359974365234375, 0.53570458984375, 0.5359288330078125, 0.5356820678710937, 0.5356134643554687, 0.5357014770507813, 0.53581005859375, 0.536121337890625, 0.5363609619140625, 0.5361182861328125, 0.535931884765625, 0.5358868408203125, 0.5356769409179688, 0.5360240478515625, 0.5362145385742187, 0.53619091796875, 0.535984130859375, 0.5360619506835937, 0.5362288818359375, 0.5361019287109375, 0.535953369140625, 0.5360773315429688, 0.5358131103515625, 0.5361151733398437, 0.53602099609375, 0.5363035888671875, 0.5361448974609375, 0.537607177734375, 0.5360465698242187, 1.152490478515625, 0.5355888671875, 0.5357168579101562, 0.5357189331054687, 0.5358223266601563, 0.5357875366210938, 0.5354557495117187, 0.535419921875, 0.535910400390625, 0.5357783203125, 0.535656494140625, 0.5357096557617187, 0.5355775756835938, 0.535868408203125, 0.5354669799804688, 0.5359646606445313, 0.5356707763671875, 0.5358212890625, 0.5357373657226563, 0.5356103515625, 0.535888916015625, 0.5358203125, 0.53557861328125, 0.5374166870117187, 0.5358920288085938, 0.5357567749023437, 0.53572607421875, 0.535762939453125, 0.5360291748046875, 0.5355325317382813, 0.5358171997070312, 0.535667724609375, 0.535636962890625, 0.535841796875, 0.5359800415039062, 0.5355899047851562, 0.535858154296875, 0.535857177734375, 0.5360076904296875, 0.536141845703125, 0.5357691040039062, 0.535562255859375, 0.5359052734375, 0.535930908203125, 0.53589501953125, 0.535868408203125, 0.5358192749023437, 0.5358561401367188, 0.535779296875, 0.5360015258789063, 0.5359196166992187, 0.5360353393554688, 0.5360230102539062, 0.5360404663085937, 0.5359431762695313, 0.5360660400390626, 0.535836669921875, 0.5378590698242187, 0.536036376953125, 0.5359093627929687, 0.536047607421875, 0.5361490478515625, 0.5359584350585938, 1.15319189453125, 0.5356452026367188, 0.5356328735351562, 0.5357281494140625, 0.5355192260742188, 0.535783447265625, 0.5356636352539063, 0.5359830932617188, 0.535942138671875, 0.5360783081054687, 0.535846923828125, 0.535783447265625, 0.5356226806640625, 0.5357219848632813, 0.5355213012695312, 0.5356687622070313, 0.536869873046875, 0.535815185546875, 0.5357864990234374, 0.5356318969726562, 0.5359800415039062, 0.535752685546875, 0.5358387451171875, 0.5357025146484375, 0.5358203125, 0.53583154296875, 0.535783447265625, 0.5358837890625, 0.5360977783203125, 0.5357271118164062, 0.5360045776367187, 0.5359462280273437, 0.53570458984375, 0.5356165161132812, 0.5363179931640625, 0.5356410522460937, 0.5359022216796875, 0.5356492919921875, 0.5357659912109375, 0.5357659912109375, 0.5358069458007813, 0.5357701416015626, 0.5360128173828125, 0.5359564819335938, 0.5359616088867187, 0.5359513549804688, 0.5358837890625, 0.5359401245117188, 0.5361172485351563, 0.53602099609375, 0.5360137939453125, 0.535984130859375, 0.5357609252929687, 0.53802392578125, 0.5357936401367187, 0.5359380493164062, 0.5357763061523437, 0.5359359130859375, 0.53591552734375, 0.5360404663085937, 0.536057861328125, 0.5359656982421875, 0.5361121215820313, 1.1536312255859376, 0.5356072998046875, 0.5357742309570312, 0.53574658203125, 0.53595751953125, 0.5358848266601562, 0.5356380004882813, 0.5355888671875, 0.5355899047851562, 0.5358806762695313, 0.5357998046875, 0.5358756103515625, 0.5356881713867188, 0.5356103515625, 0.5356820678710937, 0.5357035522460938, 0.5355663452148437, 0.5374146728515625, 0.5358694458007812, 0.535605224609375, 0.5359165649414063, 0.5356728515625, 0.53570458984375, 0.5357557983398438, 0.5357874755859375, 0.5357854614257812, 0.5357496948242187, 0.5358018188476562, 0.5360455932617187, 0.53602099609375, 0.536015869140625, 0.5357619018554688, 0.5357301635742188, 0.5356718139648438, 0.5361336059570313, 0.5355591430664063, 0.536069091796875, 0.5360435180664063, 0.5357219848632813, 0.5357691040039062, 0.5356461791992188, 0.5358233642578125, 0.536005615234375, 0.5359605712890625, 0.535826416015625, 0.5359052734375, 0.535815185546875, 0.5358131103515625, 0.535836669921875, 0.5360036010742187, 0.5360506591796875, 0.5360066528320313, 0.5359401245117188, 0.5358776245117187, 0.5359585571289063, 0.5360834350585938, 0.5360015258789063, 0.5387274169921875, 0.5362770385742187, 0.5362575073242187, 0.5361879272460938, 0.5358714599609375, 0.5359554443359374, 1.152668701171875, 0.5355140991210937, 0.5354434814453125, 0.536067138671875, 0.5357444458007813, 0.5358714599609375, 0.5354506225585938, 0.535709716796875, 0.53574755859375, 0.5359656982421875, 0.5358479614257813, 0.53581005859375, 0.53583154296875, 0.5357014770507813, 0.5356124267578125, 0.5356380004882813, 0.5356932983398438, 0.5358069458007813, 0.53576806640625, 0.5355909423828125, 0.5359882202148437, 0.5357588500976562, 0.5356349487304688, 0.5356339111328124, 0.5357086791992187, 0.5358038940429688, 0.5379215087890625, 0.5360066528320313, 0.5362769775390624, 0.53560009765625, 0.5356708984375, 0.5355826416015625, 0.535667724609375, 0.53564208984375, 0.535888916015625, 0.5357485961914062, 0.5358643798828125, 0.5358673706054687, 0.535709716796875, 0.5357373657226563, 0.53574755859375, 0.5356687622070313, 0.5359216918945312, 0.5356953735351563, 0.5358694458007812, 0.5359226684570313, 0.5359185791015625, 0.5357824096679688, 0.535889892578125, 0.536216552734375, 0.5360506591796875, 0.5361254272460938, 0.5361408081054687, 0.5358909301757813, 0.5358079833984375, 0.5359800415039062, 0.5356195678710938, 0.53587353515625, 0.5358345947265625, 0.5359073486328125, 0.535858154296875, 0.5357772827148437, 0.5359022216796875]",tokens/s,1.83571099958843,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,5086.445568,5956.435968,0.0,5309.988864,5047.4112,s,10,3.3951274108886715,0.33951274108886714,0.00038503905864011734,0.33945068359375,0.33990344848632814,0.34012059326171873,0.34029430908203123,"[0.33982321166992185, 0.3391507568359375, 0.33938748168945315, 0.33922314453125, 0.34033773803710937, 0.3395138854980469, 0.33896197509765624, 0.3398551940917969, 0.3395834045410156, 0.3392906188964844]",tokens/s,754.0217759692035,kWh,4.00638022908458e-06,2.1953130347537806e-06,1.9920034454532867e-05,2.6121727718371226e-05,tokens/kWh,9800270.593126083,MB,5086.445568,5956.435968,0.0,5309.988864,5098.333184,s,10,188.221390625,18.822139062500003,0.0011505271698466547,18.8222666015625,18.8231517578125,18.82387861328125,18.82446009765625,"[18.821546875, 18.82124609375, 18.822708984375, 18.82460546875, 18.822560546875, 18.82055859375, 18.8222265625, 18.822306640625, 18.820640625, 18.822990234375]",tokens/s,3.3471222261616944,kWh,0.00022222505159952025,0.00012179794381854662,0.0010774225934188715,0.0014214455888369383,tokens/kWh,44321.077426219425,,s,629,190.97790853881838,0.3036214762143376,0.0403576770019219,0.29874688720703124,0.2989168701171875,0.2990350280761719,0.6382961645507812,"[0.2991718444824219, 0.29857177734375, 0.2985912170410156, 0.2988011474609375, 0.29874789428710935, 0.2987694091796875, 0.29878680419921877, 0.29858917236328125, 0.29871923828125, 0.29875506591796874, 0.29862911987304686, 0.29869976806640625, 0.29868133544921877, 0.29870693969726564, 0.2988308410644531, 0.2986854248046875, 0.2986752014160156, 0.2989803466796875, 0.29859634399414064, 0.2987222900390625, 0.2987008056640625, 0.2986854248046875, 0.2986875, 0.2986332092285156, 0.2987663269042969, 0.29869259643554685, 0.2985953369140625, 0.29869976806640625, 0.298566650390625, 0.29867929077148436, 0.29865780639648437, 0.29856668090820315, 0.29851031494140623, 0.2987509765625, 0.2991912841796875, 0.2989486083984375, 0.29886465454101563, 0.2987008056640625, 0.29902334594726565, 0.2987386779785156, 0.29886257934570315, 0.29884622192382815, 0.29862298583984376, 0.2987376708984375, 0.29878475952148437, 0.2987100219726562, 0.2987274169921875, 0.29880218505859374, 0.2987960205078125, 0.29913088989257813, 0.2988216247558594, 0.29875405883789063, 0.2986885070800781, 0.29871206665039063, 0.2987673645019531, 0.29861785888671877, 0.29866598510742187, 0.2988287963867188, 0.2988287963867188, 0.29879705810546875, 0.2987857971191406, 0.2987950134277344, 0.6382120971679688, 0.29854412841796873, 0.2985953369140625, 0.2985646057128906, 0.29854104614257815, 0.298724365234375, 0.29859634399414064, 0.29881240844726564, 0.29854617309570314, 0.29873358154296875, 0.2988380126953125, 0.29890969848632815, 0.2988410949707031, 0.29858612060546874, 0.2986250305175781, 0.2987253723144531, 0.29870693969726564, 0.29874893188476564, 0.2987386779785156, 0.2988083190917969, 0.29867315673828126, 0.2986280822753906, 0.29873358154296875, 0.2987376708984375, 0.2986905517578125, 0.298777587890625, 0.29880523681640625, 0.29867828369140625, 0.2988226623535156, 0.2988072814941406, 0.2987427978515625, 0.29883905029296876, 0.29878680419921877, 0.2985922546386719, 0.29876837158203123, 0.29872332763671877, 0.29868646240234376, 0.2986690673828125, 0.2988482666015625, 0.29870181274414065, 0.29881857299804687, 0.29884518432617185, 0.2989219970703125, 0.2987427978515625, 0.2989219970703125, 0.2988748779296875, 0.2987110290527344, 0.29886566162109374, 0.2988779602050781, 0.29880523681640625, 0.29869464111328126, 0.298745849609375, 0.2987427978515625, 0.29859942626953123, 0.2986700744628906, 0.29881857299804687, 0.2987734985351562, 0.29879806518554686, 0.298703857421875, 0.2987110290527344, 0.298756103515625, 0.2991769714355469, 0.2987376708984375, 0.6383523559570312, 0.2986219482421875, 0.29856051635742187, 0.29854617309570314, 0.2987694091796875, 0.29871923828125, 0.298534912109375, 0.2986680297851563, 0.29889434814453125, 0.2986250305175781, 0.29877044677734377, 0.29878375244140626, 0.2986680297851563, 0.2991124572753906, 0.2987796630859375, 0.2987530212402344, 0.2987222900390625, 0.2987888488769531, 0.2988287963867188, 0.2987950134277344, 0.2988482666015625, 0.29875405883789063, 0.29862603759765627, 0.2985953369140625, 0.2988308410644531, 0.2987632751464844, 0.2986342468261719, 0.299082763671875, 0.2986588134765625, 0.2988001403808594, 0.29873358154296875, 0.29896295166015624, 0.29876123046875, 0.2986895446777344, 0.298682373046875, 0.2987694091796875, 0.29866494750976563, 0.298819580078125, 0.2986956787109375, 0.2985513000488281, 0.29881240844726564, 0.2988011474609375, 0.298814453125, 0.29863833618164065, 0.29881753540039063, 0.2988830871582031, 0.29885440063476565, 0.29881036376953124, 0.29866290283203123, 0.298682373046875, 0.29867315673828126, 0.29896395874023435, 0.29883700561523435, 0.2987694091796875, 0.2988328857421875, 0.29885748291015624, 0.2988564453125, 0.29895269775390626, 0.2987806701660156, 0.29891278076171873, 0.2988830871582031, 0.29890765380859374, 0.2989168701171875, 0.6389586181640625, 0.29872128295898437, 0.2986536865234375, 0.29879910278320315, 0.2986854248046875, 0.29872128295898437, 0.29872845458984376, 0.2987356262207031, 0.29869464111328126, 0.29878375244140626, 0.2991523742675781, 0.2986168212890625, 0.29872640991210936, 0.2986588134765625, 0.29912472534179685, 0.29872845458984376, 0.2986956787109375, 0.29874688720703124, 0.2986332092285156, 0.29874789428710935, 0.29879806518554686, 0.29864346313476564, 0.29869876098632814, 0.29862399291992187, 0.29865676879882813, 0.298787841796875, 0.2986106872558594, 0.29867724609375, 0.2988134460449219, 0.298756103515625, 0.29874072265625, 0.2988482666015625, 0.2986065979003906, 0.29880422973632814, 0.298756103515625, 0.2998691711425781, 0.2986905517578125, 0.298714111328125, 0.2987028503417969, 0.2987509765625, 0.2988083190917969, 0.2988206176757813, 0.2987182006835937, 0.298703857421875, 0.29875201416015623, 0.29893121337890627, 0.2988267517089844, 0.2989168701171875, 0.2986741638183594, 0.2989014892578125, 0.29890866088867185, 0.2989270935058594, 0.29876837158203123, 0.29879910278320315, 0.29877862548828127, 0.29884622192382815, 0.29892095947265623, 0.2989322204589844, 0.298756103515625, 0.2989404296875, 0.29872845458984376, 0.2988206176757813, 0.29878375244140626, 0.63944189453125, 0.29866494750976563, 0.29873458862304686, 0.29941351318359377, 0.2987222900390625, 0.2986700744628906, 0.298650634765625, 0.29863015747070315, 0.29849395751953123, 0.2986536865234375, 0.29863729858398436, 0.29870489501953124, 0.2986311950683594, 0.2986546936035156, 0.298714111328125, 0.298745849609375, 0.29884210205078127, 0.29883187866210936, 0.2987581481933594, 0.2987724914550781, 0.2988584899902344, 0.298640380859375, 0.2987530212402344, 0.29866598510742187, 0.2987734985351562, 0.2988441467285156, 0.29874993896484375, 0.2986393737792969, 0.2987663269042969, 0.2986833801269531, 0.29886053466796875, 0.29871206665039063, 0.29884622192382815, 0.2987632751464844, 0.2985902099609375, 0.29863116455078126, 0.29862911987304686, 0.29875506591796874, 0.2986752014160156, 0.2986854248046875, 0.29865472412109373, 0.298703857421875, 0.2987304992675781, 0.29888204956054687, 0.2987386779785156, 0.2986557312011719, 0.2988277893066406, 0.29886465454101563, 0.2989363098144531, 0.298903564453125, 0.2989363098144531, 0.2989404296875, 0.29923941040039065, 0.29895782470703125, 0.298714111328125, 0.2987028503417969, 0.2989168701171875, 0.2988216247558594, 0.29880935668945313, 0.29886465454101563, 0.29877658081054687, 0.2988164978027344, 0.29866598510742187, 0.6385172729492188, 0.2987161560058594, 0.29870181274414065, 0.29873458862304686, 0.2985277404785156, 0.29870181274414065, 0.29862399291992187, 0.2987222900390625, 0.29852978515625, 0.2984530029296875, 0.29863833618164065, 0.2985779113769531, 0.2986393737792969, 0.2987673645019531, 0.2987304992675781, 0.2987028503417969, 0.2987581481933594, 0.298576904296875, 0.2986188659667969, 0.29863116455078126, 0.29871307373046874, 0.2985830383300781, 0.29869259643554685, 0.29868441772460935, 0.29859634399414064, 0.29861785888671877, 0.2987222900390625, 0.2986536865234375, 0.2988011474609375, 0.2985912170410156, 0.2987427978515625, 0.29869259643554685, 0.2987960205078125, 0.2987673645019531, 0.2987694091796875, 0.29872946166992187, 0.29888204956054687, 0.29878475952148437, 0.2987581481933594, 0.29880422973632814, 0.2987530212402344, 0.29877044677734377, 0.298819580078125, 0.29879910278320315, 0.29877862548828127, 0.29886465454101563, 0.29878271484375, 0.29871923828125, 0.298576904296875, 0.2992762756347656, 0.29873663330078126, 0.2988359680175781, 0.29874688720703124, 0.29913394165039064, 0.2987376708984375, 0.29887387084960937, 0.29878988647460936, 0.29888101196289063, 0.29867724609375, 0.29872128295898437, 0.2988380126953125, 0.2989219970703125, 0.29882366943359373, 0.6379100341796875, 0.29849395751953123, 0.2987642822265625, 0.2986956787109375, 0.2986536865234375, 0.2986342468261719, 0.29868646240234376, 0.29869259643554685, 0.2986905517578125, 0.298967041015625, 0.2986536865234375, 0.2986956787109375, 0.29867111206054686, 0.2986670227050781, 0.29867724609375, 0.29880218505859374, 0.2986833801269531, 0.2987028503417969, 0.29863116455078126, 0.29869363403320315, 0.2986977233886719, 0.29881753540039063, 0.2988114013671875, 0.2987325439453125, 0.2990878601074219, 0.2988226623535156, 0.2988031921386719, 0.29867623901367185, 0.2988410949707031, 0.29869158935546875, 0.2987182006835937, 0.29871514892578127, 0.29878988647460936, 0.2987304992675781, 0.29869363403320315, 0.29874789428710935, 0.298756103515625, 0.2987591552734375, 0.2987161560058594, 0.29865676879882813, 0.29869876098632814, 0.29875711059570315, 0.2986875, 0.29886566162109374, 0.29875711059570315, 0.2986752014160156, 0.2987008056640625, 0.29882571411132813, 0.2986741638183594, 0.29876019287109373, 0.29889434814453125, 0.2987161560058594, 0.29932748413085936, 0.29880422973632814, 0.2988687438964844, 0.29856051635742187, 0.29914111328125, 0.29886566162109374, 0.29886566162109374, 0.298956787109375, 0.29882366943359373, 0.29876123046875, 0.29887387084960937, 0.6389237670898438, 0.2986998291015625, 0.29864031982421874, 0.2987253723144531, 0.2986250305175781, 0.2985779113769531, 0.2987386779785156, 0.2985850830078125, 0.29863629150390625, 0.29863729858398436, 0.29869158935546875, 0.29864141845703124, 0.2990602111816406, 0.2987724914550781, 0.2986608581542969, 0.2989906005859375, 0.2988503112792969, 0.29879705810546875, 0.29870489501953124, 0.298724365234375, 0.29857073974609377, 0.29866494750976563, 0.29862405395507813, 0.29871405029296877, 0.2985922546386719, 0.29864346313476564, 0.29879296875, 0.299109375, 0.2994646911621094, 0.2987386779785156, 0.29884722900390626, 0.29866598510742187, 0.29875711059570315, 0.2986332092285156, 0.29882571411132813, 0.2988114013671875, 0.29872332763671877, 0.2986680297851563, 0.2987653198242187, 0.29900082397460936, 0.2987509765625, 0.29862399291992187, 0.2986475524902344, 0.29878170776367186, 0.2986905517578125, 0.29885336303710935, 0.298756103515625, 0.29877044677734377, 0.2987724914550781, 0.29862911987304686, 0.2986526794433594, 0.298903564453125, 0.298745849609375, 0.29878988647460936, 0.2988226623535156, 0.2988164978027344, 0.2988072814941406, 0.29877862548828127, 0.29900390625, 0.29874072265625, 0.29883392333984377, 0.2988072814941406, 0.298956787109375, 0.6386237182617187, 0.29877658081054687, 0.2986854248046875, 0.29864651489257815, 0.2986045532226563, 0.2988031921386719, 0.2985267333984375, 0.2986618957519531, 0.29867315673828126, 0.29864651489257815, 0.29855230712890624, 0.29865472412109373, 0.2986639404296875, 0.2986680297851563, 0.2987386779785156, 0.2985472106933594, 0.2987100219726562, 0.2987182006835937, 0.2986219482421875, 0.29865164184570314, 0.29867724609375, 0.2986342468261719, 0.2985973815917969, 0.2986598510742188, 0.29906228637695315, 0.2986956787109375, 0.29896600341796875, 0.29861785888671877, 0.2990868530273437, 0.298724365234375, 0.2988769226074219, 0.2987591552734375, 0.29865164184570314, 0.298787841796875, 0.29884005737304686, 0.298682373046875, 0.29856768798828126, 0.2986956787109375, 0.29869158935546875, 0.2987182006835937, 0.2987448425292969, 0.2987745361328125, 0.2987950134277344, 0.29871923828125, 0.29873663330078126, 0.29879193115234376, 0.29870489501953124, 0.2987960205078125, 0.2988062744140625, 0.29881549072265623, 0.29877658081054687, 0.2988441467285156, 0.2987806701660156, 0.2988707885742187, 0.29862399291992187, 0.2986117248535156, 0.2987694091796875, 0.29873663330078126, 0.29883905029296876, 0.29911654663085935, 0.29874176025390625, 0.29878375244140626, 0.29891787719726565, 0.638328857421875, 0.2986250305175781, 0.2985748596191406, 0.29873663330078126, 0.29864346313476564, 0.29865472412109373, 0.29872845458984376, 0.2986977233886719, 0.29866598510742187, 0.29880935668945313, 0.29860147094726563, 0.29870693969726564, 0.29879193115234376, 0.2987386779785156, 0.29876019287109373, 0.2987950134277344, 0.29864141845703124, 0.29870693969726564, 0.29863629150390625, 0.29877044677734377, 0.29883905029296876, 0.29858917236328125, 0.2987734985351562, 0.29867623901367185, 0.29884518432617185, 0.2988001403808594, 0.298745849609375, 0.29894451904296876, 0.29879910278320315, 0.298703857421875, 0.29870489501953124, 0.29902334594726565, 0.2988717956542969, 0.298745849609375, 0.29904281616210937, 0.29879193115234376, 0.2988031921386719, 0.2987653198242187, 0.2987325439453125, 0.2988206176757813, 0.29881036376953124, 0.29879092407226565, 0.298650634765625, 0.29869158935546875, 0.29877862548828127, 0.29889739990234376, 0.29879705810546875, 0.29879193115234376, 0.29879296875, 0.29889022827148437, 0.29869876098632814, 0.2989056091308594, 0.2989189147949219, 0.2988441467285156, 0.29881857299804687, 0.29889022827148437, 0.29874789428710935, 0.29886257934570315, 0.2987694091796875, 0.298851318359375, 0.2988492736816406, 0.29875711059570315, 0.2988687438964844]",tokens/s,3.293574659040466,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,MB,1751.871488,7019.692032,0.0,6371.147776,6308.565504,s,10,1.2995915527343749,0.12995915527343752,0.00010629331555176086,0.12992606353759767,0.1301084671020508,0.13013301467895508,0.13015265274047852,"[0.13010301208496095, 0.12998130798339844, 0.13005027770996094, 0.12987632751464845, 0.12987759399414062, 0.12983052062988282, 0.12996211242675781, 0.1298900146484375, 0.13015756225585937, 0.12986282348632813]",tokens/s,1969.8496767031857,kWh,1.536989179428164e-06,8.421931469931006e-07,6.240510042909041e-06,8.619692369330305e-06,tokens/kWh,29699435.783911806,MB,1774.911488,7021.789184,0.0,6373.244928,6308.568064,s,10,18.235262329101566,1.8235262329101563,0.0053724342566725284,1.8236018676757813,1.8299890258789062,1.8311459411621094,1.8320714733886718,"[1.821221923828125, 1.823341552734375, 1.8137137451171874, 1.821026611328125, 1.82973193359375, 1.8247183837890626, 1.8323028564453125, 1.817093994140625, 1.8238621826171875, 1.8282491455078125]",tokens/s,34.54844732310685,kWh,2.1608392675293537e-05,1.1841649985889348e-05,6.936480044129123e-05,0.0001028148431024741,tokens/kWh,612751.9927955227,,s,630,18.233413635253903,0.028941926405164926,0.00036640385775941743,0.028865535736083983,0.029282304763793947,0.02959416265487671,0.030582825374603273,"[0.0289751033782959, 0.02879078483581543, 0.02918604850769043, 0.0285296630859375, 0.02856345558166504, 0.028519424438476562, 0.02915839958190918, 0.028868608474731446, 0.02905292892456055, 0.02879795265197754, 0.029087743759155273, 0.028656639099121094, 0.028816383361816408, 0.02853376007080078, 0.02878361511230469, 0.028653568267822265, 0.028646400451660156, 0.02854297637939453, 0.02879795265197754, 0.02898124885559082, 0.028824575424194337, 0.028735488891601563, 0.02854092788696289, 0.029090816497802735, 0.0288143367767334, 0.028932096481323243, 0.028717056274414062, 0.028865535736083983, 0.028693504333496093, 0.028701696395874023, 0.02878463935852051, 0.028876800537109375, 0.02876927947998047, 0.028604415893554686, 0.028659711837768553, 0.029861888885498046, 0.031046655654907225, 0.029558784484863283, 0.028917760848999025, 0.028492799758911135, 0.028908544540405274, 0.029309951782226562, 0.0287825927734375, 0.028631040573120117, 0.02882252883911133, 0.02896691131591797, 0.02896384048461914, 0.028676095962524413, 0.02855014419555664, 0.028807167053222657, 0.028857343673706053, 0.028647424697875977, 0.02859212875366211, 0.02894540786743164, 0.028827648162841796, 0.02902016067504883, 0.02938368034362793, 0.029916160583496092, 0.028964864730834962, 0.02897817611694336, 0.028865535736083983, 0.029266944885253908, 0.028907520294189453, 0.02896895980834961, 0.02874777603149414, 0.028935167312622072, 0.029082624435424805, 0.028932096481323243, 0.02916966438293457, 0.028951551437377928, 0.028854272842407228, 0.029328384399414063, 0.0287774715423584, 0.029026304244995117, 0.028847103118896485, 0.028837888717651368, 0.028767232894897462, 0.029237247467041014, 0.02893414306640625, 0.028791807174682618, 0.028851200103759765, 0.028923904418945313, 0.029068288803100587, 0.028692480087280273, 0.028873727798461913, 0.02877644729614258, 0.028967935562133788, 0.02899660873413086, 0.02878873634338379, 0.030908416748046875, 0.029042688369750977, 0.029282304763793947, 0.02879283142089844, 0.02889727973937988, 0.029263872146606446, 0.02872319984436035, 0.02879283142089844, 0.02891366386413574, 0.02900480079650879, 0.028854272842407228, 0.028833791732788085, 0.029770751953125, 0.02899865531921387, 0.028828672409057617, 0.029229055404663085, 0.028647424697875977, 0.029173759460449217, 0.02876108741760254, 0.028976127624511717, 0.028778495788574218, 0.0288143367767334, 0.028809215545654295, 0.028983295440673826, 0.02872012710571289, 0.028854272842407228, 0.028634111404418947, 0.02905292892456055, 0.0285665283203125, 0.028702720642089844, 0.028645376205444335, 0.028435455322265626, 0.028626943588256838, 0.028911615371704103, 0.029191167831420898, 0.02877132797241211, 0.028833791732788085, 0.028884992599487305, 0.02903654479980469, 0.03017932891845703, 0.02912563133239746, 0.02881331253051758, 0.028681215286254884, 0.02860032081604004, 0.028862464904785157, 0.028727296829223634, 0.028665855407714845, 0.029211648941040037, 0.029117439270019533, 0.028499967575073244, 0.02832691192626953, 0.028686336517333984, 0.028493824005126952, 0.02838015937805176, 0.02842624092102051, 0.02853785514831543, 0.028770303726196288, 0.028686336517333984, 0.02882252883911133, 0.028836864471435547, 0.028872703552246092, 0.0286167049407959, 0.029073408126831055, 0.028721151351928712, 0.028767232894897462, 0.028633087158203126, 0.029042688369750977, 0.02869964790344238, 0.028874752044677734, 0.029038591384887694, 0.028910591125488282, 0.028896287918090822, 0.028859359741210937, 0.028862464904785157, 0.028824575424194337, 0.02877644729614258, 0.028871679306030275, 0.028840959548950194, 0.028845056533813477, 0.02858598327636719, 0.028941312789916993, 0.028679168701171875, 0.029077503204345705, 0.028824575424194337, 0.028454912185668944, 0.028852224349975586, 0.02850099182128906, 0.028646400451660156, 0.028803071975708007, 0.02900480079650879, 0.028917760848999025, 0.02872422409057617, 0.028694528579711914, 0.02877644729614258, 0.028721151351928712, 0.028727296829223634, 0.02830438423156738, 0.028331008911132813, 0.02883072090148926, 0.028727296829223634, 0.028893184661865235, 0.02875494384765625, 0.02892083168029785, 0.028884992599487305, 0.02853273582458496, 0.028737535476684572, 0.028791807174682618, 0.028884992599487305, 0.028839935302734376, 0.028670976638793946, 0.028956672668457032, 0.02875699234008789, 0.028964864730834962, 0.028696575164794923, 0.028861440658569337, 0.028852224349975586, 0.030459903717041017, 0.029494272232055665, 0.028890111923217773, 0.028811264038085937, 0.02873651123046875, 0.028735488891601563, 0.028597248077392577, 0.02875187110900879, 0.029277183532714843, 0.02878873634338379, 0.028476415634155275, 0.029085695266723634, 0.029066240310668946, 0.02855833625793457, 0.02893721580505371, 0.028770303726196288, 0.02920243263244629, 0.028609535217285157, 0.028932096481323243, 0.02872319984436035, 0.02892083168029785, 0.028983295440673826, 0.028871679306030275, 0.028867584228515625, 0.02897715187072754, 0.028786687850952147, 0.028804096221923828, 0.0287825927734375, 0.02898124885559082, 0.028879871368408205, 0.02890547180175781, 0.029244415283203123, 0.028838911056518556, 0.028621824264526367, 0.02895257568359375, 0.02921881675720215, 0.02880512046813965, 0.029128704071044922, 0.02921062469482422, 0.028823551177978517, 0.029016063690185546, 0.028856319427490236, 0.029049856185913086, 0.028932096481323243, 0.02880512046813965, 0.028640256881713868, 0.029030399322509767, 0.0289751033782959, 0.028750848770141602, 0.028823551177978517, 0.029055999755859374, 0.028444671630859376, 0.030841856002807616, 0.030049280166625978, 0.02981068801879883, 0.029970432281494142, 0.028856319427490236, 0.028807167053222657, 0.028949535369873047, 0.028584928512573243, 0.02860339164733887, 0.029297664642333986, 0.028885055541992187, 0.028797887802124025, 0.028693504333496093, 0.028617727279663087, 0.029106176376342774, 0.028820480346679687, 0.028858367919921874, 0.028819456100463867, 0.029165567398071288, 0.028864511489868162, 0.02875289535522461, 0.028851200103759765, 0.028734464645385743, 0.029259775161743166, 0.02878054428100586, 0.028850175857543944, 0.02896076774597168, 0.02914508819580078, 0.028817407608032225, 0.028894208908081056, 0.029035520553588868, 0.029533184051513672, 0.029876224517822264, 0.028865535736083983, 0.028951551437377928, 0.029073408126831055, 0.029282304763793947, 0.028598272323608398, 0.028807167053222657, 0.02894643211364746, 0.029019136428833008, 0.028634111404418947, 0.030237695693969727, 0.03058995246887207, 0.029755392074584962, 0.029097984313964844, 0.029080575942993164, 0.02851430320739746, 0.028753919601440428, 0.02882252883911133, 0.028439552307128906, 0.02859212875366211, 0.02898227119445801, 0.028853248596191407, 0.029017087936401367, 0.028786687850952147, 0.029080575942993164, 0.02883072090148926, 0.029138944625854493, 0.029035520553588868, 0.028886016845703126, 0.028611583709716795, 0.028837888717651368, 0.02915635108947754, 0.029039615631103514, 0.028480512619018555, 0.028329984664916992, 0.028669952392578125, 0.028464128494262695, 0.028688383102416993, 0.028948480606079102, 0.028862464904785157, 0.028916736602783204, 0.029080575942993164, 0.028682239532470705, 0.028721151351928712, 0.02891263961791992, 0.02850201606750488, 0.028811264038085937, 0.028708864212036132, 0.028878847122192384, 0.028666879653930662, 0.029037567138671876, 0.02879078483581543, 0.02877644729614258, 0.028648448944091798, 0.02854604721069336, 0.0285614070892334, 0.02912563133239746, 0.02873139190673828, 0.028762111663818358, 0.029484031677246093, 0.029238271713256835, 0.02889727973937988, 0.02878873634338379, 0.028787712097167968, 0.028898303985595702, 0.02896384048461914, 0.028791807174682618, 0.028791807174682618, 0.028669952392578125, 0.029425664901733397, 0.029499391555786132, 0.029080575942993164, 0.029420543670654296, 0.029608959197998046, 0.028705791473388673, 0.029037567138671876, 0.029594623565673828, 0.029593599319458007, 0.029005823135375978, 0.029278207778930664, 0.02894540786743164, 0.029344768524169923, 0.029558784484863283, 0.029249536514282228, 0.02893824005126953, 0.029262847900390625, 0.02910310363769531, 0.029105152130126953, 0.029451263427734374, 0.029912063598632813, 0.029526016235351563, 0.029149183273315428, 0.029163520812988283, 0.02923520088195801, 0.02959769630432129, 0.02915839958190918, 0.029285375595092773, 0.029638656616210936, 0.028847103118896485, 0.029130752563476563, 0.02957414436340332, 0.029655040740966795, 0.029318143844604492, 0.029146112442016602, 0.028902399063110353, 0.028898303985595702, 0.029609983444213867, 0.029486080169677735, 0.029491199493408202, 0.02953011131286621, 0.029029375076293946, 0.028980224609375, 0.029107200622558595, 0.028786687850952147, 0.029413375854492187, 0.029050880432128907, 0.02896076774597168, 0.02956492805480957, 0.029155328750610353, 0.029240320205688477, 0.028940288543701172, 0.029608959197998046, 0.02892083168029785, 0.029066240310668946, 0.028889087677001952, 0.029045759201049806, 0.028923904418945313, 0.029045759201049806, 0.02833510398864746, 0.028695552825927735, 0.028941312789916993, 0.028892160415649414, 0.028859392166137695, 0.028733440399169922, 0.028951551437377928, 0.028717056274414062, 0.028823551177978517, 0.028838911056518556, 0.028933120727539063, 0.028628992080688476, 0.029154304504394532, 0.02872012710571289, 0.02891263961791992, 0.029039615631103514, 0.028838911056518556, 0.028818431854248046, 0.028924928665161134, 0.028976127624511717, 0.02880512046813965, 0.028791807174682618, 0.028910591125488282, 0.028890111923217773, 0.028876800537109375, 0.028923904418945313, 0.028926975250244142, 0.028475391387939454, 0.0287825927734375, 0.02891263961791992, 0.028914688110351562, 0.028668928146362304, 0.02852249526977539, 0.028964864730834962, 0.029128704071044922, 0.02879897689819336, 0.028848127365112306, 0.029046783447265623, 0.02854195213317871, 0.028430335998535155, 0.028602367401123048, 0.02904473686218262, 0.029035520553588868, 0.02879897689819336, 0.028817407608032225, 0.029029375076293946, 0.029014015197753908, 0.028753919601440428, 0.028676095962524413, 0.028427263259887696, 0.029050880432128907, 0.028770303726196288, 0.028854272842407228, 0.028485631942749022, 0.028844032287597656, 0.028482559204101563, 0.02877952003479004, 0.028902399063110353, 0.029059072494506837, 0.028644351959228515, 0.028656639099121094, 0.028762111663818358, 0.02892799949645996, 0.028988479614257812, 0.02888185691833496, 0.028856319427490236, 0.029099008560180665, 0.028844032287597656, 0.02872217559814453, 0.028886016845703126, 0.029026304244995117, 0.028527616500854492, 0.028668928146362304, 0.028613632202148437, 0.02917888069152832, 0.028871679306030275, 0.028916736602783204, 0.029321216583251954, 0.029249536514282228, 0.028999679565429686, 0.02897305679321289, 0.02881331253051758, 0.028956672668457032, 0.029070335388183592, 0.02880102348327637, 0.02876108741760254, 0.028701696395874023, 0.028682239532470705, 0.02914303970336914, 0.028884992599487305, 0.028890111923217773, 0.029024255752563476, 0.028845056533813477, 0.02903654479980469, 0.028824575424194337, 0.0287324161529541, 0.028844032287597656, 0.02878361511230469, 0.028580863952636718, 0.028564479827880858, 0.028636159896850585, 0.028675071716308592, 0.031180799484252928, 0.029594623565673828, 0.02877644729614258, 0.028926975250244142, 0.028827648162841796, 0.02919526481628418, 0.028821504592895508, 0.02900377655029297, 0.029698047637939453, 0.03100057601928711, 0.02909388732910156, 0.02943180847167969, 0.028597248077392577, 0.028655616760253907, 0.028725248336791992, 0.02917888069152832, 0.02891263961791992, 0.0287774715423584, 0.028787712097167968, 0.028872703552246092, 0.02877542304992676, 0.028661760330200195, 0.029030399322509767, 0.028791807174682618, 0.02857574462890625, 0.02897100830078125, 0.028993535995483398, 0.028816383361816408, 0.02872217559814453, 0.028816383361816408, 0.028663808822631837, 0.028841983795166014, 0.02897100830078125, 0.028864511489868162, 0.028516351699829103, 0.02878156852722168, 0.028826623916625976, 0.02893414306640625, 0.02879897689819336, 0.029214719772338867, 0.028898303985595702, 0.02874777603149414, 0.028673023223876954, 0.028980224609375, 0.02954444885253906, 0.028702720642089844, 0.028419071197509766, 0.02894233512878418, 0.02902016067504883, 0.028462080001831053, 0.029246463775634765, 0.029022207260131837, 0.029066240310668946, 0.028854272842407228, 0.028976127624511717, 0.028828672409057617, 0.02889625549316406, 0.028762111663818358, 0.0291778564453125, 0.0288143367767334, 0.028923904418945313, 0.029063167572021483, 0.02977894401550293, 0.030877695083618165, 0.029084672927856447, 0.028637184143066406, 0.02876620864868164, 0.02869964790344238, 0.02872319984436035, 0.029259775161743166, 0.02896384048461914, 0.02876825523376465, 0.028856319427490236, 0.028839935302734376, 0.02899865531921387, 0.028818431854248046, 0.029078527450561522, 0.02878361511230469, 0.028450815200805665, 0.028794879913330077, 0.029014015197753908, 0.028877824783325196, 0.02883072090148926, 0.02880614471435547, 0.028762111663818358, 0.028901376724243165, 0.028633087158203126, 0.028802047729492186, 0.028917760848999025, 0.028785663604736327, 0.02904473686218262, 0.029116416931152345, 0.03056537628173828, 0.029884416580200194, 0.029007871627807616, 0.02911846351623535, 0.02882252883911133, 0.028901376724243165, 0.029457408905029295, 0.028849151611328124, 0.028476415634155275, 0.028641279220581056, 0.029297664642333986, 0.030112768173217775, 0.029592575073242186, 0.029062143325805666, 0.029040639877319335, 0.02878976058959961, 0.0289751033782959, 0.02875801658630371, 0.02892799949645996]",tokens/s,34.551950205413476,,,,,,gemma,MB,4750.196736,6422.003712,0.0,5792.333824,5786.358272,s,1,11.1402783203125,11.1402783203125,0.0,11.1402783203125,11.1402783203125,11.1402783203125,11.1402783203125,[11.1402783203125],,kWh,5.05503895604161e-05,2.7689732429005306e-05,7.249394688400163e-05,0.00015073406887342302, @@ -3617,48 +3560,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,4108.419072,5287.44448,0.0,4640.997376,4330.668544,s,10,3.1997629089355466,0.31997629089355467,0.0010652547133575853,0.31959405517578127,0.3214676086425781,0.32149762878417965,0.32152164489746093,"[0.31874813842773436, 0.3210158386230469, 0.3191706848144531, 0.3194938354492188, 0.31946047973632813, 0.31844869995117187, 0.31969427490234376, 0.32074237060546873, 0.3214609375, 0.32152764892578123]",tokens/s,800.0592771580148,kWh,3.7604298649562735e-06,2.060550245869101e-06,1.8770327516251517e-05,2.4591307627076894e-05,tokens/kWh,10410182.487332419,MB,4108.419072,5287.44448,0.0,4640.997376,4515.667456,s,10,175.235673828125,17.5235673828125,0.0014736007130784414,17.5240908203125,17.525130859375,17.525280273437502,17.525399804687503,"[17.52509765625, 17.52158984375, 17.524423828125, 17.523767578125, 17.5244140625, 17.521322265625, 17.524623046875, 17.5235234375, 17.5254296875, 17.521482421875]",tokens/s,3.5951583729344847,kWh,0.00020692485703776285,0.00011341188782000648,0.001005294630623929,0.0013256313754816985,tokens/kWh,47524.523910055694,,s,629,177.83855282592782,0.28273219845139547,0.038060612034765025,0.2781388854980469,0.2783006591796875,0.2783752197265625,0.5978882275390625,"[0.27858328247070313, 0.27810714721679686, 0.27806002807617186, 0.2781603698730469, 0.2780487670898438, 0.2779842529296875, 0.27810406494140627, 0.27796890258789064, 0.27810302734375, 0.2779637756347656, 0.27821774291992185, 0.2781737060546875, 0.2781051025390625, 0.278034423828125, 0.2780057678222656, 0.27806719970703125, 0.27839590454101565, 0.27814501953125, 0.27807232666015624, 0.27819009399414063, 0.2782167053222656, 0.2782105712890625, 0.27820748901367187, 0.278297607421875, 0.27826995849609376, 0.2780712890625, 0.2780426330566406, 0.2781808776855469, 0.27829556274414063, 0.27816754150390627, 0.2781163635253906, 0.2781972351074219, 0.2782863464355469, 0.2780477294921875, 0.2781614074707031, 0.2781419372558594, 0.2782822265625, 0.27810302734375, 0.2781747131347656, 0.278181884765625, 0.27818597412109375, 0.27807949829101564, 0.2783580017089844, 0.27806002807617186, 0.27818496704101564, 0.278265869140625, 0.27827813720703126, 0.27825357055664063, 0.27825457763671874, 0.2781265869140625, 0.2781214599609375, 0.2780579833984375, 0.278192138671875, 0.27812353515625, 0.2781327209472656, 0.2782576599121094, 0.278287353515625, 0.27825357055664063, 0.27815731811523436, 0.27832012939453127, 0.2783068237304688, 0.2782637939453125, 0.599541748046875, 0.27804568481445313, 0.27802215576171874, 0.27826483154296877, 0.278076416015625, 0.2781112365722656, 0.277928955078125, 0.2780845947265625, 0.2779053955078125, 0.27808050537109374, 0.27793508911132814, 0.27789413452148437, 0.27800677490234377, 0.2780477294921875, 0.27790643310546875, 0.2779893798828125, 0.2780702819824219, 0.27814093017578123, 0.2780211181640625, 0.27843890380859376, 0.2781225280761719, 0.2781531982421875, 0.2780579833984375, 0.27811019897460937, 0.278002685546875, 0.27822796630859375, 0.2780057678222656, 0.2780579833984375, 0.27817266845703126, 0.27817266845703126, 0.27806002807617186, 0.27816754150390627, 0.27811224365234377, 0.278287353515625, 0.2780968933105469, 0.2781757507324219, 0.278150146484375, 0.2783088684082031, 0.2780928039550781, 0.27805490112304687, 0.2782443542480469, 0.278150146484375, 0.27803750610351563, 0.27807437133789065, 0.27804571533203126, 0.2782361145019531, 0.27828326416015625, 0.2781214599609375, 0.27816244506835935, 0.2782443542480469, 0.27803237915039064, 0.2781583251953125, 0.27821774291992185, 0.2781603698730469, 0.27803955078125, 0.27840408325195315, 0.27810202026367187, 0.27821978759765625, 0.2781163635253906, 0.27807333374023435, 0.2780528564453125, 0.27824639892578124, 0.27811431884765625, 0.59765966796875, 0.2781757507324219, 0.27797607421875, 0.2782167053222656, 0.2780559387207031, 0.2779996032714844, 0.2779862976074219, 0.2779637756347656, 0.27797503662109374, 0.27801702880859375, 0.2779002990722656, 0.27802215576171874, 0.27807537841796875, 0.27807537841796875, 0.2780313720703125, 0.2780508117675781, 0.2779525146484375, 0.2781388854980469, 0.27807232666015624, 0.27817779541015625, 0.2780928039550781, 0.2780682373046875, 0.27811737060546876, 0.2781368408203125, 0.27811737060546876, 0.2782392272949219, 0.2781337585449219, 0.278392822265625, 0.2782310485839844, 0.278265869140625, 0.27814093017578123, 0.27818496704101564, 0.27824026489257814, 0.27826483154296877, 0.27806719970703125, 0.27855975341796874, 0.27825665283203127, 0.2781685791015625, 0.2782085266113281, 0.2781481018066406, 0.27810919189453126, 0.27824331665039065, 0.278034423828125, 0.2783047790527344, 0.2781388854980469, 0.27817166137695315, 0.2782822265625, 0.2782945251464844, 0.27814910888671873, 0.27833242797851565, 0.27806207275390626, 0.2781890563964844, 0.27830374145507814, 0.27817984008789065, 0.27814093017578123, 0.2780979309082031, 0.2780887145996094, 0.2783211669921875, 0.27862014770507815, 0.27816549682617187, 0.2781839294433594, 0.2782105712890625, 0.27824639892578124, 0.5981255493164063, 0.278150146484375, 0.27785626220703125, 0.2780334167480469, 0.2780712890625, 0.27800167846679685, 0.2779176940917969, 0.27832321166992186, 0.2780078125, 0.2780569458007813, 0.2779176940917969, 0.27797708129882814, 0.2779504699707031, 0.2781634521484375, 0.27796990966796875, 0.277923828125, 0.2779791259765625, 0.2781951904296875, 0.2780712890625, 0.2781603698730469, 0.2780784606933594, 0.27812762451171874, 0.2781521911621094, 0.2781368408203125, 0.27850137329101565, 0.2782167053222656, 0.27812045288085935, 0.2780579833984375, 0.2780405883789063, 0.27810919189453126, 0.2783846435546875, 0.2782300109863281, 0.2781163635253906, 0.278108154296875, 0.2781839294433594, 0.27849114990234375, 0.27811328125, 0.27817880249023436, 0.2781163635253906, 0.27801190185546876, 0.2782033996582031, 0.27817266845703126, 0.27813580322265624, 0.27843585205078125, 0.27829144287109375, 0.2781808776855469, 0.2781808776855469, 0.2782023620605469, 0.2781583251953125, 0.27829656982421874, 0.2781194152832031, 0.27825357055664063, 0.2780968933105469, 0.27834878540039065, 0.278118408203125, 0.2780631103515625, 0.27816244506835935, 0.27819314575195314, 0.27817880249023436, 0.27821466064453126, 0.2786324462890625, 0.2782525329589844, 0.278255615234375, 0.5979771118164062, 0.27813784790039064, 0.27801089477539065, 0.278066162109375, 0.27801190185546876, 0.278086669921875, 0.2779729919433594, 0.27828326416015625, 0.2779862976074219, 0.27811328125, 0.2778961791992188, 0.2779228210449219, 0.2780139465332031, 0.2780231628417969, 0.27794329833984377, 0.2781051025390625, 0.2780784606933594, 0.2783068237304688, 0.2780641174316406, 0.27816754150390627, 0.27818801879882815, 0.2781562805175781, 0.2779955139160156, 0.27814501953125, 0.27816448974609376, 0.27828121948242185, 0.27810919189453126, 0.2782033996582031, 0.27807537841796875, 0.2782801818847656, 0.27810406494140627, 0.27813784790039064, 0.27835186767578124, 0.27854437255859377, 0.278118408203125, 0.27813479614257813, 0.27823822021484373, 0.2782248840332031, 0.2781470642089844, 0.2781808776855469, 0.27812762451171874, 0.278139892578125, 0.27823614501953126, 0.27833139038085936, 0.27812249755859375, 0.2782443542480469, 0.2781890563964844, 0.27816549682617187, 0.27817266845703126, 0.27839590454101565, 0.27812762451171874, 0.2781603698730469, 0.27825665283203127, 0.2782945251464844, 0.2781419372558594, 0.2781552734375, 0.2782054443359375, 0.27824639892578124, 0.2780979309082031, 0.27818292236328124, 0.27818292236328124, 0.27840103149414064, 0.27840103149414064, 0.59683837890625, 0.2780917663574219, 0.27800372314453126, 0.277970947265625, 0.27823410034179685, 0.2782843017578125, 0.27795455932617186, 0.278044677734375, 0.27804568481445313, 0.2779945068359375, 0.27793508911132814, 0.27800167846679685, 0.27796685791015624, 0.27805490112304687, 0.2780139465332031, 0.27813067626953125, 0.2780088195800781, 0.2781214599609375, 0.27797402954101563, 0.27808154296875, 0.27807537841796875, 0.278118408203125, 0.2779258728027344, 0.27817062377929686, 0.2780487670898438, 0.2783273010253906, 0.2781265869140625, 0.2780579833984375, 0.2780211181640625, 0.27823822021484373, 0.2781255798339844, 0.27804568481445313, 0.2780559387207031, 0.27821978759765625, 0.27810302734375, 0.27798016357421873, 0.27811019897460937, 0.27810302734375, 0.2781695861816406, 0.27807232666015624, 0.2781419372558594, 0.27825457763671874, 0.27811019897460937, 0.2782085266113281, 0.2781265869140625, 0.2782320556640625, 0.2781665344238281, 0.27819622802734373, 0.27819827270507813, 0.27824844360351564, 0.2780887145996094, 0.27829556274414063, 0.2781214599609375, 0.27816549682617187, 0.2781255798339844, 0.27803237915039064, 0.2779996032714844, 0.27817984008789065, 0.2780989379882812, 0.2781603698730469, 0.27807949829101564, 0.2785577087402344, 0.278192138671875, 0.59820849609375, 0.27816754150390627, 0.2782525329589844, 0.27804672241210937, 0.27811737060546876, 0.27817266845703126, 0.2780159912109375, 0.27801089477539065, 0.2779852905273438, 0.27800473022460936, 0.27799755859375, 0.2780313720703125, 0.2780928039550781, 0.2781419372558594, 0.2780712890625, 0.2780354614257812, 0.27805389404296876, 0.27812762451171874, 0.27807745361328123, 0.2782248840332031, 0.27809588623046877, 0.2781470642089844, 0.27810098266601563, 0.27829351806640623, 0.2781583251953125, 0.278287353515625, 0.278150146484375, 0.2780835876464844, 0.2781767578125, 0.27830987548828123, 0.27811019897460937, 0.278029296875, 0.2782105712890625, 0.2782822265625, 0.2781481018066406, 0.278034423828125, 0.2781163635253906, 0.27817779541015625, 0.2781245422363281, 0.27811532592773436, 0.27822592163085935, 0.2782637939453125, 0.27812762451171874, 0.27823309326171874, 0.2781532287597656, 0.2782105712890625, 0.27814297485351563, 0.2782689208984375, 0.278319091796875, 0.27834060668945315, 0.2782003173828125, 0.2781808776855469, 0.278192138671875, 0.2784429931640625, 0.27812249755859375, 0.2782085266113281, 0.27821875, 0.278245361328125, 0.2782423095703125, 0.2782637939453125, 0.278245361328125, 0.2781634521484375, 0.278361083984375, 0.5991997680664063, 0.27812249755859375, 0.27820748901367187, 0.27797503662109374, 0.2780415954589844, 0.27798834228515623, 0.27795559692382815, 0.2780201110839844, 0.27805490112304687, 0.27803955078125, 0.2779053955078125, 0.27794842529296876, 0.278066162109375, 0.2780702819824219, 0.2779914245605469, 0.27794638061523436, 0.27797503662109374, 0.2781532287597656, 0.2780139465332031, 0.27818801879882815, 0.2781051025390625, 0.278108154296875, 0.27811224365234377, 0.2780989379882812, 0.27801803588867186, 0.27859457397460935, 0.27808767700195314, 0.278076416015625, 0.2780999755859375, 0.2781941833496094, 0.2783006591796875, 0.2781562805175781, 0.2780907592773437, 0.27820440673828123, 0.2780057678222656, 0.27811328125, 0.27811019897460937, 0.27817779541015625, 0.27850445556640624, 0.2780948486328125, 0.27814093017578123, 0.27819314575195314, 0.2780907592773437, 0.2782013549804688, 0.27804672241210937, 0.27818496704101564, 0.2781634521484375, 0.2782720031738281, 0.2782156677246094, 0.27832525634765626, 0.2783416442871094, 0.2781368408203125, 0.2780928039550781, 0.2781245422363281, 0.27811224365234377, 0.278624267578125, 0.27811328125, 0.2781767578125, 0.27811328125, 0.2783078308105469, 0.27821466064453126, 0.27869287109375, 0.27829248046875, 0.5999595947265625, 0.278151123046875, 0.2783006591796875, 0.2781767578125, 0.2781562805175781, 0.278287353515625, 0.27801190185546876, 0.27812762451171874, 0.27812249755859375, 0.2780569458007813, 0.27806515502929685, 0.27807745361328123, 0.27812249755859375, 0.27806207275390626, 0.2780282897949219, 0.27807232666015624, 0.27803851318359374, 0.278192138671875, 0.2781460571289063, 0.2781552734375, 0.27821875, 0.2780917663574219, 0.27806002807617186, 0.27822283935546877, 0.278044677734375, 0.27817984008789065, 0.27797607421875, 0.2783211669921875, 0.2781972351074219, 0.2782423095703125, 0.2780641174316406, 0.2781972351074219, 0.27817880249023436, 0.2783529052734375, 0.27826791381835936, 0.27813067626953125, 0.2781532287597656, 0.2781481018066406, 0.27818701171875, 0.2780784606933594, 0.27830169677734373, 0.2783303527832031, 0.2781839294433594, 0.2782423095703125, 0.27816244506835935, 0.27821875, 0.2782740478515625, 0.27820645141601563, 0.2781583251953125, 0.278361083984375, 0.27816754150390627, 0.2782095336914063, 0.2782668762207031, 0.278319091796875, 0.2781327209472656, 0.2781767578125, 0.27830374145507814, 0.27822900390625, 0.2781603698730469, 0.27811737060546876, 0.27818801879882815, 0.2782023620605469, 0.2782637939453125, 0.6001541137695312, 0.27810202026367187, 0.27797708129882814, 0.27800167846679685, 0.2780498046875, 0.27807949829101564, 0.2780989379882812, 0.2780559387207031, 0.27792486572265623, 0.2780498046875, 0.2779361267089844, 0.27795559692382815, 0.2781327209472656, 0.27822592163085935, 0.2779873352050781, 0.27796786499023435, 0.2779811706542969, 0.278044677734375, 0.27802215576171874, 0.2781317138671875, 0.27807949829101564, 0.2781245422363281, 0.2779422607421875, 0.27812966918945314, 0.2780569458007813, 0.2781808776855469, 0.2780313720703125, 0.27810714721679686, 0.2781757507324219, 0.27822796630859375, 0.2781767578125, 0.27811431884765625, 0.27811532592773436, 0.27820440673828123, 0.27810098266601563, 0.27806207275390626, 0.27838668823242185, 0.2781460571289063, 0.27807232666015624, 0.27805389404296876, 0.27812249755859375, 0.2782218322753906, 0.27806207275390626, 0.27811328125, 0.27808050537109374, 0.278255615234375, 0.2782023620605469, 0.27814910888671873, 0.27822900390625, 0.2783365173339844, 0.27817266845703126, 0.27810406494140627, 0.2780968933105469, 0.27818597412109375, 0.27826791381835936, 0.27807232666015624, 0.27808563232421873, 0.27834368896484374, 0.27808050537109374, 0.27809381103515624, 0.27811224365234377, 0.2781562805175781, 0.2782167053222656]",tokens/s,3.5369158711928956,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch @@ -4810,6 +4738,33 @@ ChildProcessError: Traceback (most recent call last): OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,MB,2120.851456,5971.116032,0.0,5324.668928,5080.87552,s,10,3.168598297119141,0.3168598297119141,0.0006192749728236364,0.31659298706054684,0.31771173400878905,0.3178386398315429,0.31794016448974605,"[0.31654159545898436, 0.3159651184082031, 0.3165306091308594, 0.3165748596191406, 0.3166111145019531, 0.3162603149414063, 0.3169653625488281, 0.317500244140625, 0.31768353271484373, 0.31796554565429686]",tokens/s,807.9282256534466,kWh,3.7360648158937697e-06,2.0465616732053604e-06,1.909266284356309e-05,2.487528933266222e-05,tokens/kWh,10291337.583111528,MB,2120.851456,5971.116032,0.0,5324.668928,5264.397312,s,10,173.229013671875,17.3229013671875,0.001442343405648258,17.322640625,17.324698828124998,17.32475859375,17.32480640625,"[17.322501953125, 17.320978515625, 17.322779296875, 17.324615234375, 17.3214375, 17.3238359375, 17.32092578125, 17.324685546875, 17.324818359375, 17.322435546875]",tokens/s,3.636804174116735,kWh,0.0002045347902665122,0.0001121027537912869,0.0010436602984692331,0.0013602978425270322,tokens/kWh,46313.38669402326,,s,629,175.80616687011712,0.27950106020686355,0.03764112668054027,0.27495321655273436,0.2751682434082031,0.275264111328125,0.5915128442382812,"[0.27523379516601565, 0.274661376953125, 0.274724853515625, 0.27491839599609375, 0.274951171875, 0.27510580444335936, 0.2748528747558594, 0.2747484130859375, 0.27507199096679685, 0.27504229736328123, 0.2748876647949219, 0.27475149536132815, 0.27495013427734377, 0.2747965393066406, 0.2748590087890625, 0.27514981079101564, 0.27481396484375, 0.2747125854492187, 0.27480474853515624, 0.2747955322265625, 0.27480474853515624, 0.27484466552734377, 0.27495013427734377, 0.27517031860351565, 0.27505868530273436, 0.2749849548339844, 0.2749153137207031, 0.27494195556640627, 0.27481396484375, 0.27494091796875, 0.27504434204101563, 0.27508633422851564, 0.27493887329101563, 0.2749696044921875, 0.2749286499023437, 0.2749849548339844, 0.274840576171875, 0.2750412902832031, 0.2748518371582031, 0.27493887329101563, 0.27504537963867187, 0.2749624328613281, 0.27506072998046877, 0.27502590942382815, 0.27498904418945314, 0.27502386474609375, 0.2750044250488281, 0.27496551513671874, 0.27491839599609375, 0.27494808959960937, 0.2749266052246094, 0.275230712890625, 0.2749358215332031, 0.27504229736328123, 0.27491122436523435, 0.2749696044921875, 0.2749091796875, 0.2750750732421875, 0.2750627746582031, 0.2750382080078125, 0.27517745971679686, 0.2751324157714844, 0.5909329833984375, 0.2748221435546875, 0.274850830078125, 0.27494808959960937, 0.2748866577148438, 0.27481396484375, 0.27481497192382814, 0.2746982421875, 0.27488357543945313, 0.2749910888671875, 0.2749542541503906, 0.2747555847167969, 0.27491226196289065, 0.2749593505859375, 0.27475762939453124, 0.27509963989257813, 0.2749644775390625, 0.2748221435546875, 0.27491021728515624, 0.2749347839355469, 0.2748374938964844, 0.27479244995117186, 0.2748334045410156, 0.27480575561523435, 0.275162109375, 0.27477297973632814, 0.2749972839355469, 0.2748272399902344, 0.27477297973632814, 0.27502285766601564, 0.27500238037109376, 0.27488973999023436, 0.2749706115722656, 0.27487026977539064, 0.27494091796875, 0.27481497192382814, 0.27490103149414064, 0.2748630676269531, 0.27495526123046876, 0.27509963989257813, 0.2748713073730469, 0.2749706115722656, 0.27490509033203125, 0.27491021728515624, 0.27486514282226565, 0.27494091796875, 0.275009521484375, 0.2749347839355469, 0.2750474243164063, 0.27482009887695313, 0.2749542541503906, 0.2753064880371094, 0.27495321655273436, 0.2748876647949219, 0.2750382080078125, 0.27522457885742185, 0.2749296569824219, 0.2749931640625, 0.27509762573242186, 0.2750084838867187, 0.2749081726074219, 0.27519180297851564, 0.27540069580078125, 0.591584228515625, 0.2747432861328125, 0.2747330627441406, 0.2748590087890625, 0.2747699279785156, 0.2747914123535156, 0.27490304565429685, 0.2747688903808594, 0.2747484130859375, 0.2749573059082031, 0.27480679321289064, 0.27498291015625, 0.2748293151855469, 0.2748395385742187, 0.2748948364257813, 0.27490304565429685, 0.27479345703125, 0.2749020080566406, 0.2748395385742187, 0.27503411865234373, 0.2749962158203125, 0.27483648681640627, 0.2746746826171875, 0.27502490234375, 0.27492556762695314, 0.27492352294921873, 0.2751109008789063, 0.2748160095214844, 0.27496856689453125, 0.2748374938964844, 0.27486822509765624, 0.2749992980957031, 0.2751129455566406, 0.2752378845214844, 0.27494503784179686, 0.2747821960449219, 0.2750269470214844, 0.2747474060058594, 0.27489178466796876, 0.27480679321289064, 0.274988037109375, 0.27508633422851564, 0.2751375427246094, 0.2750771179199219, 0.2753034362792969, 0.27517132568359376, 0.2750689392089844, 0.2749839477539063, 0.2750013427734375, 0.27504537963867187, 0.2750392456054688, 0.2749071350097656, 0.27496755981445314, 0.2750474243164063, 0.27524710083007814, 0.2750433349609375, 0.275378173828125, 0.2749747314453125, 0.27503717041015624, 0.275198974609375, 0.2752286682128906, 0.27506585693359376, 0.2751170654296875, 0.591656982421875, 0.27494705200195313, 0.2747955322265625, 0.2749982604980469, 0.27489178466796876, 0.27486004638671874, 0.2748631591796875, 0.2749070739746094, 0.27490609741210936, 0.27486822509765624, 0.27512933349609375, 0.2749849548339844, 0.2750013427734375, 0.2750648193359375, 0.2749154052734375, 0.2749736328125, 0.27503411865234373, 0.2749020080566406, 0.27490509033203125, 0.2748538818359375, 0.2748743591308594, 0.2748590087890625, 0.275314697265625, 0.2750074768066406, 0.27493170166015624, 0.27485592651367186, 0.2748713073730469, 0.2749706115722656, 0.274850830078125, 0.27483544921875, 0.2748692626953125, 0.2748876647949219, 0.2751795349121094, 0.2749942321777344, 0.2750115051269531, 0.2749593505859375, 0.27538330078125, 0.2749153137207031, 0.275051513671875, 0.2748231811523438, 0.2750392456054688, 0.2750965881347656, 0.27516314697265626, 0.27500338745117187, 0.2748989562988281, 0.2750270080566406, 0.2750924072265625, 0.2749429626464844, 0.2748641357421875, 0.27500543212890627, 0.2751047668457031, 0.27499725341796877, 0.27502490234375, 0.2749624328613281, 0.2754734191894531, 0.27508428955078124, 0.2749716491699219, 0.274951171875, 0.275056640625, 0.27501055908203126, 0.27507815551757814, 0.2753699951171875, 0.2751201171875, 0.5915985717773438, 0.27489791870117186, 0.27461221313476564, 0.27466751098632813, 0.27515802001953127, 0.2747515258789063, 0.27481903076171876, 0.2746787719726563, 0.27523687744140624, 0.2747658386230469, 0.27481497192382814, 0.27480062866210936, 0.27491226196289065, 0.27502081298828124, 0.275230712890625, 0.27494503784179686, 0.27481292724609374, 0.2748610534667969, 0.27488973999023436, 0.2748989562988281, 0.274808837890625, 0.2750873718261719, 0.27476275634765623, 0.2749798278808594, 0.27482827758789063, 0.27477197265625, 0.27517745971679686, 0.27487335205078123, 0.27484466552734377, 0.27493170166015624, 0.2751436767578125, 0.27479345703125, 0.27482009887695313, 0.2749327392578125, 0.27509146118164063, 0.2748467102050781, 0.274904052734375, 0.2750013427734375, 0.27485592651367186, 0.27484466552734377, 0.27487640380859374, 0.2750904235839844, 0.27502386474609375, 0.27488357543945313, 0.2749859924316406, 0.2749624328613281, 0.2749624328613281, 0.27490509033203125, 0.2749992980957031, 0.2749296569824219, 0.2750894165039062, 0.2750320739746094, 0.2751068115234375, 0.274904052734375, 0.2750330810546875, 0.2751928405761719, 0.27508633422851564, 0.27502899169921874, 0.2750545959472656, 0.27521844482421876, 0.2751047668457031, 0.27494195556640627, 0.27498086547851563, 0.5913292846679687, 0.27486514282226565, 0.2747463684082031, 0.2746695556640625, 0.2749358215332031, 0.27481805419921873, 0.2750351257324219, 0.2749276123046875, 0.2748866577148438, 0.27499725341796877, 0.27486514282226565, 0.27480575561523435, 0.2748221435546875, 0.27527987670898435, 0.2748590087890625, 0.27488973999023436, 0.2749429626464844, 0.2750392456054688, 0.27490304565429685, 0.2748948364257813, 0.2749368286132812, 0.27494091796875, 0.2750382080078125, 0.27531878662109377, 0.27500238037109376, 0.27481497192382814, 0.274914306640625, 0.274956298828125, 0.27495013427734377, 0.2748590087890625, 0.27500543212890627, 0.274882568359375, 0.2749440002441406, 0.2747535400390625, 0.27498086547851563, 0.27493374633789064, 0.2751109008789063, 0.27491839599609375, 0.27491943359375, 0.27500543212890627, 0.27502899169921874, 0.2749910888671875, 0.2749020080566406, 0.27501773071289065, 0.274951171875, 0.2748794860839844, 0.275056640625, 0.2749296569824219, 0.2749542541503906, 0.27533209228515626, 0.27512832641601564, 0.2751447143554687, 0.2750904235839844, 0.27548773193359377, 0.2752286682128906, 0.2750412902832031, 0.27516110229492186, 0.2749358215332031, 0.27501568603515625, 0.27505255126953126, 0.2750392456054688, 0.27514163208007814, 0.27496038818359375, 0.5919866943359375, 0.27475762939453124, 0.27476480102539064, 0.27481292724609374, 0.2749163513183594, 0.2747904052734375, 0.27476480102539064, 0.2746921081542969, 0.2748784790039063, 0.2748374938964844, 0.2749306945800781, 0.27477297973632814, 0.2747740173339844, 0.2748692626953125, 0.2748272705078125, 0.2748190612792969, 0.274840576171875, 0.27484979248046876, 0.2747699279785156, 0.27494808959960937, 0.2749429626464844, 0.2747361145019531, 0.27487744140625, 0.2748190612792969, 0.27489791870117186, 0.27484161376953126, 0.27487539672851563, 0.2749286499023437, 0.2748272705078125, 0.2749624328613281, 0.2749542541503906, 0.2749819030761719, 0.27498086547851563, 0.27487744140625, 0.27491738891601564, 0.2748313598632813, 0.27493170166015624, 0.27491122436523435, 0.27477197265625, 0.27493374633789064, 0.2753536071777344, 0.27502590942382815, 0.27512832641601564, 0.2752010192871094, 0.2750873718261719, 0.2749706115722656, 0.2750044250488281, 0.2748794860839844, 0.2749593505859375, 0.27493374633789064, 0.2750330810546875, 0.2750064697265625, 0.274914306640625, 0.27502899169921874, 0.27496450805664063, 0.2750494384765625, 0.2749378662109375, 0.27504434204101563, 0.27510272216796877, 0.27506689453125, 0.2752511901855469, 0.27526861572265626, 0.27529931640625, 0.592447509765625, 0.27481396484375, 0.27480172729492186, 0.27488351440429687, 0.275167236328125, 0.27490609741210936, 0.2748323974609375, 0.27473101806640626, 0.2749521789550781, 0.2751825866699219, 0.2748968811035156, 0.2750433349609375, 0.2752634887695313, 0.2750279541015625, 0.27487026977539064, 0.2749358215332031, 0.275135498046875, 0.27480780029296875, 0.2748538818359375, 0.27485797119140626, 0.27519180297851564, 0.27487539672851563, 0.27492556762695314, 0.27502590942382815, 0.2747965393066406, 0.27502490234375, 0.2750382080078125, 0.2750330810546875, 0.2749992980957031, 0.27501669311523436, 0.275409912109375, 0.2749296569824219, 0.2748374938964844, 0.27500543212890627, 0.27533822631835936, 0.2748436584472656, 0.2749624328613281, 0.2749716491699219, 0.27507815551757814, 0.2749358215332031, 0.2750003051757812, 0.2750044250488281, 0.27506072998046877, 0.27513650512695315, 0.2749245300292969, 0.2750771179199219, 0.2749491271972656, 0.27497573852539064, 0.27496450805664063, 0.27504022216796875, 0.2749910888671875, 0.2749276123046875, 0.27496038818359375, 0.27504229736328123, 0.2751733703613281, 0.27498904418945314, 0.2750955505371094, 0.27518157958984374, 0.2749716491699219, 0.27494503784179686, 0.27501568603515625, 0.27508633422851564, 0.275051513671875, 0.5926707153320312, 0.2748610534667969, 0.27503717041015624, 0.27477914428710937, 0.27481396484375, 0.27488973999023436, 0.27527374267578125, 0.27483444213867186, 0.27480575561523435, 0.2750044250488281, 0.2749819030761719, 0.2750013427734375, 0.27490304565429685, 0.27549899291992186, 0.2750013427734375, 0.2748989562988281, 0.2749081726074219, 0.27493887329101563, 0.2749327392578125, 0.2749849548339844, 0.27481292724609374, 0.2752481384277344, 0.27504537963867187, 0.2749747314453125, 0.27479449462890626, 0.27488461303710937, 0.27496551513671874, 0.275009521484375, 0.2749921264648437, 0.2749091796875, 0.2747484130859375, 0.2750965881347656, 0.27488973999023436, 0.27474432373046875, 0.2750126037597656, 0.2748866577148438, 0.27504229736328123, 0.27475762939453124, 0.2748252258300781, 0.27492556762695314, 0.27502081298828124, 0.2748671875, 0.2751129455566406, 0.2752235412597656, 0.2750269470214844, 0.275019775390625, 0.2750249328613281, 0.2750699157714844, 0.2751201171875, 0.2750044250488281, 0.27518157958984374, 0.27500543212890627, 0.27503103637695314, 0.27491839599609375, 0.27523480224609376, 0.27506689453125, 0.2749921264648437, 0.27528192138671875, 0.2751682434082031, 0.2751682434082031, 0.27506689453125, 0.2754119567871094, 0.27491738891601564, 0.5930465087890625, 0.2747432861328125, 0.27484774780273435, 0.27488461303710937, 0.2747125854492187, 0.27473715209960936, 0.27486209106445314, 0.27480780029296875, 0.27513446044921874, 0.2748692626953125, 0.2748784790039063, 0.27492044067382815, 0.2748395385742187, 0.27494808959960937, 0.27470745849609374, 0.2749358215332031, 0.274808837890625, 0.2747218017578125, 0.2748190612792969, 0.2749542541503906, 0.274956298828125, 0.2748334045410156, 0.2748590087890625, 0.2749368286132812, 0.274840576171875, 0.27478732299804687, 0.27496755981445314, 0.2749214782714844, 0.2749378662109375, 0.27477297973632814, 0.2749306945800781, 0.2749358215332031, 0.27493170166015624, 0.27503616333007813, 0.27505255126953126, 0.2748671875, 0.2749573059082031, 0.2749542541503906, 0.2749521789550781, 0.2748374938964844, 0.27497882080078123, 0.27516415405273437, 0.27495013427734377, 0.27504229736328123, 0.27498904418945314, 0.27497573852539064, 0.2751590270996094, 0.2752645263671875, 0.27510885620117187, 0.2749091796875, 0.27504537963867187, 0.2753761291503906, 0.27506585693359376, 0.275051513671875, 0.27513958740234373, 0.27502590942382815, 0.2751457214355469, 0.275009521484375, 0.27514163208007814, 0.27507302856445315, 0.27506585693359376, 0.27519384765625, 0.2751242370605469]",tokens/s,3.577803959884384,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -5108,48 +5063,33 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 86.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -5220,39 +5160,33 @@ OSError: 1 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3710, in from_pretrained - model = cls(config, *model_args, **model_kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm-20b/80729bcf52fbc4553d965926b27304ac5e156d98/modeling_internlm.py"", line 906, in __init__ - self.model = InternLMModel(config) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm-20b/80729bcf52fbc4553d965926b27304ac5e156d98/modeling_internlm.py"", line 729, in __init__ - self.layers = nn.ModuleList([InternLMDecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm-20b/80729bcf52fbc4553d965926b27304ac5e156d98/modeling_internlm.py"", line 729, in - self.layers = nn.ModuleList([InternLMDecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm-20b/80729bcf52fbc4553d965926b27304ac5e156d98/modeling_internlm.py"", line 545, in __init__ - self.self_attn = INTERNLM_ATTENTION_CLASSES[config.attn_implementation](config=config) -KeyError: 'sdpa' + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,internlm,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,,MB,3137.929216,13793.492992,0.0,13147.045888,12781.924864,s,10,8.99033642578125,0.8990336425781249,0.001209152583374722,0.8988238220214844,0.9007634643554687,0.9011640808105469,0.9014845739746095,"[0.898102294921875, 0.8976475830078126, 0.89875732421875, 0.8988903198242187, 0.89919384765625, 0.8981031494140626, 0.8995744018554688, 0.9006744384765625, 0.901564697265625, 0.897828369140625]",tokens/s,284.7501893987843,kWh,1.0606409098815037e-05,5.811931839480167e-06,5.64317581083322e-05,7.285009904662741e-05,tokens/kWh,3514065.229151552,MB,3137.929216,13793.492992,0.0,13147.045888,12906.706432,s,10,495.9275859375,49.59275859375,0.0018491618139201286,49.592791015625,49.594271093749995,49.595561328125,49.596593515625,"[49.5933984375, 49.590046875, 49.59246875, 49.5923359375, 49.593984375, 49.59295703125, 49.5899609375, 49.592640625, 49.5968515625, 49.59294140625]",tokens/s,1.2703467559866628,kWh,0.0005854946296717283,0.00032090146071840533,0.0030803726957776686,0.003986768786167803,tokens/kWh,15802.270806017175,,s,629,503.2330070800783,0.8000524754850208,0.10683804867936002,0.78717236328125,0.7874144165039062,0.7875112915039062,1.685771015625,"[0.78681396484375, 0.787056640625, 0.7872000122070313, 0.7871314086914063, 0.7870761108398437, 0.7871968994140625, 0.7870853271484375, 0.7871027221679687, 0.7870504760742187, 0.7872420043945313, 0.7869163818359375, 0.7871047973632812, 0.7869481201171875, 0.7870822143554688, 0.787114990234375, 0.7870955810546875, 0.78725634765625, 0.7873034057617188, 0.7869389038085938, 0.787019775390625, 0.7870382080078125, 0.7876044921875, 0.7871713256835937, 0.7871876831054687, 0.7872317504882812, 0.7870443725585937, 0.787030029296875, 0.7872450561523437, 0.7871856689453125, 0.7870392456054688, 0.7871764526367188, 0.7872102661132813, 0.7872501831054688, 0.7873095703125, 0.7874119873046875, 0.787078125, 0.7872286987304687, 0.7874293823242188, 0.787230712890625, 0.7871242065429688, 0.7872471313476562, 0.7872798461914062, 0.7871795043945312, 0.7870382080078125, 0.7872808837890625, 0.7874662475585937, 0.7871713256835937, 0.7872593994140625, 0.7871129760742187, 0.7872962646484375, 0.7874160766601562, 0.7872471313476562, 0.7874232177734375, 0.787145751953125, 0.78729931640625, 0.7873556518554687, 0.7873126220703125, 0.7872225341796875, 0.7873648681640625, 0.7873320922851562, 0.7873925170898437, 0.787409912109375, 1.6845506591796875, 0.7865989379882813, 0.7869122314453125, 0.7870771484375, 0.786904052734375, 0.786988037109375, 0.7868211059570313, 0.7866941528320313, 0.7869010009765625, 0.7872276611328125, 0.7870422973632812, 0.7869706420898438, 0.78719384765625, 0.786861083984375, 0.78710888671875, 0.7869931640625, 0.7870863647460937, 0.7872379150390625, 0.7870750732421875, 0.786935791015625, 0.7871508178710938, 0.7870403442382813, 0.7873442993164063, 0.7869214477539063, 0.7870392456054688, 0.7870474243164063, 0.786840576171875, 0.7870504760742187, 0.78713037109375, 0.7870873413085937, 0.7869808349609375, 0.7869808349609375, 0.787267578125, 0.786956298828125, 0.787124267578125, 0.7870125732421875, 0.786892822265625, 0.7871385498046874, 0.78703515625, 0.7871498413085938, 0.7871344604492188, 0.78739453125, 0.7875164184570312, 0.7874877319335938, 0.7870628051757812, 0.787272705078125, 0.787435546875, 0.7874375610351563, 0.787373046875, 0.7877529296875, 0.7874273071289063, 0.787282958984375, 0.7873843383789062, 0.7874119873046875, 0.7875717163085938, 0.7875112915039062, 0.7872808837890625, 0.7871426391601563, 0.7872798461914062, 0.7872686157226563, 0.7872399291992187, 0.7873178100585938, 0.78739453125, 1.68635595703125, 0.7868969116210938, 0.787056640625, 0.7871334228515625, 0.7868477172851562, 0.7871549682617187, 0.787078125, 0.7870904541015625, 0.7870392456054688, 0.787114990234375, 0.7871692504882812, 0.7877386474609375, 0.7869378662109375, 0.7871314086914063, 0.7872255859375, 0.7870453491210937, 0.7869655151367188, 0.786951171875, 0.7870986328125, 0.7870658569335938, 0.7872614135742187, 0.7874027709960938, 0.7870945434570312, 0.78715185546875, 0.7872450561523437, 0.7871549682617187, 0.7872819213867187, 0.787478515625, 0.7873597412109375, 0.7869859619140624, 0.7874692993164063, 0.7872286987304687, 0.7869798583984375, 0.787325927734375, 0.7869276123046876, 0.7874866943359375, 0.7870545654296875, 0.787198974609375, 0.7871416015625, 0.787103759765625, 0.7871918334960938, 0.787167236328125, 0.7870433349609375, 0.78729931640625, 0.7875491943359375, 0.7871170654296875, 0.787009521484375, 0.7875952758789062, 0.7872061157226562, 0.7870023803710937, 0.7870484619140625, 0.7871426391601563, 0.787188720703125, 0.7870167236328125, 0.7873402709960937, 0.7872317504882812, 0.7870904541015625, 0.7873126220703125, 0.7873116455078125, 0.78721533203125, 0.7872767944335938, 0.7872000122070313, 0.7872696533203125, 1.686043701171875, 0.7870218505859375, 0.7873935546875, 0.7871447143554687, 0.787346435546875, 0.7868549194335938, 0.7871016845703125, 0.7867627563476562, 0.7869368286132813, 0.7868671875, 0.7868692626953125, 0.786977783203125, 0.7869685668945312, 0.7871201171875, 0.7870084838867187, 0.7870965576171876, 0.7871498413085938, 0.7871580200195313, 0.7870658569335938, 0.7869450073242188, 0.7869378662109375, 0.787483642578125, 0.7873648681640625, 0.7870310668945313, 0.7869224853515625, 0.7871365966796875, 0.787136474609375, 0.7871539306640625, 0.787125244140625, 0.7869398803710937, 0.78703515625, 0.7872276611328125, 0.7871282958984375, 0.7870628051757812, 0.7872921752929688, 0.7870003051757812, 0.7870494995117188, 0.7871273193359375, 0.7872399291992187, 0.787398681640625, 0.78772119140625, 0.78713037109375, 0.787267578125, 0.7872276611328125, 0.7871488037109375, 0.787314697265625, 0.7874241943359375, 0.7870822143554688, 0.7873843383789062, 0.7872819213867187, 0.7872849731445313, 0.78731982421875, 0.7872379150390625, 0.787399658203125, 0.7873013916015625, 0.7877857055664063, 0.7872245483398438, 0.7873157348632812, 0.7872655639648437, 0.7875112915039062, 0.7873760986328125, 0.787336181640625, 0.78727783203125, 1.68654541015625, 0.7870822143554688, 0.7871160278320313, 0.7870023803710937, 0.7870228271484375, 0.7869972534179688, 0.786967529296875, 0.7874447631835938, 0.7872890625, 0.7869368286132813, 0.7870853271484375, 0.7870689086914062, 0.7870709838867187, 0.78703515625, 0.787505126953125, 0.7871375122070312, 0.7872081909179688, 0.7871528930664062, 0.7871610717773437, 0.787219482421875, 0.7872010498046875, 0.7872266235351563, 0.787040283203125, 0.7871918334960938, 0.7872061157226562, 0.7872471313476562, 0.7872041015625, 0.787078125, 0.7871713256835937, 0.7871109008789062, 0.7871968994140625, 0.7870965576171876, 0.7871549682617187, 0.78739453125, 0.787336181640625, 0.7870084838867187, 0.7874140014648437, 0.7870863647460937, 0.7871047973632812, 0.7871129760742187, 0.7871273193359375, 0.78738330078125, 0.7872819213867187, 0.7871928100585938, 0.7872757568359375, 0.7873280029296875, 0.7872593994140625, 0.7870822143554688, 0.7872921752929688, 0.7870812377929688, 0.78708837890625, 0.7870607299804687, 0.7874990234375, 0.7872245483398438, 0.78721435546875, 0.7873689575195313, 0.7873966064453125, 0.7875502319335937, 0.7875901489257813, 0.7873382568359375, 0.7871918334960938, 0.7875266723632812, 0.7872737426757812, 1.68506982421875, 0.7869859619140624, 0.7871129760742187, 0.7870054321289063, 0.7870218505859375, 0.78700341796875, 0.787030029296875, 0.7872501831054688, 0.7870105590820312, 0.786934814453125, 0.7869972534179688, 0.786946044921875, 0.787056640625, 0.787325927734375, 0.787252197265625, 0.7869522094726562, 0.7875215454101563, 0.7873843383789062, 0.78717236328125, 0.7871324462890625, 0.7873607788085938, 0.78706689453125, 0.7871395874023438, 0.78725634765625, 0.78721435546875, 0.7872757568359375, 0.7873546142578125, 0.7874447631835938, 0.7869368286132813, 0.78706689453125, 0.7870453491210937, 0.7871375122070312, 0.7872071533203125, 0.7872235717773437, 0.7871477661132813, 0.7870607299804687, 0.78715185546875, 0.7871478271484375, 0.7873289794921875, 0.7873003540039063, 0.787083251953125, 0.7871488037109375, 0.78725732421875, 0.7871795043945312, 0.78721533203125, 0.7876454467773437, 0.787178466796875, 0.7871815795898438, 0.7872112426757812, 0.78717236328125, 0.7872327880859376, 0.7870750732421875, 0.7871897583007812, 0.7872420043945313, 0.787294189453125, 0.7872420043945313, 0.7872491455078126, 0.7874774780273438, 0.7874058227539062, 0.7875645141601563, 0.7871815795898438, 0.7874334716796875, 0.7871918334960938, 1.687065673828125, 0.7868405151367187, 0.7869450073242188, 0.7869102172851562, 0.7868426513671875, 0.7869398803710937, 0.7869041137695313, 0.7869142456054687, 0.786946044921875, 0.786820068359375, 0.78691943359375, 0.787078125, 0.7869788208007813, 0.786951171875, 0.7870924682617187, 0.78710986328125, 0.7872061157226562, 0.787040283203125, 0.7869204711914063, 0.7873966064453125, 0.7871181030273438, 0.7870392456054688, 0.7870648193359375, 0.787114013671875, 0.7872133178710937, 0.7872122802734375, 0.7869757690429687, 0.7872870483398438, 0.7871006469726562, 0.7871600341796875, 0.7870965576171876, 0.7870996704101563, 0.7871242065429688, 0.7870771484375, 0.7871447143554687, 0.7871078491210938, 0.7870576782226563, 0.7873760986328125, 0.7871334228515625, 0.7871221923828124, 0.7870075073242188, 0.7875072021484375, 0.7872061157226562, 0.7871661987304688, 0.7876239624023438, 0.787252197265625, 0.7872225341796875, 0.7872481079101562, 0.78718359375, 0.7872593994140625, 0.7872921752929688, 0.7870996704101563, 0.7873925170898437, 0.78729931640625, 0.7872388916015625, 0.7874826049804687, 0.7871661987304688, 0.78746728515625, 0.7872122802734375, 0.7872655639648437, 0.787220458984375, 0.787262451171875, 0.7870750732421875, 1.6877587890625, 0.7871488037109375, 0.7872276611328125, 0.787061767578125, 0.7871856689453125, 0.7868661499023437, 0.7870802001953126, 0.7870126342773438, 0.787072998046875, 0.7868098754882813, 0.7869603881835937, 0.7869419555664062, 0.7873136596679687, 0.78706689453125, 0.787056640625, 0.7869767456054687, 0.7870965576171876, 0.7870637817382813, 0.7873054809570312, 0.7870341186523437, 0.7869490966796875, 0.7870863647460937, 0.7871129760742187, 0.78717236328125, 0.7871334228515625, 0.7870802001953126, 0.7873003540039063, 0.7872020263671875, 0.7871979370117187, 0.7872655639648437, 0.7871344604492188, 0.7875747680664062, 0.7872235717773437, 0.7872450561523437, 0.787177490234375, 0.7873760986328125, 0.787162109375, 0.7870628051757812, 0.7869501342773437, 0.7871918334960938, 0.7872214965820312, 0.7871826171875, 0.7872737426757812, 0.78723583984375, 0.7870596923828125, 0.7872890625, 0.7871928100585938, 0.7872849731445313, 0.7872379150390625, 0.7874027709960938, 0.7871641845703125, 0.7871498413085938, 0.7873095703125, 0.7872429809570313, 0.7872481079101562, 0.7872973022460937, 0.7877294311523437, 0.7872665405273438, 0.7872921752929688, 0.78738330078125, 0.7872399291992187, 0.78740478515625, 0.7873894653320312, 1.6892119140625, 0.7867801513671875, 0.7874058227539062, 0.787093505859375, 0.7871375122070312, 0.7874641723632813, 0.7870320434570313, 0.7867852783203125, 0.7870791625976562, 0.7871867065429687, 0.7870504760742187, 0.7871549682617187, 0.7870392456054688, 0.7870269165039062, 0.7870709838867187, 0.7869685668945312, 0.7870894165039063, 0.7870361328125, 0.78729931640625, 0.7869368286132813, 0.7870637817382813, 0.7870525512695312, 0.78691943359375, 0.7871181030273438, 0.7872645263671875, 0.787114013671875, 0.7869951782226563, 0.78718359375, 0.7872911376953124, 0.787188720703125, 0.7874611206054688, 0.7871201171875, 0.7871590576171875, 0.7871764526367188, 0.7872604370117188, 0.7872041015625, 0.787367919921875, 0.787430419921875, 0.7872379150390625, 0.7872767944335938, 0.7871437377929688, 0.7872286376953125, 0.7871508178710938, 0.7872225341796875, 0.7872214965820312, 0.78721435546875, 0.7874846801757812, 0.787336181640625, 0.787209228515625, 0.7872471313476562, 0.78731982421875, 0.787367919921875, 0.7874590454101562, 0.7876321411132813, 0.787188720703125, 0.7873546142578125, 0.7873054809570312, 0.7873382568359375, 0.7870945434570312, 0.7873085327148438, 0.7871641845703125, 0.787177490234375, 0.7871334228515625, 1.6885872802734374, 0.7868467407226563, 0.7870679321289062, 0.78702490234375, 0.7873843383789062, 0.7869224853515625, 0.7869368286132813, 0.7871651611328125, 0.7870330810546875, 0.7871365356445312, 0.7869951782226563, 0.787177490234375, 0.7871631469726562, 0.78697265625, 0.7870187377929687, 0.7872450561523437, 0.7869921264648437, 0.7871190795898437, 0.7870576782226563, 0.7870761108398437, 0.78708837890625, 0.7871754150390625, 0.7870689086914062, 0.787504150390625, 0.7870955810546875, 0.7871426391601563, 0.7869767456054687, 0.7875563354492188, 0.7872174072265625, 0.7870545654296875, 0.7871416015625, 0.787072021484375, 0.7871344604492188, 0.7871682739257813, 0.7871314086914063, 0.787078125, 0.7870269165039062, 0.7872860107421875, 0.7870443725585937, 0.7871928100585938, 0.787251220703125, 0.7872686157226563, 0.7878246459960937, 0.7873013916015625, 0.7872030639648437, 0.787198974609375, 0.7872061157226562, 0.78739453125, 0.7873382568359375, 0.7876372680664062, 0.787146728515625, 0.787293212890625, 0.7874549560546875, 0.7871702880859375, 0.7872808837890625, 0.7872481079101562, 0.7872327880859376, 0.7873157348632812, 0.7872184448242188, 0.7873003540039063, 0.7872634887695312, 0.7872921752929688, 0.7874928588867187]",tokens/s,1.2499180124325768,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-30b,facebook/opt-30b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -5308,48 +5242,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2_moe,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,mixtral,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,MB,1341.939712,1292.36992,0.0,645.922816,583.517696,s,10,0.33003395080566406,0.033003395080566406,0.0009490722519561337,0.032697120666503904,0.033060840225219726,0.034453939247131346,0.03556841846466064,"[0.035847038269042966, 0.03269696044921875, 0.0326627197265625, 0.03274041748046875, 0.032751262664794924, 0.032656383514404294, 0.032584129333496095, 0.032697280883789065, 0.03266003036499023, 0.032737728118896484]",tokens/s,7756.777730747528,kWh,3.892594550828836e-07,2.13294823252478e-07,1.2394189374407605e-06,1.8419732157761221e-06,tokens/kWh,138981391.15564364,MB,1341.939712,1292.36992,0.0,645.922816,617.915392,s,10,20.267458740234375,2.0267458740234376,0.007313239917815637,2.025048645019531,2.03470361328125,2.0397496337890626,2.0437864501953125,"[2.044795654296875, 2.020955810546875, 2.0253223876953124, 2.029792724609375, 2.019128662109375, 2.0225035400390623, 2.020671142578125, 2.02477490234375, 2.033582275390625, 2.025931640625]",tokens/s,31.08431146078231,kWh,2.4787268574888778e-05,1.3584073468072775e-05,7.156591708455636e-05,0.00010993725912751792,tokens/kWh,573054.1265079688,,s,629,20.527176671981813,0.03263462110013006,0.003914486757335895,0.032086017608642575,0.032351640319824224,0.033026662445068355,0.06488547119140625,"[0.032965633392333986, 0.033142784118652346, 0.03345100784301758, 0.033375232696533204, 0.0333199348449707, 0.03362508773803711, 0.03331891250610351, 0.03328307342529297, 0.03301068878173828, 0.03359743881225586, 0.033181697845458984, 0.032979969024658204, 0.032715774536132815, 0.03390156936645508, 0.03343667221069336, 0.033037311553955076, 0.032322559356689456, 0.032176128387451174, 0.0319866886138916, 0.03207167816162109, 0.032007167816162106, 0.031942655563354495, 0.032176128387451174, 0.032004096984863284, 0.03197644805908203, 0.03201638412475586, 0.032069633483886716, 0.03198054313659668, 0.0320819206237793, 0.03197747230529785, 0.03204710388183594, 0.03205017471313477, 0.03205120086669922, 0.03233484649658203, 0.032089088439941404, 0.03201638412475586, 0.031940607070922854, 0.03196518325805664, 0.03219353485107422, 0.03215871810913086, 0.03197747230529785, 0.032091136932373046, 0.03205222320556641, 0.032072704315185545, 0.03213516616821289, 0.03204710388183594, 0.03202560043334961, 0.0320706558227539, 0.03204198455810547, 0.032328704833984374, 0.03220889663696289, 0.032048126220703126, 0.032674816131591795, 0.032985088348388675, 0.03251302337646484, 0.03211468887329102, 0.0320552978515625, 0.032077823638916016, 0.03209523010253906, 0.03212492752075195, 0.03207372665405273, 0.032254974365234376, 0.06495231628417969, 0.03237580871582031, 0.03187612724304199, 0.03206140899658203, 0.03218124771118164, 0.03197337532043457, 0.03209625625610352, 0.03205836868286133, 0.032023551940917966, 0.032107521057128906, 0.032069633483886716, 0.032031742095947266, 0.03211775970458984, 0.03206655883789063, 0.032097278594970705, 0.032092159271240234, 0.03202969741821289, 0.03243315124511719, 0.03213721466064453, 0.03245568084716797, 0.03208703994750976, 0.032036865234375, 0.03220479965209961, 0.032077823638916016, 0.03195289611816406, 0.03199385643005371, 0.03206655883789063, 0.031974399566650394, 0.03200102233886719, 0.03200307083129883, 0.03195699119567871, 0.032130046844482424, 0.03212799835205078, 0.03219769668579102, 0.03176134490966797, 0.03197952079772949, 0.03210444641113281, 0.032161792755126956, 0.03207167816162109, 0.03207372665405273, 0.032113662719726564, 0.031941631317138675, 0.03214438247680664, 0.032054271697998044, 0.032077823638916016, 0.032043006896972655, 0.03208499145507813, 0.03201331329345703, 0.03213516616821289, 0.03211980819702148, 0.03219148635864258, 0.03205222320556641, 0.03206860733032227, 0.032094207763671875, 0.0318474235534668, 0.03189043235778809, 0.03202252960205078, 0.03202252960205078, 0.03219251251220703, 0.03199283218383789, 0.0320634880065918, 0.03205017471313477, 0.032233470916748046, 0.06501068878173828, 0.032080894470214845, 0.031971328735351565, 0.0320634880065918, 0.03197952079772949, 0.032020481109619144, 0.032099327087402346, 0.03209625625610352, 0.03205120086669922, 0.0321638412475586, 0.03222323226928711, 0.03202969741821289, 0.0320819206237793, 0.03201536178588867, 0.032123905181884765, 0.03211264038085938, 0.032107521057128906, 0.032054271697998044, 0.03208806228637695, 0.03202764892578125, 0.03203891372680664, 0.03204608154296875, 0.03206655883789063, 0.032031742095947266, 0.032173057556152344, 0.032159744262695314, 0.032178176879882815, 0.03203071975708008, 0.032, 0.03219251251220703, 0.03220991897583008, 0.03215052795410156, 0.03212287902832031, 0.03214438247680664, 0.03208806228637695, 0.032189441680908204, 0.03358617782592774, 0.032282623291015625, 0.03224883270263672, 0.032129024505615236, 0.03215359878540039, 0.03218329620361328, 0.03209011077880859, 0.03211775970458984, 0.03201638412475586, 0.03225600051879883, 0.03211673736572265, 0.03223859024047852, 0.03225088119506836, 0.03216793441772461, 0.03224883270263672, 0.03214438247680664, 0.03214233779907227, 0.03219968032836914, 0.032086017608642575, 0.03218841552734375, 0.03218329620361328, 0.0322344970703125, 0.032053249359130856, 0.03201740646362305, 0.03223756790161133, 0.032105472564697264, 0.032145408630371096, 0.06503936004638672, 0.03211161422729492, 0.03218227386474609, 0.03204915237426758, 0.03210137557983399, 0.0321003532409668, 0.032008190155029294, 0.03204608154296875, 0.032110591888427735, 0.032113662719726564, 0.03195699119567871, 0.032056320190429685, 0.033364990234375, 0.03484672164916992, 0.032486400604248046, 0.03219558334350586, 0.032039936065673826, 0.03206655883789063, 0.032054271697998044, 0.03215462493896484, 0.03211264038085938, 0.03203276824951172, 0.0319815673828125, 0.03220787048339844, 0.032091136932373046, 0.032097278594970705, 0.03231846237182617, 0.03222016143798828, 0.03202150344848633, 0.032233470916748046, 0.03211775970458984, 0.03213824081420898, 0.032110591888427735, 0.03215155029296875, 0.032023551940917966, 0.0320634880065918, 0.03200614547729492, 0.032024574279785153, 0.03201740646362305, 0.03204095840454101, 0.0321710090637207, 0.03213516616821289, 0.03200511932373047, 0.032059391021728514, 0.03214745712280274, 0.03209011077880859, 0.03227852630615234, 0.03380428695678711, 0.03240345764160156, 0.03237171173095703, 0.03226521682739258, 0.0320706558227539, 0.03210649490356445, 0.03227443313598633, 0.032080894470214845, 0.0321525764465332, 0.03206655883789063, 0.032039936065673826, 0.032299007415771484, 0.032091136932373046, 0.03222016143798828, 0.032230400085449216, 0.03210956954956055, 0.06500863647460937, 0.03205734252929687, 0.03189452743530274, 0.03201740646362305, 0.03198464012145996, 0.03199692726135254, 0.032059391021728514, 0.032039936065673826, 0.03207372665405273, 0.032036865234375, 0.03203891372680664, 0.03213926315307617, 0.03204403305053711, 0.032121856689453124, 0.03199078369140625, 0.03199692726135254, 0.03200614547729492, 0.03198361587524414, 0.03177984046936035, 0.0318474235534668, 0.03203481674194336, 0.03219148635864258, 0.031959039688110355, 0.03210649490356445, 0.032053249359130856, 0.03217510223388672, 0.03201228713989258, 0.03213619232177734, 0.03213516616821289, 0.03202969741821289, 0.03199590492248535, 0.03210137557983399, 0.031922176361083986, 0.032301055908203126, 0.03201433563232422, 0.03198054313659668, 0.03193343925476074, 0.032036865234375, 0.03209625625610352, 0.032007167816162106, 0.03206758499145508, 0.032023551940917966, 0.03206860733032227, 0.03202150344848633, 0.03199283218383789, 0.03218431854248047, 0.032007167816162106, 0.03203379058837891, 0.03207167816162109, 0.03210854339599609, 0.032146430969238284, 0.03204710388183594, 0.03199283218383789, 0.03204608154296875, 0.03216486358642578, 0.03210649490356445, 0.03213516616821289, 0.032031742095947266, 0.03211673736572265, 0.03215462493896484, 0.03202969741821289, 0.03196518325805664, 0.032059391021728514, 0.06488473510742188, 0.032023551940917966, 0.03192422485351563, 0.03199897575378418, 0.03199488067626953, 0.03191705513000488, 0.032036865234375, 0.03181977653503418, 0.03184435272216797, 0.03196006393432617, 0.03202867126464844, 0.03201433563232422, 0.03199385643005371, 0.03192422485351563, 0.032020481109619144, 0.032026622772216795, 0.03197849655151367, 0.032054271697998044, 0.03197952079772949, 0.032043006896972655, 0.03210956954956055, 0.032107521057128906, 0.03200511932373047, 0.0320552978515625, 0.03214438247680664, 0.03209523010253906, 0.032145408630371096, 0.03196928024291992, 0.03198975944519043, 0.03209318542480469, 0.03209011077880859, 0.032132095336914065, 0.03204403305053711, 0.03206655883789063, 0.03199692726135254, 0.03212595367431641, 0.032026622772216795, 0.03201331329345703, 0.032080894470214845, 0.03209625625610352, 0.03203891372680664, 0.03206758499145508, 0.03210854339599609, 0.03203071975708008, 0.03220377731323242, 0.03426508712768555, 0.032353279113769534, 0.03223961639404297, 0.03213312149047851, 0.032159744262695314, 0.03218841552734375, 0.03218022537231445, 0.03231027221679687, 0.032031742095947266, 0.03197135925292969, 0.03226620864868164, 0.03202150344848633, 0.03208499145507813, 0.03227545547485351, 0.03205734252929687, 0.032123905181884765, 0.032077823638916016, 0.032130046844482424, 0.06488575744628906, 0.03196928024291992, 0.031887359619140625, 0.031970304489135744, 0.032115711212158206, 0.03201536178588867, 0.031959039688110355, 0.03208294296264649, 0.032271358489990236, 0.032254974365234376, 0.0325662727355957, 0.032036865234375, 0.03214438247680664, 0.032148479461669925, 0.032162849426269534, 0.031946720123291014, 0.03196006393432617, 0.0319180793762207, 0.0319682559967041, 0.03204915237426758, 0.03192934417724609, 0.03199180793762207, 0.03207372665405273, 0.03199488067626953, 0.03211468887329102, 0.03208703994750976, 0.032080894470214845, 0.03203788757324219, 0.032026622772216795, 0.032091136932373046, 0.0319866886138916, 0.03201433563232422, 0.03199283218383789, 0.03202560043334961, 0.03208499145507813, 0.03210342407226562, 0.032048126220703126, 0.03196416091918945, 0.03204095840454101, 0.03218841552734375, 0.0322242546081543, 0.032102401733398435, 0.032010238647460935, 0.032053249359130856, 0.03212595367431641, 0.03209523010253906, 0.032056320190429685, 0.03198361587524414, 0.03203891372680664, 0.03215155029296875, 0.03204403305053711, 0.0320819206237793, 0.03201433563232422, 0.032156673431396485, 0.03220889663696289, 0.032118785858154295, 0.03220684814453125, 0.03203891372680664, 0.032118785858154295, 0.03203379058837891, 0.032118785858154295, 0.032099327087402346, 0.03203481674194336, 0.06507417297363281, 0.03202764892578125, 0.03196928024291992, 0.03220377731323242, 0.032061439514160156, 0.03224883270263672, 0.03297177505493164, 0.03252326583862305, 0.03234918212890625, 0.032121856689453124, 0.032026622772216795, 0.032045055389404296, 0.03210956954956055, 0.03206041717529297, 0.03198873519897461, 0.03208703994750976, 0.03215359878540039, 0.032113662719726564, 0.03207372665405273, 0.03208294296264649, 0.03223756790161133, 0.03221609497070312, 0.03226518249511719, 0.03219046401977539, 0.03213312149047851, 0.03204403305053711, 0.03203379058837891, 0.03311820983886719, 0.032105472564697264, 0.032247806549072264, 0.03196518325805664, 0.032053249359130856, 0.032043006896972655, 0.03206553649902344, 0.03221913528442383, 0.032140289306640625, 0.031958015441894534, 0.03200921630859375, 0.032077823638916016, 0.03201433563232422, 0.03206246566772461, 0.032036865234375, 0.032126976013183595, 0.032299007415771484, 0.032056320190429685, 0.032130046844482424, 0.03199590492248535, 0.03198975944519043, 0.0321638412475586, 0.03199180793762207, 0.032075775146484374, 0.03214438247680664, 0.032026622772216795, 0.0321607666015625, 0.03220889663696289, 0.03221913528442383, 0.03200204849243164, 0.03199385643005371, 0.03206041717529297, 0.03203276824951172, 0.03217919921875, 0.03206655883789063, 0.03205222320556641, 0.06487859344482422, 0.03203891372680664, 0.03199488067626953, 0.031939584732055666, 0.033501182556152344, 0.034561023712158204, 0.03289702224731445, 0.03199283218383789, 0.03206758499145508, 0.03205734252929687, 0.03200614547729492, 0.0321341438293457, 0.03210649490356445, 0.03194675254821777, 0.03201945495605469, 0.03200921630859375, 0.03199897575378418, 0.03207680130004883, 0.032140289306640625, 0.03202969741821289, 0.03214131164550781, 0.0320552978515625, 0.03226828765869141, 0.03204403305053711, 0.03199692726135254, 0.03222630310058594, 0.03198464012145996, 0.032, 0.03209011077880859, 0.032118785858154295, 0.0319498233795166, 0.032161792755126956, 0.031925247192382815, 0.032215038299560544, 0.03199590492248535, 0.032024574279785153, 0.03210137557983399, 0.032505855560302735, 0.03497983932495117, 0.03360768127441406, 0.03216691207885742, 0.03209625625610352, 0.03213516616821289, 0.03210342407226562, 0.03218534469604492, 0.03214233779907227, 0.0327116813659668, 0.032121856689453124, 0.03208806228637695, 0.03220889663696289, 0.0323583984375, 0.03224371337890625, 0.032077823638916016, 0.0321341438293457, 0.03221811294555664, 0.032686080932617184, 0.03235123062133789, 0.03210956954956055, 0.03247923278808594, 0.03230822372436523, 0.032247806549072264, 0.03220787048339844, 0.032107521057128906, 0.0652933120727539, 0.03203276824951172, 0.031871999740600586, 0.031926271438598636, 0.03214131164550781, 0.03189555168151856, 0.031697919845581055, 0.03218636703491211, 0.03216281509399414, 0.03213516616821289, 0.032086017608642575, 0.03211161422729492, 0.03218124771118164, 0.03226726531982422, 0.0321638412475586, 0.03203891372680664, 0.03204710388183594, 0.03205017471313477, 0.032023551940917966, 0.03236044692993164, 0.03211673736572265, 0.03203481674194336, 0.0321945915222168, 0.032360416412353515, 0.03255295944213867, 0.03209830474853516, 0.03205120086669922, 0.03196928024291992, 0.032061439514160156, 0.03204198455810547, 0.031987712860107424, 0.03217407989501953, 0.032043006896972655, 0.03206041717529297, 0.032056320190429685, 0.03204710388183594, 0.03218022537231445, 0.0319866886138916, 0.03201331329345703, 0.032130046844482424, 0.03208396911621094, 0.03199897575378418, 0.032069633483886716, 0.033242111206054685, 0.0324956169128418, 0.0324956169128418, 0.03204198455810547, 0.032077823638916016, 0.03203583908081055, 0.03224166488647461, 0.03228876876831055, 0.03223961639404297, 0.032086048126220706, 0.03266454315185547, 0.03249868774414062, 0.03214233779907227, 0.03209318542480469, 0.03201433563232422, 0.03204608154296875, 0.03210342407226562, 0.03222016143798828, 0.03211468887329102, 0.03254476928710937]",tokens/s,30.642304592162546,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -5900,48 +5819,33 @@ ValueError: FalconForCausalLM does not support an attention implementation throu ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.42.1,,0.31.0,,,,1.20.0,,,,0.11.1,,,MB,1272.561664,1017.643008,0.0,371.195904,315.552256,s,10,0.17266409492492674,0.01726640949249268,0.0007675808488520977,0.016947951316833496,0.017721705245971678,0.018613541030883786,0.019327009658813475,"[0.019505376815795897, 0.016936832427978516, 0.016892736434936523, 0.017105247497558592, 0.016948831558227538, 0.01689910316467285, 0.01690847969055176, 0.01752351951599121, 0.016996896743774415, 0.016947071075439454]",tokens/s,14826.475655596325,kWh,2.060496687104524e-07,1.129046661832092e-07,5.11595453797845e-07,8.305497886915066e-07,tokens/kWh,308229564.9046114,MB,1272.561664,1017.643008,0.0,371.195904,315.554816,s,10,10.474469848632811,1.047446984863281,0.008630898655165174,1.0474097900390626,1.05653818359375,1.0603552734375,1.0634089453125002,"[1.06417236328125, 1.0373424072265625, 1.0374180908203126, 1.0385887451171876, 1.046391357421875, 1.0409013671875, 1.04842822265625, 1.0554637451171875, 1.0500736083984374, 1.05568994140625]",tokens/s,60.14624215871234,kWh,1.275105717653597e-05,6.9871232323215676e-06,2.9252345023997115e-05,4.899052543285465e-05,tokens/kWh,1285962.9375958915,,s,629,10.609158132553102,0.01686670609308919,0.0020826478100592063,0.016492544174194337,0.01711534004211426,0.0173123592376709,0.03376529418945313,"[0.018044927597045898, 0.017146879196166993, 0.017113088607788086, 0.016909311294555664, 0.01741209602355957, 0.017143808364868163, 0.01705779266357422, 0.017091583251953125, 0.01789132881164551, 0.016710655212402344, 0.017185792922973633, 0.01699839973449707, 0.017154048919677735, 0.017137664794921875, 0.016920576095581053, 0.017048576354980468, 0.016884735107421875, 0.017006591796875, 0.017116159439086915, 0.01743769645690918, 0.018395135879516602, 0.01801420783996582, 0.01739468765258789, 0.01720012855529785, 0.017091583251953125, 0.016885759353637696, 0.016760831832885743, 0.017328128814697266, 0.017426431655883787, 0.017473535537719728, 0.01698307228088379, 0.016822240829467774, 0.016887807846069337, 0.016673791885375978, 0.016417823791503906, 0.016383968353271484, 0.016471040725708007, 0.016681983947753908, 0.01655193519592285, 0.016497663497924805, 0.016518144607543944, 0.016408575057983397, 0.01641574478149414, 0.0164270076751709, 0.016439327239990233, 0.016429023742675783, 0.016446464538574217, 0.01641267204284668, 0.01640447998046875, 0.016479232788085937, 0.01642393684387207, 0.01640550422668457, 0.01640550422668457, 0.01638502311706543, 0.01641164779663086, 0.01640345573425293, 0.016447488784790038, 0.01644339179992676, 0.01643929672241211, 0.016630783081054687, 0.01643110466003418, 0.01642393684387207, 0.03379916763305664, 0.01641164779663086, 0.01641574478149414, 0.016420896530151368, 0.016477151870727538, 0.01653555107116699, 0.016623615264892578, 0.016562175750732423, 0.016541696548461913, 0.01641267204284668, 0.01643519973754883, 0.01643110466003418, 0.016397312164306642, 0.016480255126953124, 0.01642188835144043, 0.016451583862304688, 0.016448511123657226, 0.016508928298950197, 0.01641983985900879, 0.01642188835144043, 0.01643519973754883, 0.016417791366577148, 0.01641267204284668, 0.01640447998046875, 0.016474111557006836, 0.016441343307495117, 0.016570367813110352, 0.01648031997680664, 0.016432064056396484, 0.016438272476196288, 0.016416767120361327, 0.016416767120361327, 0.016455680847167968, 0.01644339179992676, 0.016519168853759765, 0.016416767120361327, 0.016457727432250976, 0.0164270076751709, 0.01641574478149414, 0.016365568161010743, 0.016473087310791015, 0.016448511123657226, 0.016550912857055664, 0.016434175491333008, 0.01644441604614258, 0.01646080017089844, 0.016490495681762696, 0.01640243148803711, 0.01641881561279297, 0.01642803192138672, 0.016541696548461913, 0.016527360916137695, 0.016464895248413085, 0.016539648056030275, 0.016506879806518555, 0.016465919494628906, 0.01640550422668457, 0.016491519927978517, 0.01644339179992676, 0.01643110466003418, 0.016487424850463867, 0.016539648056030275, 0.016473087310791015, 0.03371417617797851, 0.016503807067871093, 0.0164771842956543, 0.016470016479492186, 0.01646899223327637, 0.016463872909545898, 0.016446464538574217, 0.0164454402923584, 0.016488447189331054, 0.01643212890625, 0.016579584121704103, 0.016518144607543944, 0.016522239685058594, 0.016476160049438478, 0.016488447189331054, 0.0164771842956543, 0.016481279373168945, 0.016521215438842773, 0.016462848663330077, 0.016474111557006836, 0.016492544174194337, 0.01641983985900879, 0.016450559616088867, 0.016500736236572267, 0.016479232788085937, 0.016440319061279296, 0.01646899223327637, 0.01638297653198242, 0.016430080413818358, 0.016467967987060548, 0.01637887954711914, 0.016446464538574217, 0.01640652847290039, 0.016470016479492186, 0.01639936065673828, 0.01640140724182129, 0.016447488784790038, 0.01642291259765625, 0.01640959930419922, 0.01643724822998047, 0.01641164779663086, 0.01637580871582031, 0.016389120101928712, 0.016472063064575194, 0.01643519973754883, 0.01638707160949707, 0.01644441604614258, 0.016458751678466797, 0.016448511123657226, 0.01643724822998047, 0.016447488784790038, 0.016451583862304688, 0.016476160049438478, 0.016519168853759765, 0.016479232788085937, 0.016507904052734376, 0.016459775924682618, 0.016454656600952147, 0.016476160049438478, 0.016559104919433593, 0.016471040725708007, 0.01655193519592285, 0.016534528732299804, 0.03410124969482422, 0.016536575317382812, 0.016515071868896485, 0.016485376358032225, 0.016470016479492186, 0.01659187126159668, 0.016485376358032225, 0.016447488784790038, 0.016483327865600587, 0.016476160049438478, 0.016464895248413085, 0.016546815872192384, 0.016486400604248046, 0.016496639251708984, 0.016488447189331054, 0.016465919494628906, 0.016507904052734376, 0.016489471435546875, 0.01646080017089844, 0.016476160049438478, 0.016508928298950197, 0.01644339179992676, 0.016475135803222657, 0.016447488784790038, 0.016458751678466797, 0.016450559616088867, 0.01644339179992676, 0.016467967987060548, 0.016504831314086914, 0.016473087310791015, 0.016475135803222657, 0.01640447998046875, 0.01639628791809082, 0.016474111557006836, 0.016563199996948243, 0.01660211181640625, 0.016331775665283203, 0.01636764717102051, 0.016565216064453124, 0.016497663497924805, 0.01642291259765625, 0.016447488784790038, 0.016458751678466797, 0.016501760482788085, 0.016513023376464844, 0.016465919494628906, 0.016482303619384766, 0.016463872909545898, 0.016433151245117187, 0.016491519927978517, 0.016524288177490236, 0.01641164779663086, 0.016477216720581056, 0.01645871925354004, 0.016456703186035156, 0.016447488784790038, 0.0164454402923584, 0.016433151245117187, 0.01660825538635254, 0.016490495681762696, 0.016510976791381835, 0.016484352111816408, 0.016484352111816408, 0.0338155517578125, 0.016489471435546875, 0.016448511123657226, 0.016450559616088867, 0.01646899223327637, 0.016492544174194337, 0.016563199996948243, 0.016493568420410155, 0.016849920272827147, 0.017140735626220704, 0.017802240371704102, 0.017163263320922852, 0.017125375747680666, 0.01714995193481445, 0.016508928298950197, 0.016515071868896485, 0.0164270076751709, 0.016563199996948243, 0.016475135803222657, 0.016446464538574217, 0.016465919494628906, 0.016476160049438478, 0.016479232788085937, 0.016449535369873047, 0.016465919494628906, 0.0165928955078125, 0.016491519927978517, 0.016514047622680664, 0.016529407501220703, 0.016520191192626953, 0.016493568420410155, 0.016480255126953124, 0.016478208541870116, 0.01657548713684082, 0.016433151245117187, 0.016492544174194337, 0.016470016479492186, 0.016504831314086914, 0.0164771842956543, 0.016692224502563476, 0.016459775924682618, 0.016574464797973632, 0.016495616912841796, 0.016503807067871093, 0.016446464538574217, 0.0164771842956543, 0.016455680847167968, 0.01644339179992676, 0.016417791366577148, 0.016545791625976563, 0.016451583862304688, 0.016441343307495117, 0.016448511123657226, 0.016476160049438478, 0.016487424850463867, 0.01703424072265625, 0.017331199645996095, 0.017337343215942384, 0.017126399993896483, 0.016506879806518555, 0.016497663497924805, 0.016466943740844727, 0.016486400604248046, 0.03378278350830078, 0.016478208541870116, 0.016470016479492186, 0.016462848663330077, 0.016499711990356446, 0.016526336669921874, 0.016574464797973632, 0.016481279373168945, 0.016525312423706053, 0.016504831314086914, 0.016474111557006836, 0.016473087310791015, 0.016480255126953124, 0.016523263931274415, 0.016495616912841796, 0.016548864364624022, 0.016459775924682618, 0.016884735107421875, 0.0170250244140625, 0.016882688522338866, 0.016536575317382812, 0.016530431747436524, 0.016450559616088867, 0.016538623809814454, 0.016465919494628906, 0.016454656600952147, 0.016454656600952147, 0.016509952545166014, 0.016750591278076172, 0.01651513671875, 0.016491455078125, 0.016459775924682618, 0.016397312164306642, 0.0164454402923584, 0.016509952545166014, 0.016466943740844727, 0.016454656600952147, 0.016483327865600587, 0.016485376358032225, 0.016458751678466797, 0.016501760482788085, 0.016461824417114256, 0.016459775924682618, 0.01646899223327637, 0.016497663497924805, 0.01646080017089844, 0.016533504486083983, 0.016457727432250976, 0.016490495681762696, 0.016505887985229492, 0.016468959808349608, 0.016545791625976563, 0.0165980167388916, 0.01653353691101074, 0.0164751033782959, 0.016479232788085937, 0.016467967987060548, 0.016507904052734376, 0.01658470344543457, 0.016465919494628906, 0.016486400604248046, 0.016506879806518555, 0.016464895248413085, 0.03372032165527344, 0.016523263931274415, 0.016459775924682618, 0.016480255126953124, 0.01654374313354492, 0.016504831314086914, 0.016498687744140626, 0.01659391975402832, 0.016497663497924805, 0.016447488784790038, 0.016566272735595702, 0.016504831314086914, 0.016497663497924805, 0.016505855560302735, 0.016487424850463867, 0.016534528732299804, 0.01646899223327637, 0.016547840118408205, 0.016464895248413085, 0.016497663497924805, 0.01656831932067871, 0.01646080017089844, 0.016440319061279296, 0.016514047622680664, 0.016447488784790038, 0.016484352111816408, 0.016533504486083983, 0.016571392059326173, 0.016488447189331054, 0.016450559616088867, 0.0164454402923584, 0.016466943740844727, 0.016448511123657226, 0.0164454402923584, 0.016479232788085937, 0.016516096115112306, 0.016849920272827147, 0.016763904571533202, 0.017097728729248047, 0.016528383255004882, 0.016474111557006836, 0.016526336669921874, 0.016483327865600587, 0.016450559616088867, 0.016497663497924805, 0.016440319061279296, 0.01657347106933594, 0.016730079650878905, 0.017118207931518553, 0.017361919403076173, 0.017110015869140623, 0.016995328903198242, 0.017312768936157227, 0.01661235237121582, 0.01661440086364746, 0.01660211181640625, 0.016501760482788085, 0.016514047622680664, 0.017301504135131835, 0.017311744689941407, 0.017344512939453126, 0.017085439682006837, 0.01704652786254883, 0.03498291015625, 0.01660416030883789, 0.016484352111816408, 0.016539648056030275, 0.016464895248413085, 0.016505855560302735, 0.016472063064575194, 0.016533504486083983, 0.01646080017089844, 0.016534528732299804, 0.016520191192626953, 0.016480255126953124, 0.01642803192138672, 0.01683046340942383, 0.01799884796142578, 0.017342464447021484, 0.016716800689697265, 0.016493568420410155, 0.016482303619384766, 0.01643929672241211, 0.016434175491333008, 0.01643622398376465, 0.016492544174194337, 0.016495616912841796, 0.016455680847167968, 0.016524288177490236, 0.016508928298950197, 0.016502784729003905, 0.016611328125, 0.016478208541870116, 0.016517120361328123, 0.016503807067871093, 0.0164771842956543, 0.016915456771850586, 0.017076223373413087, 0.01702809524536133, 0.017115135192871094, 0.01703321647644043, 0.01704243278503418, 0.017071104049682616, 0.017157119750976564, 0.01722777557373047, 0.017131519317626954, 0.017146879196166993, 0.017185792922973633, 0.017120256423950195, 0.017085439682006837, 0.01701068878173828, 0.01699839973449707, 0.01697996711730957, 0.0170199031829834, 0.01704652786254883, 0.016681983947753908, 0.01662668800354004, 0.016473087310791015, 0.016683008193969725, 0.01702092742919922, 0.01702604866027832, 0.016519168853759765, 0.016507904052734376, 0.01644339179992676, 0.016493568420410155, 0.01641881561279297, 0.033909759521484374, 0.016473087310791015, 0.016453632354736326, 0.01658060836791992, 0.016533504486083983, 0.016505855560302735, 0.01662668800354004, 0.016434175491333008, 0.016395263671875, 0.016492544174194337, 0.016561151504516602, 0.016702463150024414, 0.016501760482788085, 0.016482303619384766, 0.016515071868896485, 0.016467967987060548, 0.016466943740844727, 0.016429119110107422, 0.016498624801635744, 0.016440319061279296, 0.016476160049438478, 0.01643110466003418, 0.016664575576782227, 0.016502784729003905, 0.01643622398376465, 0.01641164779663086, 0.0164771842956543, 0.016417791366577148, 0.01646080017089844, 0.016480255126953124, 0.01644339179992676, 0.016461824417114256, 0.01643724822998047, 0.016574464797973632, 0.01643622398376465, 0.016476160049438478, 0.016520191192626953, 0.016473087310791015, 0.016493568420410155, 0.016529407501220703, 0.016494592666625976, 0.017533952713012696, 0.0170700798034668, 0.016489471435546875, 0.017136640548706054, 0.017500160217285156, 0.018059263229370116, 0.017138687133789063, 0.017060863494873048, 0.01720319938659668, 0.017084415435791016, 0.016487424850463867, 0.01654374313354492, 0.016476160049438478, 0.016623615264892578, 0.016487424850463867, 0.01654374313354492, 0.016915456771850586, 0.017087488174438475, 0.01701785659790039, 0.01703424072265625, 0.01700044822692871, 0.01702809524536133, 0.03388518524169922, 0.016950271606445313, 0.017107967376708985, 0.016590848922729492, 0.016510976791381835, 0.017054719924926756, 0.017101823806762697, 0.01706598472595215, 0.016545791625976563, 0.016669696807861328, 0.01657548713684082, 0.01655705642700195, 0.016527360916137695, 0.016513023376464844, 0.016556032180786134, 0.016525312423706053, 0.016498687744140626, 0.016489471435546875, 0.016495616912841796, 0.016548864364624022, 0.016484352111816408, 0.016470016479492186, 0.016471040725708007, 0.016473087310791015, 0.016485376358032225, 0.016454656600952147, 0.01661747169494629, 0.016450559616088867, 0.016463872909545898, 0.016464895248413085, 0.0164454402923584, 0.016505855560302735, 0.016496639251708984, 0.016509952545166014, 0.016450559616088867, 0.0166297607421875, 0.016496639251708984, 0.016539648056030275, 0.016425983428955078, 0.01644339179992676, 0.016450559616088867, 0.016711679458618164, 0.017238016128540038, 0.017059839248657227, 0.017031167984008787, 0.01701273536682129, 0.01719398307800293, 0.01784320068359375, 0.018110464096069336, 0.01717350387573242, 0.017152000427246093, 0.01718169593811035, 0.01724723243713379, 0.017096704483032226, 0.01704243278503418, 0.017192960739135742, 0.01704243278503418, 0.017093631744384767, 0.016695295333862305, 0.016524288177490236, 0.01656729507446289, 0.01699430465698242, 0.016478208541870116]",tokens/s,59.28839896070347,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch - raise ChildProcessError(response[""traceback""]) -ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target - report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,3992.109056,12218.53184,0.0,11572.084736,11067.6992,s,10,6.179316650390626,0.6179316650390625,0.0013633687573851503,0.6172748413085938,0.6201095886230469,0.6201251068115234,0.6201375213623047,"[0.6168117065429688, 0.6166927490234375, 0.617082763671875, 0.6173292846679688, 0.6172203979492188, 0.6164153442382813, 0.6180755615234375, 0.6194420776367188, 0.620140625, 0.6201061401367187]",tokens/s,414.28529153594513,kWh,7.284492750962576e-06,3.991580694684457e-06,3.91370084337663e-05,5.041308187941334e-05,tokens/kWh,5078047.015898467,MB,3992.109056,12218.53184,0.0,11572.084736,11417.442816,s,10,337.7582734375,33.77582734375,0.0018514865161095365,33.776255859375,33.777652734375,33.7777775390625,33.7778773828125,"[33.775578125, 33.77729296875, 33.77443359375, 33.777484375, 33.77612109375, 33.77790234375, 33.7730078125, 33.7724375, 33.776390625, 33.777625]",tokens/s,1.8652392836694713,kWh,0.00039873445866836445,0.00021854115969749565,0.002154927330149635,0.002772202948515495,tokens/kWh,22725.60890022005,,s,629,342.78417449951223,0.544966891096203,0.07347847726131472,0.5360609130859375,0.5363870361328125,0.5365364990234375,1.1536408984375002,"[0.5359658203125, 0.5360260009765625, 0.5359830932617188, 0.5357762451171875, 0.5358079833984375, 0.5357578125, 0.536057861328125, 0.5358643188476563, 0.5358817138671875, 0.5360762939453125, 0.5358797607421875, 0.5359031982421875, 0.5359114379882812, 0.535762939453125, 0.5361029052734375, 0.5359216918945312, 0.5361571655273437, 0.5362728881835938, 0.5359862060546875, 0.5363558349609375, 0.53625341796875, 0.5360670776367188, 0.5360250854492188, 0.5361551513671875, 0.5358868408203125, 0.5363343505859375, 0.53604248046875, 0.53614599609375, 0.5357413330078125, 0.5361039428710938, 0.5360609130859375, 0.535994384765625, 0.5358919677734375, 0.5362943725585938, 0.5358212890625, 0.5360732421875, 0.5360598754882813, 0.5360128173828125, 0.536158203125, 0.5361387329101562, 0.53642138671875, 0.5362493286132812, 0.536437744140625, 0.5364603881835938, 0.5359768676757812, 0.5361715698242188, 0.5362175903320312, 0.5359380493164062, 0.53612646484375, 0.53635888671875, 0.5363887939453125, 0.5363740844726562, 0.5362288818359375, 0.5360537719726562, 0.5364336547851563, 0.53607421875, 0.5363271484375, 0.5363067016601563, 0.5362882690429688, 0.5363281860351562, 0.5362197265625, 0.536455078125, 1.1534130859375, 0.5356964111328125, 0.5358591918945312, 0.5360977783203125, 0.5360404663085937, 0.5358120727539063, 0.5357598876953125, 0.5360036010742187, 0.535930908203125, 0.5361203002929688, 0.5360250854492188, 0.5360271606445313, 0.5363189697265625, 0.5360609130859375, 0.5358591918945312, 0.5361387939453125, 0.5359646606445313, 0.53595458984375, 0.5362307739257812, 0.535994384765625, 0.5361828002929687, 0.5358101806640625, 0.535956298828125, 0.536256591796875, 0.536070068359375, 0.5358377075195313, 0.536141845703125, 0.5360025634765625, 0.536183837890625, 0.5360322265625, 0.5363148803710938, 0.5361500244140625, 0.5360394287109375, 0.5360117797851562, 0.5362032470703125, 0.5359974365234375, 0.536437744140625, 0.5358848266601562, 0.5359247436523438, 0.5360087280273438, 0.5358848266601562, 0.5360732421875, 0.5361243896484374, 0.5362575073242187, 0.5362042846679688, 0.53837109375, 0.5366231079101562, 0.5364879150390625, 0.536089599609375, 0.5362257690429687, 0.5363108520507812, 0.5363219604492188, 0.53629541015625, 0.536052734375, 0.5360117797851562, 0.5363743896484375, 0.5360311279296875, 0.5361807250976562, 0.5362575073242187, 0.5364439086914062, 0.5363804931640626, 0.5361827392578125, 0.5361879272460938, 1.1537294921875, 0.5356585083007812, 0.5357578125, 0.5362330932617188, 0.5359236450195313, 0.5359400024414063, 0.5355130615234375, 0.5358325805664063, 0.5358253784179687, 0.5359175415039062, 0.5359646606445313, 0.5358131103515625, 0.5359237060546875, 0.5357967529296875, 0.53574755859375, 0.536037353515625, 0.535878662109375, 0.5359575805664063, 0.5363414306640625, 0.5359759521484375, 0.5363261108398437, 0.535931884765625, 0.5358356323242187, 0.5376511840820313, 0.5358756103515625, 0.5360435180664063, 0.5359667358398438, 0.5363836669921875, 0.5364918823242187, 0.5357455444335938, 0.5357609252929687, 0.5358714599609375, 0.5358612670898437, 0.5359185791015625, 0.5364593505859375, 0.53573828125, 0.5363804321289063, 0.5361234130859375, 0.5359493408203125, 0.536099853515625, 0.5359144897460938, 0.5359063110351563, 0.5361397705078125, 0.5362360229492188, 0.536052734375, 0.53610595703125, 0.5360025634765625, 0.5383301391601563, 0.53597900390625, 0.5362012329101562, 0.5361643676757812, 0.5362411499023437, 0.5361654052734375, 0.5362390747070312, 0.5361377563476563, 0.5363753662109375, 0.536123291015625, 0.5360568237304687, 0.5363026123046875, 0.5362759399414062, 0.5363353881835937, 0.5361090698242188, 0.5360189208984375, 1.1538984375, 0.5357557983398438, 0.5358960571289062, 0.5358387451171875, 0.5358746948242188, 0.5359072265625, 0.5356656494140625, 0.5360302124023437, 0.5358325805664063, 0.5361961059570313, 0.5362575073242187, 0.536289306640625, 0.5370880126953125, 0.5360404663085937, 0.5358018798828125, 0.5358909301757813, 0.5359216918945312, 0.5360066528320313, 0.5361715087890625, 0.5361909790039062, 0.5363128051757813, 0.5360137939453125, 0.53574755859375, 0.5359697875976562, 0.5360650634765625, 0.5359359130859375, 0.5359750366210938, 0.53585498046875, 0.536374267578125, 0.5358345947265625, 0.5361674194335937, 0.536142822265625, 0.5358745727539063, 0.5359216918945312, 0.5363722534179688, 0.5359923095703125, 0.536289306640625, 0.5359442138671875, 0.5359288330078125, 0.5362186889648437, 0.535847900390625, 0.5360865478515625, 0.5361991577148437, 0.5361500244140625, 0.5386311645507813, 0.5361819458007813, 0.536072021484375, 0.5363650512695313, 0.5362186279296876, 0.5365278930664062, 0.5365155639648438, 0.53625244140625, 0.5360865478515625, 0.5360722045898437, 0.5359431762695313, 0.536431640625, 0.5360281372070312, 0.5362841796875, 0.5362257690429687, 0.536131591796875, 0.5363865966796875, 0.536090576171875, 0.5364408569335938, 1.15415966796875, 0.5359022216796875, 0.5359144897460938, 0.53595751953125, 0.5359749145507813, 0.5358868408203125, 0.535404541015625, 0.53585205078125, 0.5359554443359374, 0.5361510620117188, 0.536099853515625, 0.5363251342773437, 0.5363619995117187, 0.5359820556640625, 0.5358171997070312, 0.5359042358398437, 0.535952392578125, 0.5376307373046875, 0.5360568237304687, 0.5358981323242188, 0.5363967895507813, 0.53595458984375, 0.5358468017578125, 0.53597900390625, 0.5359411010742188, 0.5358264770507812, 0.5360383911132812, 0.535878662109375, 0.5361961059570313, 0.5358909301757813, 0.5359595336914063, 0.5360497436523437, 0.5359052124023438, 0.5358837890625, 0.5363169555664062, 0.5359349975585938, 0.5359779663085937, 0.5358458862304688, 0.535952392578125, 0.5360803833007812, 0.535984130859375, 0.5360537719726562, 0.5359790649414062, 0.5361837158203125, 0.5359412231445313, 0.5360147705078125, 0.5358458862304688, 0.5359851684570313, 0.5362432250976562, 0.5363035888671875, 0.5361275024414063, 0.5362432250976562, 0.5360271606445313, 0.53608447265625, 0.5359513549804688, 0.536268798828125, 0.535962646484375, 0.5394749145507812, 0.5364561767578125, 0.5365913696289063, 0.5363251342773437, 0.536248291015625, 0.5362032470703125, 1.153287109375, 0.535741455078125, 0.5358274536132812, 0.536648681640625, 0.536184814453125, 0.5359554443359374, 0.5358244018554688, 0.535973876953125, 0.53565234375, 0.5358510131835937, 0.5360148315429687, 0.5357537231445313, 0.5359800415039062, 0.536056884765625, 0.5358929443359375, 0.5360916748046874, 0.5358028564453124, 0.5358479614257813, 0.5360568237304687, 0.5358387451171875, 0.5360291748046875, 0.5361879272460938, 0.5361356811523438, 0.5360722045898437, 0.536173583984375, 0.535920654296875, 0.5360414428710938, 0.535952392578125, 0.5360814208984375, 0.53586328125, 0.5363804321289063, 0.5363519897460938, 0.5361846313476563, 0.5362749633789062, 0.5363671264648437, 0.53599951171875, 0.5362022705078126, 0.5360322265625, 0.5388390502929687, 0.5362084350585937, 0.53614794921875, 0.5361172485351563, 0.5362667236328125, 0.53610595703125, 0.5361387329101562, 0.536110107421875, 0.5359862060546875, 0.536263671875, 0.5363753051757812, 0.5363077392578125, 0.5363046264648438, 0.5363292236328125, 0.5362575073242187, 0.5360609130859375, 0.5361285400390625, 0.5363517456054687, 0.5360169067382813, 0.5362207641601563, 0.5361736450195312, 0.5364979858398438, 0.536511474609375, 0.53607421875, 0.536079345703125, 1.15462451171875, 0.5355130615234375, 0.5357547607421875, 0.5357742309570312, 0.5357783203125, 0.5361571655273437, 0.5357271118164062, 0.5361858520507813, 0.5361121215820313, 0.5359820556640625, 0.535794677734375, 0.5359134521484376, 0.5360670776367188, 0.5358827514648438, 0.5358397216796875, 0.5360496826171876, 0.53564111328125, 0.5359298706054687, 0.536079345703125, 0.5356728515625, 0.5358930053710937, 0.535846923828125, 0.5357905883789063, 0.53806591796875, 0.5360537719726562, 0.5360701293945312, 0.5361746215820312, 0.536268798828125, 0.5364520874023437, 0.5355530395507813, 0.5357025146484375, 0.5359697875976562, 0.535841796875, 0.5360537719726562, 0.5362442016601563, 0.5358919677734375, 0.5360272216796875, 0.5358335571289062, 0.5359073486328125, 0.5360445556640625, 0.5359134521484376, 0.5360814208984375, 0.5361162109375, 0.5362063598632812, 0.5361193237304688, 0.5360322265625, 0.5362032470703125, 0.5360588989257813, 0.5359401245117188, 0.5361489868164062, 0.5363446044921875, 0.5365084228515625, 0.5364049682617188, 0.5359832153320313, 0.5360147094726563, 0.5362421875, 0.5362145385742187, 0.5363138427734375, 0.5365606689453125, 0.536320068359375, 0.536520751953125, 0.5364028930664062, 0.5362575073242187, 1.1561614990234375, 0.5356492919921875, 0.5358203125, 0.5358499755859375, 0.5358787231445312, 0.5363731689453125, 0.5356687622070313, 0.535857177734375, 0.5358069458007813, 0.5358602294921875, 0.535900146484375, 0.5358806762695313, 0.53600048828125, 0.5359329223632813, 0.5359974365234375, 0.5360414428710938, 0.536551513671875, 0.5360792846679687, 0.5361510620117188, 0.5359216918945312, 0.5360773315429688, 0.5359267578125, 0.535736328125, 0.53596875, 0.5359646606445313, 0.5360250854492188, 0.5358765869140625, 0.5357721557617188, 0.536036376953125, 0.5356483154296875, 0.535942138671875, 0.5359749145507813, 0.53587353515625, 0.5358593139648438, 0.536328125, 0.5358448486328125, 0.5362452392578125, 0.5360291748046875, 0.5359083251953125, 0.536192138671875, 0.5358662109375, 0.5360487060546875, 0.5363189086914063, 0.5367244873046875, 0.53640087890625, 0.5360160522460937, 0.5360556030273438, 0.5362503662109375, 0.5359749145507813, 0.5363671264648437, 0.5362462768554688, 0.5367183227539063, 0.5365473022460937, 0.5361336059570313, 0.5361151733398437, 0.5364818115234375, 0.5360332641601563, 0.5359431762695313, 0.536322021484375, 0.536247314453125, 0.5361448974609375, 0.536072265625, 0.53621142578125, 1.1571527099609376, 0.53568408203125, 0.5358028564453124, 0.5358428344726562, 0.5359677734375, 0.5360455932617187, 0.53608447265625, 0.5359380493164062, 0.5360446166992188, 0.53616943359375, 0.536326171875, 0.5360281372070312, 0.536005615234375, 0.535910400390625, 0.5358336181640625, 0.5360783081054687, 0.5360752563476563, 0.5370399169921874, 0.5360987548828124, 0.5359862060546875, 0.5362216796875, 0.53591552734375, 0.5358817138671875, 0.5361991577148437, 0.5363834838867187, 0.5362493286132812, 0.536322021484375, 0.5362063598632812, 0.5362421875, 0.535826416015625, 0.5361561889648437, 0.53603125, 0.5362677612304687, 0.5359451904296875, 0.5363026123046875, 0.5362135009765625, 0.5364592895507813, 0.535962646484375, 0.5360015258789063, 0.5361715698242188, 0.5358529663085938, 0.5358407592773438, 0.5359871826171875, 0.5362933959960937, 0.536326171875, 0.536173583984375, 0.5360372924804687, 0.5361603393554687, 0.5360833740234375, 0.5362227172851562, 0.536310791015625, 0.5365032958984375, 0.536352783203125, 0.5360598754882813, 0.536079345703125, 0.5365145874023437, 0.5358397216796875, 0.5360926513671875, 0.536224853515625, 0.535985107421875, 0.5362647094726563, 0.53623193359375, 0.536079345703125, 1.1569970703125, 0.5358653564453125, 0.53596875, 0.5359288330078125, 0.5359646606445313, 0.5359473266601562, 0.5357546997070313, 0.5363394775390625, 0.5362810668945313, 0.5360557861328125, 0.5361285400390625, 0.535868408203125, 0.5360045776367187, 0.5358489379882813, 0.5356820678710937, 0.536005615234375, 0.5358970947265626, 0.5360179443359375, 0.5362861938476563, 0.5363015747070312, 0.5365145874023437, 0.5360169067382813, 0.5357752075195312, 0.5362216796875, 0.536158203125, 0.5358131103515625, 0.5376022338867188, 0.536208251953125, 0.536542236328125, 0.5358131103515625, 0.5360281372070312, 0.5361243896484374, 0.5358212890625, 0.5360148315429687, 0.536268798828125, 0.5359216918945312, 0.5361305541992187, 0.5359134521484376, 0.536026123046875, 0.536110107421875, 0.53586328125, 0.5362237548828125, 0.5360230102539062, 0.5360865478515625, 0.5360476684570312, 0.5360148315429687, 0.5359800415039062, 0.5362421875, 0.5360383911132812, 0.5362329711914062, 0.5365883178710937, 0.5367234497070312, 0.5361285400390625, 0.5362032470703125, 0.5362831420898437, 0.5368237915039062, 0.5363056640625, 0.5362196655273438, 0.53631591796875, 0.5363343505859375, 0.5364572143554688, 0.5363097534179687, 0.536158203125]",tokens/s,1.8349738605009493,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,1712.406528,5964.824576,0.0,5318.377472,5048.328704,s,10,3.4079426879882817,0.34079426879882807,0.0009322383785557938,0.3405218048095703,0.3422028137207031,0.34224659423828124,0.34228161865234374,"[0.33942413330078125, 0.33999859619140627, 0.3406260070800781, 0.3401291809082031, 0.3404176025390625, 0.3401330871582031, 0.34098886108398435, 0.34174176025390623, 0.3421930847167969, 0.3422903747558594]",tokens/s,751.1863415494162,kWh,4.007971358520013e-06,2.195722103369917e-06,2.097056307273452e-05,2.717425653462445e-05,tokens/kWh,9420680.918126099,MB,1712.406528,5964.824576,0.0,5318.377472,5098.333184,s,10,188.29865429687501,18.8298654296875,0.0006673439432085487,18.82976953125,18.8306990234375,18.83094423828125,18.831140410156248,"[18.829521484375, 18.829875, 18.82865625, 18.829970703125, 18.8296640625, 18.83021484375, 18.83064453125, 18.829525390625, 18.831189453125, 18.829392578125]",tokens/s,3.3457488177622916,kWh,0.0002223270491041519,0.00012185299643231701,0.0011337806940608672,0.0014779607395973358,tokens/kWh,42626.30143826695,,s,629,191.06832772827136,0.3037652269129912,0.04052678174686376,0.29887591552734377,0.29901171264648435,0.2990915771484375,0.6395446215820312,"[0.29933056640625, 0.2987888488769531, 0.2987939758300781, 0.2988922729492188, 0.29877658081054687, 0.2988226623535156, 0.2988328857421875, 0.29880218505859374, 0.29885440063476565, 0.2987530212402344, 0.2989803466796875, 0.298977294921875, 0.2988349304199219, 0.2987745361328125, 0.29902029418945314, 0.29885336303710935, 0.2988748779296875, 0.2988410949707031, 0.29882571411132813, 0.2988380126953125, 0.2986956787109375, 0.2987386779785156, 0.29879092407226565, 0.29889739990234376, 0.298745849609375, 0.29895782470703125, 0.2990745544433594, 0.2989486083984375, 0.2988134460449219, 0.2987888488769531, 0.29889535522460936, 0.2988922729492188, 0.29888717651367186, 0.298935302734375, 0.29891070556640625, 0.29895269775390626, 0.2988277893066406, 0.2988584899902344, 0.29889535522460936, 0.29889739990234376, 0.2988717956542969, 0.2989383544921875, 0.2988564453125, 0.29891787719726565, 0.29884210205078127, 0.2988062744140625, 0.2989690856933594, 0.29877862548828127, 0.29881036376953124, 0.29896807861328123, 0.29892098999023436, 0.29895574951171877, 0.29893939208984377, 0.2987796630859375, 0.29892095947265623, 0.2987734985351562, 0.29884722900390626, 0.2990111083984375, 0.2989178466796875, 0.2988789672851562, 0.2989742431640625, 0.2987898559570312, 0.6390538330078125, 0.29874993896484375, 0.2987274169921875, 0.2988769226074219, 0.2988164978027344, 0.29890048217773435, 0.29877658081054687, 0.29891278076171873, 0.2988912658691406, 0.29923736572265625, 0.29886669921875, 0.29883392333984377, 0.29876019287109373, 0.29876223754882814, 0.2988769226074219, 0.29878170776367186, 0.2988779602050781, 0.2988216247558594, 0.2988441467285156, 0.29885952758789064, 0.2987755432128906, 0.29880935668945313, 0.29883392333984377, 0.2988554382324219, 0.29924658203125, 0.2988380126953125, 0.29887387084960937, 0.2988840942382813, 0.2993029174804688, 0.2988584899902344, 0.29871514892578127, 0.2991124572753906, 0.298829833984375, 0.29886053466796875, 0.29897317504882814, 0.2988963928222656, 0.29880523681640625, 0.29894247436523436, 0.29871514892578127, 0.29875201416015623, 0.29894451904296876, 0.29874789428710935, 0.29877658081054687, 0.29893017578125, 0.29889434814453125, 0.2990643310546875, 0.29894451904296876, 0.29872845458984376, 0.2989383544921875, 0.29896295166015624, 0.29883187866210936, 0.2988687438964844, 0.29901412963867186, 0.29885748291015624, 0.2988984375, 0.2988963928222656, 0.29887387084960937, 0.29880935668945313, 0.29898651123046877, 0.2988912658691406, 0.2989486083984375, 0.29897830200195313, 0.2988584899902344, 0.6398330688476562, 0.298903564453125, 0.29880218505859374, 0.2988216247558594, 0.2987315063476563, 0.298777587890625, 0.29872332763671877, 0.2986977233886719, 0.2987960205078125, 0.29879092407226565, 0.298724365234375, 0.29893121337890627, 0.2992732238769531, 0.298814453125, 0.29890866088867185, 0.2988011474609375, 0.2988769226074219, 0.2987274169921875, 0.2989219970703125, 0.29873458862304686, 0.29885235595703125, 0.2989690856933594, 0.2988492736816406, 0.2988031921386719, 0.298861572265625, 0.2988554382324219, 0.2988062744140625, 0.2989547424316406, 0.29884622192382815, 0.2987438049316406, 0.2989117431640625, 0.2988840942382813, 0.29899468994140627, 0.2988564453125, 0.298814453125, 0.298819580078125, 0.29902542114257813, 0.2989270935058594, 0.2990540771484375, 0.2987673645019531, 0.298851318359375, 0.29884518432617185, 0.29885235595703125, 0.2987008056640625, 0.29885336303710935, 0.29878170776367186, 0.2989045715332031, 0.2989260864257812, 0.2990172119140625, 0.298829833984375, 0.29888101196289063, 0.2989363098144531, 0.2988584899902344, 0.2988636169433594, 0.29881753540039063, 0.2989066162109375, 0.29882571411132813, 0.29906329345703125, 0.29887387084960937, 0.2989240417480469, 0.2989045715332031, 0.29904177856445313, 0.29881857299804687, 0.6396661987304687, 0.29890866088867185, 0.29881549072265623, 0.29886465454101563, 0.29882366943359373, 0.2986588134765625, 0.2989281311035156, 0.29879806518554686, 0.2988359680175781, 0.29883905029296876, 0.2988031921386719, 0.2987530212402344, 0.2989854736328125, 0.299030517578125, 0.2989547424316406, 0.29914111328125, 0.2987806701660156, 0.29885952758789064, 0.298814453125, 0.2987591552734375, 0.2988308410644531, 0.29881857299804687, 0.2988554382324219, 0.2988011474609375, 0.2988554382324219, 0.29885440063476565, 0.29888204956054687, 0.29879910278320315, 0.2987745361328125, 0.29876837158203123, 0.29885235595703125, 0.29886770629882814, 0.2988062744140625, 0.29895065307617186, 0.29894552612304687, 0.29887591552734377, 0.29902847290039064, 0.298745849609375, 0.29884722900390626, 0.29874176025390625, 0.29884005737304686, 0.29894247436523436, 0.29886257934570315, 0.2987888488769531, 0.2990315551757812, 0.29888717651367186, 0.29897216796875, 0.29883905029296876, 0.2989117431640625, 0.2990663757324219, 0.2989895629882813, 0.29887591552734377, 0.298925048828125, 0.2989056091308594, 0.29893939208984377, 0.29892095947265623, 0.29896395874023435, 0.2988687438964844, 0.2989609069824219, 0.29890765380859374, 0.2990243835449219, 0.2988287963867188, 0.29879092407226565, 0.6396928100585938, 0.29893939208984377, 0.29887283325195313, 0.29885440063476565, 0.2989383544921875, 0.29895269775390626, 0.2988441467285156, 0.2988482666015625, 0.2989619140625, 0.29886257934570315, 0.29881549072265623, 0.2989117431640625, 0.2988072814941406, 0.29882571411132813, 0.2989240417480469, 0.29879296875, 0.2987960205078125, 0.29870181274414065, 0.29926913452148435, 0.29865475463867186, 0.2987703857421875, 0.29870181274414065, 0.2988308410644531, 0.29877658081054687, 0.2987857971191406, 0.29879193115234376, 0.298919921875, 0.2987591552734375, 0.2987806701660156, 0.298919921875, 0.2987694091796875, 0.29903359985351563, 0.29880523681640625, 0.2990408020019531, 0.2988523254394531, 0.29895782470703125, 0.2988554382324219, 0.2988769226074219, 0.2992046203613281, 0.298925048828125, 0.29894247436523436, 0.29885336303710935, 0.2989486083984375, 0.29879296875, 0.29882571411132813, 0.2989240417480469, 0.29892913818359373, 0.29900799560546876, 0.29885952758789064, 0.29900390625, 0.29886257934570315, 0.2988062744140625, 0.29906842041015624, 0.29885336303710935, 0.29888101196289063, 0.2989056091308594, 0.2989219970703125, 0.2988503112792969, 0.2989137878417969, 0.29884005737304686, 0.2988789672851562, 0.29902029418945314, 0.29878988647460936, 0.6392319946289062, 0.29863629150390625, 0.2988984375, 0.29874688720703124, 0.29883392333984377, 0.2988584899902344, 0.29881857299804687, 0.29880218505859374, 0.29896600341796875, 0.29875506591796874, 0.2988287963867188, 0.2988226623535156, 0.2989434814453125, 0.29880218505859374, 0.2989854736328125, 0.2987653198242187, 0.29884005737304686, 0.29883187866210936, 0.2989383544921875, 0.29896295166015624, 0.298893310546875, 0.2992261047363281, 0.29879296875, 0.29881036376953124, 0.2989045715332031, 0.29901412963867186, 0.2989537353515625, 0.29883905029296876, 0.2987694091796875, 0.2988349304199219, 0.29899468994140627, 0.2988861389160156, 0.29891787719726565, 0.29884518432617185, 0.2990172119140625, 0.2990745544433594, 0.29899264526367186, 0.29902029418945314, 0.29891278076171873, 0.2989342651367187, 0.2989270935058594, 0.2989557800292969, 0.2988441467285156, 0.29887081909179686, 0.2988656311035156, 0.29873358154296875, 0.29878680419921877, 0.2989465637207031, 0.2988912658691406, 0.298819580078125, 0.29887283325195313, 0.29891583251953124, 0.29893121337890627, 0.29889022827148437, 0.2988380126953125, 0.299009033203125, 0.2988410949707031, 0.2991124572753906, 0.29884622192382815, 0.2988584899902344, 0.29896395874023435, 0.29894451904296876, 0.29879705810546875, 0.6401351928710938, 0.2988328857421875, 0.2987581481933594, 0.2988840942382813, 0.29872024536132813, 0.29893939208984377, 0.2988707885742187, 0.29880523681640625, 0.2989363098144531, 0.2989342651367187, 0.2987714538574219, 0.29879092407226565, 0.2989823913574219, 0.29887387084960937, 0.2987889099121094, 0.2989301452636719, 0.2989189147949219, 0.2988789672851562, 0.2987530212402344, 0.29921383666992185, 0.2988380126953125, 0.2988083190917969, 0.29881753540039063, 0.2987274169921875, 0.2992404479980469, 0.298977294921875, 0.2987080078125, 0.2987703857421875, 0.2988779602050781, 0.2988001403808594, 0.29884722900390626, 0.2989322509765625, 0.2988431091308594, 0.2989875183105469, 0.29900698852539065, 0.298787841796875, 0.2988380126953125, 0.2987663269042969, 0.29904177856445313, 0.2989045715332031, 0.2989270935058594, 0.2989854736328125, 0.2989793395996094, 0.2989168701171875, 0.29900286865234377, 0.2988134460449219, 0.2988492736816406, 0.2989066162109375, 0.2990909729003906, 0.29900079345703123, 0.2989189147949219, 0.2989056091308594, 0.29892913818359373, 0.29894964599609375, 0.2989598693847656, 0.29893017578125, 0.29894659423828124, 0.2988861083984375, 0.29896295166015624, 0.29888204956054687, 0.2989690856933594, 0.2989066162109375, 0.2988011474609375, 0.64087451171875, 0.2988800048828125, 0.2988359680175781, 0.29870693969726564, 0.2989414367675781, 0.2989752197265625, 0.29876223754882814, 0.2988226623535156, 0.29887591552734377, 0.2989363098144531, 0.298777587890625, 0.29906228637695315, 0.2988083190917969, 0.29909197998046877, 0.2988851318359375, 0.29884210205078127, 0.29921075439453126, 0.2988840942382813, 0.2988717956542969, 0.2987806701660156, 0.29890252685546875, 0.2987325439453125, 0.2989486083984375, 0.29873358154296875, 0.2988717956542969, 0.29869158935546875, 0.29887591552734377, 0.2990520324707031, 0.2989014892578125, 0.2989066162109375, 0.2988083190917969, 0.2989014892578125, 0.29879705810546875, 0.29891278076171873, 0.29888101196289063, 0.2989956970214844, 0.29885440063476565, 0.29872845458984376, 0.29881036376953124, 0.29893939208984377, 0.2989168701171875, 0.29879092407226565, 0.2989588623046875, 0.2989014892578125, 0.29880218505859374, 0.29904794311523436, 0.2987724914550781, 0.29912576293945314, 0.29889022827148437, 0.2988779602050781, 0.2988830871582031, 0.2989260864257812, 0.2988697509765625, 0.29885235595703125, 0.2987653198242187, 0.29883392333984377, 0.29885952758789064, 0.29876837158203123, 0.29888922119140626, 0.298998779296875, 0.2989137878417969, 0.2988769226074219, 0.2987448425292969, 0.6411837158203125, 0.2987673645019531, 0.2987755432128906, 0.2987161560058594, 0.2988308410644531, 0.2989434814453125, 0.29867929077148436, 0.2986700744628906, 0.29879296875, 0.29884622192382815, 0.29911654663085935, 0.2989598693847656, 0.2990049133300781, 0.2988994445800781, 0.2992814025878906, 0.29890252685546875, 0.2989742126464844, 0.29876019287109373, 0.29890252685546875, 0.2987694091796875, 0.29888922119140626, 0.2989137878417969, 0.29885440063476565, 0.2989066162109375, 0.2988206176757813, 0.29879296875, 0.2987888488769531, 0.29874893188476564, 0.298829833984375, 0.29893734741210937, 0.29940017700195315, 0.29889434814453125, 0.298829833984375, 0.2988830871582031, 0.2988994445800781, 0.2990162048339844, 0.2989823913574219, 0.29884622192382815, 0.29896395874023435, 0.2990172119140625, 0.29889022827148437, 0.2988503112792969, 0.298745849609375, 0.2989895629882813, 0.2990100402832031, 0.29893734741210937, 0.2988861389160156, 0.29892913818359373, 0.29891583251953124, 0.29912371826171874, 0.298935302734375, 0.29889739990234376, 0.29890969848632815, 0.2988707885742187, 0.2987857971191406, 0.298977294921875, 0.29891278076171873, 0.2990110778808594, 0.2991124572753906, 0.2989363098144531, 0.2988584899902344, 0.2989465637207031, 0.29895065307617186, 0.641522705078125, 0.298787841796875, 0.29876837158203123, 0.29884722900390626, 0.29865164184570314, 0.2988779602050781, 0.29880117797851563, 0.298661865234375, 0.29881036376953124, 0.2990540771484375, 0.29873663330078126, 0.2987796630859375, 0.2988769226074219, 0.2989619140625, 0.2989916076660156, 0.298925048828125, 0.29885235595703125, 0.298861572265625, 0.29878680419921877, 0.29885235595703125, 0.2988216247558594, 0.2989045715332031, 0.29874893188476564, 0.29875711059570315, 0.29881036376953124, 0.29895269775390626, 0.29883392333984377, 0.29884518432617185, 0.2988687438964844, 0.2988963928222656, 0.29879092407226565, 0.29900390625, 0.29879705810546875, 0.2992803955078125, 0.298956787109375, 0.29879910278320315, 0.29885336303710935, 0.29882366943359373, 0.298935302734375, 0.2987694091796875, 0.298777587890625, 0.29879092407226565, 0.29875201416015623, 0.29888204956054687, 0.29894757080078127, 0.29877862548828127, 0.29879705810546875, 0.29888922119140626, 0.2988441467285156, 0.29900698852539065, 0.29907659912109374, 0.2989619140625, 0.29929470825195315, 0.2988861389160156, 0.2989270935058594, 0.29891278076171873, 0.29884210205078127, 0.2989056091308594, 0.29896600341796875, 0.2988984375, 0.29889022827148437, 0.29891583251953124, 0.2989547424316406]",tokens/s,3.2920160420021825,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,MB,1824.84992,7019.692032,0.0,6371.147776,6308.565504,s,10,1.2993753967285155,0.12993753967285154,0.00011286455710600496,0.12993553924560547,0.13006795806884766,0.13008922348022461,0.13010623580932618,"[0.12990701293945311, 0.12985215759277344, 0.12982847595214844, 0.129797119140625, 0.12978787231445313, 0.12996406555175782, 0.13005126953125, 0.13011048889160157, 0.130063232421875, 0.13001370239257812]",tokens/s,1970.177368638351,kWh,1.5363306787247698e-06,8.418326263070598e-07,6.204391687896046e-06,8.582554992927875e-06,tokens/kWh,29827947.529721275,MB,1841.987584,7021.789184,0.0,6373.244928,6308.568064,s,10,18.549695434570314,1.8549695434570315,0.02668907242410975,1.84841796875,1.8910526123046876,1.89603271484375,1.900016796875,"[1.824569580078125, 1.82419775390625, 1.8338994140625, 1.8366678466796875, 1.8768740234375, 1.8899459228515625, 1.833943115234375, 1.8684168701171875, 1.9010128173828125, 1.8601680908203124]",tokens/s,33.96282177366074,kWh,2.1637239003497494e-05,1.1855319487961205e-05,6.853525143710449e-05,0.00010202780992856319,tokens/kWh,617478.7055030458,,s,630,18.547763187408467,0.02944089394826738,0.0007386781089137757,0.029053440093994142,0.03056373748779297,0.030708736419677734,0.031011831760406495,"[0.029435903549194335, 0.028799999237060548, 0.028675071716308592, 0.0287774715423584, 0.02910310363769531, 0.028845056533813477, 0.02919628715515137, 0.028859392166137695, 0.028884992599487305, 0.029048831939697265, 0.029411327362060546, 0.028957696914672853, 0.028802047729492186, 0.02902016067504883, 0.029060096740722657, 0.029015039443969725, 0.02879795265197754, 0.028836864471435547, 0.028864511489868162, 0.02894438362121582, 0.028876800537109375, 0.028848127365112306, 0.028787712097167968, 0.02878463935852051, 0.028892160415649414, 0.0290119686126709, 0.02894438362121582, 0.02873036766052246, 0.028632064819335938, 0.028734464645385743, 0.029022207260131837, 0.028778495788574218, 0.028887039184570314, 0.028841983795166014, 0.028770303726196288, 0.028872703552246092, 0.02953625679016113, 0.028918783187866212, 0.02892185592651367, 0.02893721580505371, 0.02898841667175293, 0.029005823135375978, 0.028895231246948243, 0.028829696655273438, 0.029121536254882813, 0.029520896911621092, 0.029315071105957033, 0.029470720291137696, 0.029352960586547853, 0.029054975509643553, 0.028767232894897462, 0.028882944107055664, 0.028900352478027344, 0.02878873634338379, 0.028916736602783204, 0.028697599411010744, 0.029105152130126953, 0.029049856185913086, 0.02878976058959961, 0.028859392166137695, 0.02895462417602539, 0.02913689613342285, 0.028918783187866212, 0.0292044792175293, 0.028636159896850585, 0.02897817611694336, 0.028839935302734376, 0.02837299156188965, 0.028718080520629883, 0.029174783706665038, 0.028313600540161132, 0.028725248336791992, 0.02896281623840332, 0.0292044792175293, 0.028868608474731446, 0.02891263961791992, 0.029107200622558595, 0.02940825653076172, 0.02896895980834961, 0.028666879653930662, 0.02877235221862793, 0.0290119686126709, 0.028794879913330077, 0.028884992599487305, 0.02877337646484375, 0.02913382339477539, 0.02880102348327637, 0.028914688110351562, 0.028708864212036132, 0.02889625549316406, 0.028753919601440428, 0.02890547180175781, 0.028935167312622072, 0.028679168701171875, 0.028694528579711914, 0.028618751525878908, 0.02891263961791992, 0.02897407913208008, 0.02898944091796875, 0.02971238327026367, 0.028841983795166014, 0.028999679565429686, 0.028959743499755858, 0.028793855667114256, 0.02871603202819824, 0.028881919860839843, 0.028887039184570314, 0.02878054428100586, 0.029067264556884766, 0.02895871925354004, 0.028829696655273438, 0.02878873634338379, 0.029099008560180665, 0.02938265609741211, 0.028451839447021485, 0.028663808822631837, 0.028810239791870116, 0.029000703811645507, 0.02895052719116211, 0.028888063430786134, 0.028919807434082033, 0.029197311401367186, 0.028907520294189453, 0.030434303283691407, 0.03060121536254883, 0.029274112701416017, 0.02936627197265625, 0.029361152648925783, 0.029535232543945314, 0.029569023132324217, 0.02960383987426758, 0.029265920639038087, 0.02980454444885254, 0.029460479736328125, 0.028684288024902343, 0.029451263427734374, 0.029846527099609374, 0.029654016494750978, 0.02895257568359375, 0.02896588706970215, 0.02874675178527832, 0.029076480865478517, 0.02878054428100586, 0.02933350372314453, 0.029639680862426757, 0.029858816146850587, 0.02856038475036621, 0.028628992080688476, 0.029049856185913086, 0.02872012710571289, 0.028729343414306642, 0.028850175857543944, 0.028862464904785157, 0.028964864730834962, 0.02869964790344238, 0.0287825927734375, 0.0287774715423584, 0.02898124885559082, 0.02879692840576172, 0.02898227119445801, 0.031281152725219724, 0.029683712005615235, 0.02875699234008789, 0.029171712875366212, 0.02890547180175781, 0.02972159957885742, 0.02896281623840332, 0.028843008041381835, 0.02878976058959961, 0.028785663604736327, 0.028879871368408205, 0.02877132797241211, 0.029067264556884766, 0.02877952003479004, 0.028889087677001952, 0.028787712097167968, 0.02898841667175293, 0.028823551177978517, 0.02902835273742676, 0.029410303115844725, 0.029129728317260743, 0.02877644729614258, 0.028907520294189453, 0.029072383880615234, 0.02891366386413574, 0.0289751033782959, 0.02897407913208008, 0.02931609535217285, 0.028984319686889647, 0.028669952392578125, 0.028899328231811523, 0.02996633529663086, 0.029438976287841798, 0.02955673599243164, 0.030885887145996094, 0.02896076774597168, 0.029236223220825194, 0.029338623046875, 0.029499391555786132, 0.02918809509277344, 0.028900352478027344, 0.030670848846435547, 0.030100479125976562, 0.0291778564453125, 0.029322240829467775, 0.028851200103759765, 0.02897203254699707, 0.028859392166137695, 0.028659711837768553, 0.029129728317260743, 0.02975027275085449, 0.02891263961791992, 0.029053951263427736, 0.031267839431762694, 0.030234624862670898, 0.029413375854492187, 0.028826623916625976, 0.02874880027770996, 0.02877337646484375, 0.02891263961791992, 0.028811264038085937, 0.028666879653930662, 0.028901376724243165, 0.02880102348327637, 0.029096960067749023, 0.028684288024902343, 0.02892799949645996, 0.02914406394958496, 0.029237247467041014, 0.029673471450805664, 0.02894233512878418, 0.02880512046813965, 0.02894233512878418, 0.02875596809387207, 0.028700672149658202, 0.028852224349975586, 0.028909568786621095, 0.028880895614624022, 0.02877952003479004, 0.028807167053222657, 0.029085695266723634, 0.029009920120239258, 0.02905292892456055, 0.028632064819335938, 0.029077503204345705, 0.02911948776245117, 0.029230079650878905, 0.02892083168029785, 0.02897920036315918, 0.028864511489868162, 0.029160512924194335, 0.028853183746337892, 0.02872319984436035, 0.02895257568359375, 0.029064191818237304, 0.028918783187866212, 0.029065216064453125, 0.029054975509643553, 0.02895359992980957, 0.029087743759155273, 0.02879078483581543, 0.028686336517333984, 0.029024255752563476, 0.029067264556884766, 0.02901299285888672, 0.028825599670410155, 0.028923904418945313, 0.029023231506347655, 0.02899865531921387, 0.028693504333496093, 0.02877952003479004, 0.029692928314208986, 0.029121536254882813, 0.02891263961791992, 0.028786687850952147, 0.030319616317749022, 0.03080396842956543, 0.030843904495239258, 0.03077631950378418, 0.030473215103149414, 0.030278656005859376, 0.030461952209472655, 0.030740480422973632, 0.030567424774169922, 0.030289920806884765, 0.03038003158569336, 0.030675968170166015, 0.03039948844909668, 0.03038412857055664, 0.03063910484313965, 0.03082137680053711, 0.030446592330932616, 0.030737407684326173, 0.03054080009460449, 0.030523391723632814, 0.030623743057250977, 0.030691328048706053, 0.030826496124267577, 0.03039948844909668, 0.030118911743164063, 0.030645248413085937, 0.030887935638427736, 0.030486528396606444, 0.02875494384765625, 0.028856319427490236, 0.029266944885253908, 0.030688255310058594, 0.028855327606201173, 0.029094879150390624, 0.028906496047973632, 0.029232128143310547, 0.030708736419677734, 0.03041689682006836, 0.03011686325073242, 0.029842432022094727, 0.028956672668457032, 0.029114368438720704, 0.03080294418334961, 0.030883840560913086, 0.030266368865966797, 0.030465024948120117, 0.030712831497192384, 0.030327808380126952, 0.03035852813720703, 0.030264320373535155, 0.03042918395996094, 0.030269439697265626, 0.0305797119140625, 0.03040153694152832, 0.030717952728271485, 0.03043328094482422, 0.030300159454345704, 0.030472192764282226, 0.030353408813476562, 0.030259199142456054, 0.030254079818725587, 0.02900377655029297, 0.030596096038818358, 0.030373888015747072, 0.030338048934936523, 0.030670848846435547, 0.029414400100708008, 0.029832191467285156, 0.029116416931152345, 0.028889087677001952, 0.029254655838012695, 0.028828672409057617, 0.030299135208129883, 0.029643775939941407, 0.030287872314453124, 0.030431232452392577, 0.03054182434082031, 0.030453760147094725, 0.029018112182617187, 0.028778495788574218, 0.03084492874145508, 0.030290943145751953, 0.0305797119140625, 0.030488576889038086, 0.02892799949645996, 0.02897407913208008, 0.028999679565429686, 0.029039615631103514, 0.03061862373352051, 0.029815807342529296, 0.030456832885742188, 0.03054591941833496, 0.03058892822265625, 0.030239744186401366, 0.030327808380126952, 0.03040768051147461, 0.030505983352661133, 0.030239744186401366, 0.028889087677001952, 0.028925952911376954, 0.029418495178222655, 0.029080575942993164, 0.029162496566772462, 0.02895564842224121, 0.029017087936401367, 0.02935910415649414, 0.028815359115600587, 0.02896588706970215, 0.02892902374267578, 0.028851200103759765, 0.028985343933105468, 0.02900480079650879, 0.02896691131591797, 0.028992511749267577, 0.028856319427490236, 0.028725248336791992, 0.029074432373046875, 0.02896691131591797, 0.029024255752563476, 0.029099008560180665, 0.02890444755554199, 0.029068288803100587, 0.02874470329284668, 0.029113344192504883, 0.028859392166137695, 0.02859519958496094, 0.028871679306030275, 0.029061119079589845, 0.02903654479980469, 0.028833791732788085, 0.02896384048461914, 0.029275136947631834, 0.029422592163085938, 0.029051904678344728, 0.028951551437377928, 0.028903423309326173, 0.029014015197753908, 0.0289617919921875, 0.0287774715423584, 0.02877235221862793, 0.028848127365112306, 0.028832767486572267, 0.02877644729614258, 0.029198335647583007, 0.03034316825866699, 0.030426111221313477, 0.030445568084716795, 0.030542848587036132, 0.030349311828613282, 0.03035443115234375, 0.03041177558898926, 0.028819456100463867, 0.028891136169433593, 0.029072383880615234, 0.02891263961791992, 0.029049856185913086, 0.02910207939147949, 0.028867584228515625, 0.028860416412353516, 0.028918783187866212, 0.02895052719116211, 0.028841983795166014, 0.028658687591552736, 0.02887887954711914, 0.028800991058349608, 0.028770303726196288, 0.029241344451904298, 0.02896895980834961, 0.02878361511230469, 0.02905292892456055, 0.0289617919921875, 0.02894540786743164, 0.028914688110351562, 0.028604415893554686, 0.028856319427490236, 0.028713983535766603, 0.029275136947631834, 0.02880512046813965, 0.028919807434082033, 0.02911948776245117, 0.029121536254882813, 0.028892160415649414, 0.028852224349975586, 0.028726272583007813, 0.02894438362121582, 0.029105152130126953, 0.028899328231811523, 0.02935910415649414, 0.029290496826171877, 0.028894208908081056, 0.029138944625854493, 0.02879795265197754, 0.03019059181213379, 0.030493696212768553, 0.03061452865600586, 0.029611007690429687, 0.03078041648864746, 0.03078963279724121, 0.030885887145996094, 0.030317567825317384, 0.030547967910766603, 0.030497791290283204, 0.030501888275146483, 0.030253055572509766, 0.03154022407531738, 0.03078246307373047, 0.03077939224243164, 0.03057052803039551, 0.030074848175048827, 0.030117887496948242, 0.030563327789306642, 0.030478336334228515, 0.030517248153686522, 0.03034419250488281, 0.030648319244384766, 0.030654464721679688, 0.030445568084716795, 0.028943359375, 0.029139968872070314, 0.02898944091796875, 0.028834815979003905, 0.028804096221923828, 0.028902399063110353, 0.02902835273742676, 0.028890111923217773, 0.02930892753601074, 0.030292991638183595, 0.030505983352661133, 0.0304005126953125, 0.02976870346069336, 0.028880895614624022, 0.029182975769042968, 0.029065216064453125, 0.028932096481323243, 0.028997631072998048, 0.028851200103759765, 0.028886016845703126, 0.028762111663818358, 0.028909568786621095, 0.028931072235107422, 0.029050880432128907, 0.029086719512939452, 0.029035520553588868, 0.029228031158447267, 0.03216793441772461, 0.030608383178710938, 0.030604288101196288, 0.030047231674194336, 0.030013439178466796, 0.02997760009765625, 0.030316543579101563, 0.030269439697265626, 0.03033907127380371, 0.031006719589233397, 0.031718399047851564, 0.031013919830322267, 0.030535648345947266, 0.030478336334228515, 0.030258176803588867, 0.03042918395996094, 0.03056230354309082, 0.03058073616027832, 0.030880863189697266, 0.03057244873046875, 0.03056025505065918, 0.03059712028503418, 0.030769151687622072, 0.030682111740112306, 0.030483455657958985, 0.03081523132324219, 0.030259199142456054, 0.03055820846557617, 0.03055615997314453, 0.030100479125976562, 0.030483455657958985, 0.03042815971374512, 0.03060121536254883, 0.03042815971374512, 0.031265792846679685, 0.030349311828613282, 0.030616575241088868, 0.030475263595581056, 0.030254079818725587, 0.03021004867553711, 0.030309375762939454, 0.030220287322998047, 0.02977689552307129, 0.03056025505065918, 0.03038515281677246, 0.030373888015747072, 0.03042815971374512, 0.030319616317749022, 0.02992742347717285, 0.028992511749267577, 0.030078975677490235, 0.030649343490600587, 0.030337024688720703, 0.03034316825866699, 0.030655487060546875, 0.02897817611694336, 0.029120512008666992, 0.02894540786743164, 0.028997631072998048, 0.029040639877319335, 0.029045759201049806, 0.02879078483581543, 0.028851200103759765, 0.028949504852294923, 0.02880102348327637, 0.028957696914672853, 0.028901376724243165, 0.028867584228515625, 0.028841983795166014, 0.029035520553588868, 0.029005823135375978, 0.028895263671875, 0.02884809684753418, 0.029082624435424805, 0.02901094436645508, 0.0285849609375, 0.029014015197753908, 0.02855116844177246, 0.028697599411010744, 0.028823551177978517, 0.028859392166137695, 0.028891136169433593, 0.02897203254699707, 0.029500415802001953, 0.029064191818237304, 0.02913382339477539, 0.029066240310668946, 0.028949504852294923, 0.02940620803833008, 0.030470144271850585, 0.03040870475769043, 0.02983628845214844, 0.030511104583740234, 0.03073843193054199, 0.030438400268554686, 0.03033087921142578, 0.030329856872558594, 0.030007295608520508, 0.030294015884399415, 0.030227455139160156, 0.030708736419677734, 0.03019366455078125, 0.030324735641479493, 0.030216192245483397, 0.029859840393066408, 0.03038105583190918, 0.029019136428833008, 0.02889625549316406, 0.03035136032104492, 0.03033497619628906, 0.030620672225952147]",tokens/s,33.96635991275159,,,,,,gemma,MB,4745.924608,6422.003712,0.0,5792.333824,5786.358272,s,1,11.137408203125,11.137408203125,0.0,11.137408203125,11.137408203125,11.137408203125,11.137408203125,[11.137408203125],,kWh,5.037324592430537e-05,2.757378016947258e-05,7.052061197200532e-05,0.00014846763806578326, @@ -5976,48 +5880,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-sdpa,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,,,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,4113.342464,5287.44448,0.0,4640.997376,4330.668544,s,10,3.2019090576171876,0.3201909057617187,0.0017134996701674667,0.31955131530761716,0.3223881805419922,0.3223929733276367,0.32239680755615235,"[0.3185635986328125, 0.3218643493652344, 0.318020263671875, 0.3190321350097656, 0.3193435974121094, 0.3183670349121094, 0.319759033203125, 0.32239776611328125, 0.3221741638183594, 0.3223871154785156]",tokens/s,799.5230201525817,kWh,3.7596130847103066e-06,2.060095877009083e-06,1.8658895135439252e-05,2.447860409715864e-05,tokens/kWh,10458112.6841998,MB,4113.342464,5287.44448,0.0,4640.997376,4515.667456,s,10,175.162884765625,17.5162884765625,0.0013239684065771513,17.516396484375,17.5170955078125,17.51822841796875,17.51913474609375,"[17.5166953125, 17.51646484375, 17.51684375, 17.515330078125, 17.516587890625, 17.51393359375, 17.519361328125, 17.51530859375, 17.516328125, 17.51603125]",tokens/s,3.596652343577039,kWh,0.00020683056154391835,0.0001133576330349002,0.0010022863660645621,0.0013224745606433806,tokens/kWh,47637.96739451128,,s,629,177.76777615356434,0.2826196759198163,0.038106541032305166,0.27801907348632815,0.2782052307128906,0.2782926879882812,0.5975155737304687,"[0.27855975341796874, 0.27801702880859375, 0.2778101806640625, 0.2780354614257812, 0.27816244506835935, 0.2778398742675781, 0.2780078125, 0.27787161254882814, 0.2779852905273438, 0.27780914306640625, 0.27788491821289063, 0.2780006408691406, 0.2780405883789063, 0.27786956787109374, 0.27781939697265623, 0.27792691040039064, 0.27795147705078127, 0.27795660400390626, 0.2780201110839844, 0.27802520751953125, 0.2780088195800781, 0.2779525146484375, 0.2779207763671875, 0.2779402160644531, 0.2781061096191406, 0.27823513793945315, 0.2779893798828125, 0.27797503662109374, 0.27814501953125, 0.2779105224609375, 0.2781542358398438, 0.2779914245605469, 0.2781327209472656, 0.27813067626953125, 0.2779105224609375, 0.2780354614257812, 0.27810202026367187, 0.27796890258789064, 0.2780477294921875, 0.27807232666015624, 0.2782822265625, 0.2780149841308594, 0.27799859619140627, 0.27797402954101563, 0.2781419372558594, 0.2780426330566406, 0.2779207763671875, 0.278086669921875, 0.2782423095703125, 0.27799755859375, 0.2780201110839844, 0.27810202026367187, 0.2780508117675781, 0.27797607421875, 0.2779842529296875, 0.27804568481445313, 0.27818701171875, 0.2779914245605469, 0.27808255004882815, 0.2781388854980469, 0.2782474365234375, 0.2780313720703125, 0.6001817626953125, 0.27799859619140627, 0.27790234375, 0.2779320373535156, 0.27799655151367186, 0.27801907348632815, 0.2778521728515625, 0.27775384521484375, 0.27788595581054687, 0.2779473876953125, 0.27779583740234376, 0.2782218322753906, 0.27791156005859374, 0.27793304443359373, 0.277886962890625, 0.2779207763671875, 0.27789208984375, 0.27796890258789064, 0.27791461181640625, 0.278029296875, 0.27789312744140626, 0.2780139465332031, 0.2779402160644531, 0.2780508117675781, 0.2779996032714844, 0.2780682373046875, 0.27781939697265623, 0.2778675231933594, 0.2782105712890625, 0.27816244506835935, 0.2780088195800781, 0.27801089477539065, 0.27811737060546876, 0.27817166137695315, 0.27833856201171875, 0.27794842529296876, 0.27794329833984377, 0.27803955078125, 0.2779781188964844, 0.2778316650390625, 0.2780487670898438, 0.2780835876464844, 0.2780579833984375, 0.2780610656738281, 0.2779648132324219, 0.2781368408203125, 0.278066162109375, 0.27846142578125, 0.2780498046875, 0.2782576599121094, 0.2779914245605469, 0.27807333374023435, 0.27804568481445313, 0.27823309326171874, 0.27794842529296876, 0.2780641174316406, 0.27811431884765625, 0.2781614074707031, 0.27800985717773435, 0.2780682373046875, 0.278029296875, 0.2781337585449219, 0.27806719970703125, 0.5975459594726562, 0.2780262451171875, 0.2778736572265625, 0.2778173522949219, 0.277960693359375, 0.2780528564453125, 0.27834878540039065, 0.2778900451660156, 0.27787673950195313, 0.2779422607421875, 0.27794638061523436, 0.2780139465332031, 0.27793408203125, 0.27799655151367186, 0.27793304443359373, 0.2779371643066406, 0.2779084777832031, 0.2779934692382813, 0.27786343383789064, 0.27802520751953125, 0.2779924621582031, 0.2782013549804688, 0.27794329833984377, 0.2780282897949219, 0.2780559387207031, 0.27812966918945314, 0.27786648559570315, 0.2779525146484375, 0.27802725219726565, 0.27815731811523436, 0.27795455932617186, 0.27812249755859375, 0.2780569458007813, 0.27815731811523436, 0.277923828125, 0.2779637756347656, 0.27835391235351564, 0.27807437133789065, 0.27791256713867185, 0.2779658203125, 0.2780928039550781, 0.27812966918945314, 0.2783221740722656, 0.2781583251953125, 0.2779832458496094, 0.27802520751953125, 0.27819622802734373, 0.2780426330566406, 0.2780415954589844, 0.27818292236328124, 0.27791156005859374, 0.2779729919433594, 0.2780845947265625, 0.27821261596679686, 0.2780334167480469, 0.27809588623046877, 0.27812863159179685, 0.2781747131347656, 0.2780631103515625, 0.2779945068359375, 0.27803033447265624, 0.2781747131347656, 0.2780282897949219, 0.5974374389648438, 0.27794943237304687, 0.27812762451171874, 0.27768115234375, 0.2779719543457031, 0.27803955078125, 0.27782247924804687, 0.2778367919921875, 0.27792691040039064, 0.27802725219726565, 0.2777835388183594, 0.2779033508300781, 0.27785626220703125, 0.27808563232421873, 0.2782443542480469, 0.2781737060546875, 0.2779832458496094, 0.2779924621582031, 0.2779576416015625, 0.27801907348632815, 0.27789413452148437, 0.2780528564453125, 0.27785626220703125, 0.27797503662109374, 0.27792691040039064, 0.27814501953125, 0.2779156494140625, 0.27798016357421873, 0.2780313720703125, 0.2781163635253906, 0.27802420043945314, 0.2779832458496094, 0.27799755859375, 0.27810714721679686, 0.2778880004882813, 0.2779371643066406, 0.27798834228515623, 0.27807949829101564, 0.27803033447265624, 0.2779576416015625, 0.2780262451171875, 0.2780784606933594, 0.27793408203125, 0.27805902099609375, 0.27813784790039064, 0.2781317138671875, 0.27809381103515624, 0.2779576416015625, 0.27808154296875, 0.2782586975097656, 0.2782494812011719, 0.2780610656738281, 0.27814501953125, 0.2781214599609375, 0.2779207763671875, 0.277923828125, 0.2779832458496094, 0.27808154296875, 0.27801907348632815, 0.27799755859375, 0.2781317138671875, 0.27814910888671873, 0.2781051025390625, 0.5979115600585938, 0.2780415954589844, 0.2779361267089844, 0.27780813598632814, 0.27787774658203124, 0.2779586486816406, 0.2779361267089844, 0.27787774658203124, 0.278002685546875, 0.278002685546875, 0.2778736572265625, 0.27790234375, 0.2779648132324219, 0.2780006408691406, 0.27782861328125, 0.277970947265625, 0.2783703002929688, 0.2780405883789063, 0.27791360473632815, 0.27801702880859375, 0.2779422607421875, 0.27796786499023435, 0.2783999938964844, 0.27802520751953125, 0.278012939453125, 0.2780917663574219, 0.277907470703125, 0.2779371643066406, 0.27806207275390626, 0.2780498046875, 0.27799038696289063, 0.2780682373046875, 0.27811328125, 0.2783170471191406, 0.27801190185546876, 0.27803237915039064, 0.2780487670898438, 0.27803237915039064, 0.2779955139160156, 0.27786138916015624, 0.2780262451171875, 0.2781327209472656, 0.27797503662109374, 0.2780948486328125, 0.2779361267089844, 0.27810714721679686, 0.27803851318359374, 0.27794943237304687, 0.2780702819824219, 0.27821774291992185, 0.2779535217285156, 0.2780139465332031, 0.27819314575195314, 0.2782003173828125, 0.2779729919433594, 0.27806515502929685, 0.278012939453125, 0.2781890563964844, 0.27835595703125, 0.27808767700195314, 0.27808563232421873, 0.2781542358398438, 0.278086669921875, 0.596664306640625, 0.27794638061523436, 0.27794125366210937, 0.2780057678222656, 0.27801803588867186, 0.27811328125, 0.27796173095703125, 0.277855224609375, 0.2778839111328125, 0.278002685546875, 0.2778757019042969, 0.27769650268554685, 0.2778746948242187, 0.2780262451171875, 0.27781427001953124, 0.2780006408691406, 0.27796685791015624, 0.27805389404296876, 0.27779379272460936, 0.2778880004882813, 0.27790438842773435, 0.27795559692382815, 0.2778480529785156, 0.27800167846679685, 0.27791360473632815, 0.27811019897460937, 0.2778890380859375, 0.27791973876953124, 0.2779453430175781, 0.2782668762207031, 0.27817062377929686, 0.27803750610351563, 0.2780006408691406, 0.2780569458007813, 0.2779402160644531, 0.2778787841796875, 0.2779443054199219, 0.2780559387207031, 0.27797708129882814, 0.2779002990722656, 0.2781419372558594, 0.2780415954589844, 0.277960693359375, 0.27822592163085935, 0.27799755859375, 0.2780968933105469, 0.27796890258789064, 0.27799655151367186, 0.278034423828125, 0.2782085266113281, 0.2779279479980469, 0.27796890258789064, 0.27808050537109374, 0.27819110107421874, 0.2779781188964844, 0.2779361267089844, 0.2780231628417969, 0.2781419372558594, 0.2780088195800781, 0.27789926147460936, 0.27808563232421873, 0.2782248840332031, 0.2780631103515625, 0.5981163330078125, 0.2783621215820313, 0.2783006591796875, 0.2778358154296875, 0.27795965576171877, 0.2779729919433594, 0.2779207763671875, 0.2778347473144531, 0.27787774658203124, 0.27798016357421873, 0.277865478515625, 0.2778245239257813, 0.2778685302734375, 0.27808563232421873, 0.27831195068359377, 0.27789208984375, 0.2779105224609375, 0.27801907348632815, 0.2779402160644531, 0.27796890258789064, 0.27797503662109374, 0.27805184936523436, 0.277907470703125, 0.2780078125, 0.27791873168945314, 0.2780928039550781, 0.2779361267089844, 0.27801702880859375, 0.2780907592773437, 0.27824127197265625, 0.27799038696289063, 0.2781542358398438, 0.2781747131347656, 0.27819827270507813, 0.27798019409179686, 0.27835592651367186, 0.2779781188964844, 0.2782853088378906, 0.27827609252929686, 0.2779586486816406, 0.27801190185546876, 0.27814093017578123, 0.27794329833984377, 0.27867544555664064, 0.27796786499023435, 0.27813479614257813, 0.27802520751953125, 0.27811532592773436, 0.27813580322265624, 0.2783088684082031, 0.2785464172363281, 0.278076416015625, 0.2782156677246094, 0.27824331665039065, 0.27822900390625, 0.27794329833984377, 0.27808767700195314, 0.27831195068359377, 0.27817266845703126, 0.27806719970703125, 0.27806002807617186, 0.2782515258789062, 0.2780887145996094, 0.60076953125, 0.2780528564453125, 0.27787774658203124, 0.2780835876464844, 0.2779893798828125, 0.2780364685058594, 0.27775180053710935, 0.2778787841796875, 0.2780692443847656, 0.2780835876464844, 0.2782095336914063, 0.27780813598632814, 0.2778828735351562, 0.27790234375, 0.27783782958984377, 0.2779309997558594, 0.27794125366210937, 0.2780784606933594, 0.277897216796875, 0.2780211181640625, 0.2779945068359375, 0.27791973876953124, 0.2779525146484375, 0.278066162109375, 0.27791256713867185, 0.2780282897949219, 0.27787673950195313, 0.2778757019042969, 0.27794842529296876, 0.27817779541015625, 0.2779361267089844, 0.2780845947265625, 0.27806207275390626, 0.27812249755859375, 0.27798834228515623, 0.27791668701171873, 0.27791360473632815, 0.277960693359375, 0.2780917663574219, 0.2781583251953125, 0.27809588623046877, 0.27802725219726565, 0.27801190185546876, 0.2780579833984375, 0.27796273803710936, 0.27820440673828123, 0.2784163818359375, 0.2779811706542969, 0.278076416015625, 0.2782392272949219, 0.27798220825195313, 0.2780231628417969, 0.27797708129882814, 0.2781951904296875, 0.2779914245605469, 0.2780313720703125, 0.277970947265625, 0.2780968933105469, 0.2780917663574219, 0.278002685546875, 0.2780364685058594, 0.27814501953125, 0.278044677734375, 0.6006200561523437, 0.27797503662109374, 0.2778890380859375, 0.2780282897949219, 0.2779402160644531, 0.27803851318359374, 0.277928955078125, 0.27785931396484376, 0.2780006408691406, 0.2779156494140625, 0.2777651062011719, 0.2779658203125, 0.27800677490234377, 0.2779637756347656, 0.2778880004882813, 0.27796173095703125, 0.2779176940917969, 0.2780159912109375, 0.27834878540039065, 0.27800985717773435, 0.27794842529296876, 0.2780631103515625, 0.2779105224609375, 0.27805490112304687, 0.27783578491210936, 0.2781808776855469, 0.2778818664550781, 0.27799755859375, 0.2780211181640625, 0.27812045288085935, 0.27800473022460936, 0.27808154296875, 0.2780897216796875, 0.2780979309082031, 0.277960693359375, 0.27794329833984377, 0.2780313720703125, 0.27806719970703125, 0.2780088195800781, 0.2779719543457031, 0.27808563232421873, 0.27815731811523436, 0.2779842529296875, 0.2779084777832031, 0.2779156494140625, 0.2780907592773437, 0.27803750610351563, 0.2781327209472656, 0.27808563232421873, 0.27822283935546877, 0.27808050537109374, 0.27801803588867186, 0.2780968933105469, 0.278297607421875, 0.2783078308105469, 0.2780487670898438, 0.27810302734375, 0.2782310485839844, 0.2781327209472656, 0.27803851318359374, 0.2779381713867187, 0.2782033996582031, 0.27808050537109374, 0.6008156127929688, 0.2780989379882812, 0.27795147705078127, 0.2778542175292969, 0.2780559387207031, 0.2779955139160156, 0.2778808288574219, 0.2778542175292969, 0.2780201110839844, 0.27807745361328123, 0.27784295654296876, 0.27786956787109374, 0.27791156005859374, 0.2778890380859375, 0.27790234375, 0.27802520751953125, 0.27789926147460936, 0.27797402954101563, 0.2777374572753906, 0.27824026489257814, 0.2779862976074219, 0.27802420043945314, 0.2778818664550781, 0.27795455932617186, 0.27798220825195313, 0.2781051025390625, 0.278150146484375, 0.277886962890625, 0.2780579833984375, 0.27816754150390627, 0.2780149841308594, 0.2779729919433594, 0.27802420043945314, 0.27813784790039064, 0.27803237915039064, 0.27795147705078127, 0.27801089477539065, 0.278044677734375, 0.27811224365234377, 0.27802215576171874, 0.2780487670898438, 0.27812863159179685, 0.27805389404296876, 0.27803955078125, 0.2779996032714844, 0.27808767700195314, 0.2780692443847656, 0.27806207275390626, 0.27810302734375, 0.27824639892578124, 0.27794125366210937, 0.27805389404296876, 0.2780415954589844, 0.2782115783691406, 0.2779934692382813, 0.27796786499023435, 0.2781685791015625, 0.2781685791015625, 0.27801089477539065, 0.2780498046875, 0.2780282897949219, 0.27819314575195314, 0.27838156127929686]",tokens/s,3.538324063055383,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch @@ -7057,6 +6946,33 @@ ChildProcessError: Traceback (most recent call last): OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,2108.653568,5971.116032,0.0,5324.668928,5080.876032,s,10,3.254842498779297,0.3254842498779297,0.001114065237433989,0.3250286560058594,0.3257514434814453,0.3272673233032226,0.3284800271606445,"[0.328783203125, 0.32502313232421876, 0.3253096313476562, 0.3250040283203125, 0.32493109130859377, 0.3254145812988281, 0.3249727783203125, 0.32496087646484373, 0.3250341796875, 0.32540899658203126]",tokens/s,786.5203926027474,kWh,3.84116784859729e-06,2.1047766923490397e-06,1.920148489522528e-05,2.514742943617161e-05,tokens/kWh,10179966.928618725,MB,2108.653568,5971.116032,0.0,5324.668928,5290.118656,s,10,177.13133984375,17.713133984375002,0.001836879786445902,17.712749023437503,17.7155212890625,17.71599501953125,17.71637400390625,"[17.7094375, 17.71198828125, 17.71232421875, 17.712689453125, 17.712490234375, 17.713939453125, 17.71377734375, 17.71280859375, 17.71646875, 17.715416015625]",tokens/s,3.5566828577920293,kWh,0.00020915095838113933,0.00011463212397823582,0.0010249259615171765,0.0013487090438765517,tokens/kWh,46711.335025174216,,s,629,179.77634902954114,0.2858129555318617,0.038628015635411614,0.2811412353515625,0.2814228515625,0.28152381591796877,0.6061751391601563,"[0.2810777587890625, 0.28092620849609373, 0.28088525390625, 0.280838134765625, 0.2809333801269531, 0.28067941284179687, 0.28087091064453124, 0.2809241638183594, 0.2810931091308594, 0.28108184814453124, 0.28116070556640627, 0.28089447021484376, 0.2808821716308594, 0.28095077514648437, 0.2809036865234375, 0.280943603515625, 0.2808340454101563, 0.2809927673339844, 0.28108184814453124, 0.281091064453125, 0.28095794677734376, 0.2809938049316406, 0.2811002807617187, 0.28091595458984375, 0.28110848999023436, 0.2811279296875, 0.2810521545410156, 0.2811770935058594, 0.28101223754882815, 0.281059326171875, 0.2810245056152344, 0.2809323425292969, 0.28119961547851563, 0.2811125793457031, 0.28143206787109376, 0.281101318359375, 0.28115045166015623, 0.28138803100585935, 0.2812610473632812, 0.28120166015625, 0.28122726440429685, 0.28131329345703127, 0.28121701049804687, 0.28117913818359375, 0.2809989013671875, 0.2813675537109375, 0.28137677001953126, 0.28117913818359375, 0.2810777587890625, 0.28124774169921873, 0.281164794921875, 0.281196533203125, 0.28134707641601564, 0.2811832275390625, 0.2812538757324219, 0.28116888427734377, 0.28106854248046875, 0.28122726440429685, 0.281122802734375, 0.28126617431640627, 0.28131329345703127, 0.28124774169921873, 0.6063380737304688, 0.28098458862304687, 0.2808115234375, 0.2807510986328125, 0.28091802978515623, 0.28093646240234377, 0.28100302124023435, 0.28108084106445314, 0.2809866333007813, 0.2808965148925781, 0.2809385070800781, 0.2809046936035156, 0.2808934326171875, 0.28086273193359373, 0.28120269775390627, 0.2810859375, 0.28100811767578127, 0.28084326171875, 0.2810449829101562, 0.2810654602050781, 0.28101937866210935, 0.28106341552734376, 0.28127743530273436, 0.28122113037109375, 0.2810931091308594, 0.28118118286132815, 0.28112384033203125, 0.28111358642578127, 0.28134194946289065, 0.28118118286132815, 0.2811453552246094, 0.2811934814453125, 0.2813255615234375, 0.2810931091308594, 0.2812149658203125, 0.28111154174804687, 0.2811781005859375, 0.28105624389648437, 0.2810224609375, 0.28145050048828124, 0.28102859497070315, 0.2810921020507812, 0.2813788146972656, 0.28130712890625, 0.2813788146972656, 0.2810214538574219, 0.2815672302246094, 0.28127435302734377, 0.2810726318359375, 0.28103067016601563, 0.28111154174804687, 0.28136038208007813, 0.28119143676757813, 0.28112896728515624, 0.28117913818359375, 0.28128256225585935, 0.2813081665039063, 0.28129995727539064, 0.28135321044921874, 0.28134707641601564, 0.28152935791015626, 0.2813255615234375, 0.28147198486328123, 0.6066227416992187, 0.2809098205566406, 0.2808719482421875, 0.28102554321289064, 0.28104702758789063, 0.2808401794433594, 0.28098968505859373, 0.2811484069824219, 0.28116378784179685, 0.28091802978515623, 0.28097842407226564, 0.2810368041992187, 0.28097125244140625, 0.2808842163085937, 0.28110336303710937, 0.28096511840820315, 0.2808688659667969, 0.281027587890625, 0.2810245056152344, 0.28101119995117185, 0.2810521545410156, 0.2808965148925781, 0.281069580078125, 0.2811453552246094, 0.28131329345703127, 0.28105624389648437, 0.2812549133300781, 0.28084225463867185, 0.28120065307617187, 0.28095281982421877, 0.2810378112792969, 0.28143515014648435, 0.2811975708007812, 0.28118936157226565, 0.2811351013183594, 0.2813869934082031, 0.28113714599609374, 0.28141876220703127, 0.28113201904296875, 0.2815682678222656, 0.2815467529296875, 0.28107672119140625, 0.281275390625, 0.2812549133300781, 0.2810992736816406, 0.28117196655273435, 0.28111154174804687, 0.28140032958984373, 0.2811883544921875, 0.28148223876953127, 0.281238525390625, 0.2814341125488281, 0.28122830200195315, 0.28124465942382815, 0.2812200927734375, 0.2811576232910156, 0.2811269226074219, 0.2812252197265625, 0.2812303466796875, 0.28119961547851563, 0.28121908569335935, 0.28127743530273436, 0.2814617614746094, 0.6061670532226563, 0.2809108581542969, 0.28080230712890625, 0.2814484558105469, 0.28111358642578127, 0.28132147216796877, 0.28121701049804687, 0.2810378112792969, 0.2809323425292969, 0.2810368041992187, 0.2813255615234375, 0.28106341552734376, 0.28104702758789063, 0.28093438720703123, 0.28106036376953125, 0.281196533203125, 0.28113919067382814, 0.2808115234375, 0.2810900573730469, 0.2811146240234375, 0.2812098693847656, 0.281169921875, 0.2809169921875, 0.28105010986328127, 0.28096511840820315, 0.2809661560058594, 0.28101119995117185, 0.2809743347167969, 0.2809354248046875, 0.28098150634765623, 0.28091290283203124, 0.28116888427734377, 0.28102655029296875, 0.280943603515625, 0.281064453125, 0.2811545715332031, 0.2809794616699219, 0.2811832275390625, 0.28138189697265625, 0.2813788146972656, 0.281302001953125, 0.2813061218261719, 0.2811617431640625, 0.28119143676757813, 0.281169921875, 0.28117913818359375, 0.2815242309570313, 0.2812149658203125, 0.28117913818359375, 0.28131942749023436, 0.2813040771484375, 0.2812733459472656, 0.281275390625, 0.28110336303710937, 0.2812569580078125, 0.2811883544921875, 0.28134808349609375, 0.2814064636230469, 0.2813460388183594, 0.2813624267578125, 0.28149041748046877, 0.28145458984375, 0.28124365234375, 0.6059878540039062, 0.2807142333984375, 0.28078695678710935, 0.2806753234863281, 0.2809016418457031, 0.28094772338867186, 0.280922119140625, 0.2809876403808594, 0.28105523681640626, 0.2810040283203125, 0.2814208068847656, 0.28126617431640627, 0.2815948791503906, 0.281312255859375, 0.2813163452148438, 0.2814361572265625, 0.2813061218261719, 0.28089959716796875, 0.2814156799316406, 0.2811781005859375, 0.2810296325683594, 0.28103167724609374, 0.28088934326171877, 0.28106649780273435, 0.28101937866210935, 0.2810941467285156, 0.2810859375, 0.2814464111328125, 0.2812600402832031, 0.2811463623046875, 0.28102041625976565, 0.2810408935546875, 0.28117300415039065, 0.2810572814941406, 0.280922119140625, 0.28111871337890626, 0.28110336303710937, 0.28109619140625, 0.2811473999023438, 0.28106036376953125, 0.28131121826171873, 0.2811351013183594, 0.281248779296875, 0.28103067016601563, 0.28100506591796875, 0.2814095458984375, 0.2813429870605469, 0.2812538757324219, 0.2815825805664062, 0.2811955261230469, 0.28126412963867187, 0.28140237426757814, 0.2811934814453125, 0.28137777709960937, 0.28109619140625, 0.28129278564453125, 0.2812129211425781, 0.28120065307617187, 0.2812610473632812, 0.2811412353515625, 0.2812590026855469, 0.2810992736816406, 0.28120065307617187, 0.6068009033203124, 0.2809374694824219, 0.281122802734375, 0.2810439758300781, 0.28092825317382814, 0.2809692077636719, 0.280853515625, 0.281069580078125, 0.2810142822265625, 0.2809354248046875, 0.28104702758789063, 0.2809108581542969, 0.28106854248046875, 0.28096307373046875, 0.28088626098632813, 0.2809968566894531, 0.28083712768554686, 0.281101318359375, 0.2811217956542969, 0.28138290405273436, 0.281101318359375, 0.281059326171875, 0.2809938049316406, 0.2811545715332031, 0.28104193115234377, 0.2813081665039063, 0.28128460693359375, 0.28130508422851563, 0.2814761047363281, 0.28104193115234377, 0.2812200927734375, 0.28116888427734377, 0.2809876403808594, 0.2812538757324219, 0.28137985229492185, 0.28124774169921873, 0.2811678771972656, 0.28140032958984373, 0.2810624084472656, 0.28133273315429685, 0.2811002807617187, 0.28105831909179685, 0.2811576232910156, 0.2810726318359375, 0.28113919067382814, 0.2811269226074219, 0.28103372192382814, 0.28134808349609375, 0.28107672119140625, 0.2812047424316406, 0.2810511474609375, 0.2811412353515625, 0.2813204345703125, 0.2812037048339844, 0.2814218139648437, 0.2817884216308594, 0.28124978637695314, 0.281523193359375, 0.2813183898925781, 0.28128460693359375, 0.2815395812988281, 0.28148736572265626, 0.28139520263671874, 0.6065490112304688, 0.2811002807617187, 0.28092620849609373, 0.28095489501953125, 0.2808719482421875, 0.28084326171875, 0.2811023254394531, 0.281069580078125, 0.2810491027832031, 0.2809333801269531, 0.280995849609375, 0.2810798034667969, 0.2810992736816406, 0.2809169921875, 0.28097329711914065, 0.28101837158203125, 0.28094259643554687, 0.2809354248046875, 0.28100607299804686, 0.2810429382324219, 0.28095794677734376, 0.280880126953125, 0.2807767028808594, 0.28106649780273435, 0.2810859375, 0.28120574951171873, 0.2813941650390625, 0.28095281982421877, 0.2810931091308594, 0.2811975708007812, 0.28105831909179685, 0.28106341552734376, 0.2813562927246094, 0.28138803100585935, 0.2810921020507812, 0.28107879638671873, 0.2809671630859375, 0.28129791259765624, 0.28090573120117185, 0.28116070556640627, 0.2810798034667969, 0.28148837280273437, 0.2814392395019531, 0.2813061218261719, 0.2811269226074219, 0.2812876892089844, 0.28159078979492186, 0.2817607727050781, 0.2814699401855469, 0.2814566345214844, 0.2813061218261719, 0.2814146423339844, 0.28142489624023437, 0.28127947998046876, 0.28125595092773437, 0.2811934814453125, 0.28141055297851564, 0.2812682189941406, 0.2812600402832031, 0.2815703125, 0.2814689331054688, 0.28122113037109375, 0.28161843872070313, 0.6061782836914062, 0.2809692077636719, 0.28095693969726565, 0.2809661560058594, 0.28118014526367185, 0.2808770446777344, 0.2809989013671875, 0.28118118286132815, 0.28119961547851563, 0.281017333984375, 0.28093646240234377, 0.28108901977539064, 0.2810705871582031, 0.28113409423828123, 0.28105831909179685, 0.28111358642578127, 0.28118426513671874, 0.2812620849609375, 0.2812037048339844, 0.28109619140625, 0.28100607299804686, 0.281069580078125, 0.281059326171875, 0.28090573120117185, 0.2810521545410156, 0.28109515380859373, 0.28112997436523435, 0.28092108154296874, 0.2811074523925781, 0.2809927673339844, 0.28092724609375, 0.281491455078125, 0.28131942749023436, 0.281628662109375, 0.2811770935058594, 0.2811514892578125, 0.2810828857421875, 0.28110540771484377, 0.2809518127441406, 0.28126617431640627, 0.28103884887695313, 0.28114227294921873, 0.2810705871582031, 0.28101324462890626, 0.281322509765625, 0.28095489501953125, 0.2812886962890625, 0.2812037048339844, 0.2814197692871094, 0.2811985778808594, 0.28113714599609374, 0.2812467346191406, 0.28106854248046875, 0.2812129211425781, 0.28124978637695314, 0.2813388671875, 0.2813460388183594, 0.2814832763671875, 0.28139620971679685, 0.2815467529296875, 0.28133477783203126, 0.28151397705078124, 0.28122418212890626, 0.6063093872070312, 0.2810449829101562, 0.2814197692871094, 0.28110848999023436, 0.2810439758300781, 0.28139007568359375, 0.2809241638183594, 0.2809989013671875, 0.28083096313476563, 0.28112384033203125, 0.2810460205078125, 0.281302001953125, 0.28109515380859373, 0.2808883056640625, 0.28115966796875, 0.2810040283203125, 0.28084225463867185, 0.2809395141601562, 0.28089447021484376, 0.2809190368652344, 0.281059326171875, 0.28120779418945313, 0.2809241638183594, 0.28102655029296875, 0.28089752197265627, 0.2810101623535156, 0.28102859497070315, 0.28097537231445313, 0.28098355102539063, 0.2810992736816406, 0.2810654602050781, 0.28129995727539064, 0.2813839416503906, 0.2814535827636719, 0.2811074523925781, 0.28135833740234373, 0.28134707641601564, 0.2813183898925781, 0.28143515014648435, 0.2816296997070312, 0.28170855712890625, 0.2813061218261719, 0.2814228515625, 0.28137472534179686, 0.28131942749023436, 0.28127743530273436, 0.2817423400878906, 0.281481201171875, 0.2812590026855469, 0.2812630920410156, 0.2812231750488281, 0.2812467346191406, 0.28122113037109375, 0.2812037048339844, 0.2813644714355469, 0.281638916015625, 0.2814474182128906, 0.28127435302734377, 0.28134194946289065, 0.2812262268066406, 0.28141876220703127, 0.28152935791015626, 0.28133477783203126, 0.6068182983398438, 0.2809968566894531, 0.2809323425292969, 0.2809938049316406, 0.281027587890625, 0.28121701049804687, 0.28108389282226565, 0.2811351013183594, 0.28117401123046876, 0.28138494873046876, 0.28109722900390627, 0.2813368225097656, 0.28140850830078123, 0.28154779052734374, 0.28113409423828123, 0.2812129211425781, 0.28104806518554687, 0.2809405517578125, 0.2812672119140625, 0.281064453125, 0.28112588500976565, 0.28105523681640626, 0.28098150634765623, 0.2810439758300781, 0.28095590209960936, 0.28117300415039065, 0.28090777587890625, 0.2813061218261719, 0.2813839416503906, 0.28095077514648437, 0.2811576232910156, 0.2810654602050781, 0.2809599914550781, 0.28118118286132815, 0.2810726318359375, 0.28104193115234377, 0.28117300415039065, 0.28109515380859373, 0.281133056640625, 0.2813736877441406, 0.28131329345703127, 0.28105831909179685, 0.28130712890625, 0.2811975708007812, 0.28105624389648437, 0.2814986267089844, 0.281385986328125, 0.2814115905761719, 0.2812958679199219, 0.2812630920410156, 0.2812508239746094, 0.2813562927246094, 0.28127847290039065, 0.28153753662109376, 0.2815887451171875, 0.28126925659179686, 0.28113714599609374, 0.2811975708007812, 0.2814228515625, 0.28128564453125, 0.2811975708007812, 0.2811351013183594, 0.28143820190429686]",tokens/s,3.4987917120101386,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,1488.359424,1579.679744,0.0,933.23264,856.69632,s,10,0.4187196159362792,0.041871961593627924,0.0014808966097482894,0.04126091194152832,0.043131636810302736,0.0445920425415039,0.04576036712646484,"[0.04605244827270508, 0.04131564712524414, 0.0411776008605957, 0.04146476745605469, 0.04118272018432617, 0.04280710220336914, 0.0412061767578125, 0.0415596809387207, 0.04101833724975586, 0.04093513488769531]",tokens/s,6113.876452326468,kWh,4.934550676908758e-07,2.7027020769764927e-07,1.3917580578494868e-06,2.155483333238012e-06,tokens/kWh,118766865.90539834,MB,1488.359424,1579.679744,0.0,933.23264,899.62496,s,10,24.019910400390625,2.4019910400390625,0.019429569356012314,2.394274169921875,2.414185791015625,2.436057373046875,2.453554638671875,"[2.457928955078125, 2.409325439453125, 2.399566650390625, 2.391402587890625, 2.39863720703125, 2.391660400390625, 2.391578369140625, 2.391282958984375, 2.391639892578125, 2.396887939453125]",tokens/s,26.228241050797365,kWh,2.945148775767949e-05,1.6140594842595416e-05,7.348580184415957e-05,0.00011907788444443448,tokens/kWh,529065.4960316985,,s,629,24.35256624984743,0.03871632154188778,0.004963175047223502,0.037943294525146484,0.03889213485717773,0.03936890869140625,0.079117353515625,"[0.040226814270019534, 0.04027084732055664, 0.040318977355957034, 0.03974041748046875, 0.040825855255126955, 0.04090367889404297, 0.03897753524780274, 0.039828479766845705, 0.04030361557006836, 0.03983871841430664, 0.04004761505126953, 0.04024524688720703, 0.03980287933349609, 0.039346176147460936, 0.038795265197753906, 0.039277568817138675, 0.03892326354980469, 0.03963289642333984, 0.039182334899902346, 0.038902782440185545, 0.039244800567626956, 0.03907788848876953, 0.0392806396484375, 0.03914035034179687, 0.03907788848876953, 0.03934003067016602, 0.03828940963745117, 0.038212608337402344, 0.03859251022338867, 0.038125568389892575, 0.03932876968383789, 0.03894169616699219, 0.03829043197631836, 0.03913216018676758, 0.03880038452148438, 0.038793216705322264, 0.039011329650878904, 0.039003135681152344, 0.039147518157958985, 0.038547454833984376, 0.03806617736816406, 0.03796889495849609, 0.03785420989990235, 0.03794432067871094, 0.03905023956298828, 0.03965235137939453, 0.03999129486083984, 0.039384063720703126, 0.039397377014160156, 0.0378419189453125, 0.03808563232421875, 0.037797889709472655, 0.037797889709472655, 0.03783987045288086, 0.037891071319580076, 0.03788083267211914, 0.03912499237060547, 0.038760448455810545, 0.03786342239379883, 0.038950912475585936, 0.0378419189453125, 0.03783270263671875, 0.0792985610961914, 0.03790643310546875, 0.03781017684936523, 0.038370304107666016, 0.03831193542480469, 0.03775078582763672, 0.03785420989990235, 0.038435840606689455, 0.038365184783935545, 0.03781017684936523, 0.037905406951904294, 0.03861094284057617, 0.03815731048583984, 0.037917697906494144, 0.037776382446289065, 0.03875532913208008, 0.03827711868286133, 0.03894784164428711, 0.0380313606262207, 0.038877185821533204, 0.037923839569091795, 0.03808870315551758, 0.03812351989746094, 0.03917004776000976, 0.0390645751953125, 0.03829862213134766, 0.03843379211425781, 0.038542335510253906, 0.03800166320800781, 0.0378869743347168, 0.038740993499755856, 0.03850239944458008, 0.03803955078125, 0.038438911437988284, 0.03892838287353516, 0.03817267227172851, 0.038012928009033206, 0.03801190567016602, 0.039065601348876954, 0.03803033447265625, 0.03791462326049805, 0.03789619064331055, 0.03887923049926758, 0.038371326446533204, 0.03790848159790039, 0.037904384613037106, 0.03786956787109375, 0.037988353729248046, 0.037943294525146484, 0.03816755294799805, 0.03861503982543945, 0.03803340911865234, 0.037966846466064456, 0.03835084915161133, 0.03848396682739258, 0.03804159927368164, 0.03789926528930664, 0.038539264678955076, 0.038340606689453126, 0.03802521514892578, 0.03793203353881836, 0.038679550170898434, 0.03826892852783203, 0.07922994995117187, 0.03890790557861328, 0.03783679962158203, 0.03792281723022461, 0.03792588806152344, 0.03827507019042969, 0.03808153533935547, 0.03808768081665039, 0.03786137771606445, 0.03799552154541016, 0.0380313606262207, 0.03805388641357422, 0.03804569625854492, 0.03840409469604492, 0.03806412887573242, 0.037905406951904294, 0.03795455932617187, 0.03794636917114258, 0.03813478469848633, 0.03800985717773438, 0.03793203353881836, 0.03811123275756836, 0.038125568389892575, 0.03792793655395508, 0.03794739151000977, 0.03824435043334961, 0.0380847053527832, 0.0379257926940918, 0.03787161636352539, 0.037819393157958986, 0.03803647994995117, 0.037956607818603515, 0.039212032318115236, 0.03805184173583984, 0.0379504623413086, 0.0379156494140625, 0.04073574447631836, 0.038391807556152346, 0.03786649703979492, 0.03888947296142578, 0.0400711669921875, 0.03781222534179687, 0.03783679962158203, 0.037981182098388674, 0.03775078582763672, 0.03779891204833984, 0.03775590515136719, 0.0378419189453125, 0.03782144165039063, 0.037781505584716796, 0.03789311981201172, 0.03784396743774414, 0.03780710220336914, 0.038095870971679685, 0.037797889709472655, 0.03782451248168945, 0.03784601593017578, 0.03794124984741211, 0.037806079864501956, 0.037781505584716796, 0.037884929656982425, 0.03780710220336914, 0.03832627105712891, 0.07955046081542969, 0.03795251083374023, 0.03794947052001953, 0.03773539352416992, 0.03782860946655273, 0.03780710220336914, 0.037874687194824216, 0.037789695739746096, 0.037867584228515626, 0.03773331069946289, 0.038191104888916014, 0.03783679962158203, 0.037872638702392575, 0.03788185501098633, 0.037850112915039064, 0.0378869743347168, 0.03800371170043945, 0.037870655059814455, 0.038172607421875, 0.037921791076660154, 0.03788390350341797, 0.03808256149291992, 0.03796377563476563, 0.037819393157958986, 0.03782553482055664, 0.03784806442260742, 0.038128639221191404, 0.038128639221191404, 0.03777228927612305, 0.03794944000244141, 0.03797196960449219, 0.03817062377929688, 0.03783168029785156, 0.037855232238769534, 0.037928958892822266, 0.03805286407470703, 0.0382371826171875, 0.03802828979492188, 0.037988353729248046, 0.037961727142333986, 0.038179840087890625, 0.03786240005493164, 0.037978111267089845, 0.0379156494140625, 0.038059009552001956, 0.038112255096435545, 0.037910526275634765, 0.03792588806152344, 0.03811942291259766, 0.03815935897827148, 0.037978111267089845, 0.037874687194824216, 0.03809996795654297, 0.03806924819946289, 0.03789311981201172, 0.037894176483154296, 0.037800926208496094, 0.038100990295410156, 0.037956607818603515, 0.037953536987304685, 0.037889022827148434, 0.03792899322509766, 0.03823203277587891, 0.07917670440673828, 0.03791462326049805, 0.03787366485595703, 0.03793305587768555, 0.03789926528930664, 0.03786240005493164, 0.038117374420166016, 0.03861913681030273, 0.03791974258422852, 0.038007808685302735, 0.03791257476806641, 0.03793817520141601, 0.03802521514892578, 0.0379422721862793, 0.03785830307006836, 0.038046718597412106, 0.037959678649902344, 0.037958656311035156, 0.03790028762817383, 0.037817344665527344, 0.03794124984741211, 0.03805184173583984, 0.03784294509887695, 0.03805388641357422, 0.03809894561767578, 0.03805286407470703, 0.03784499359130859, 0.03791974258422852, 0.03803955078125, 0.03802828979492188, 0.038161407470703124, 0.03802214431762695, 0.037894142150878905, 0.03818598556518555, 0.037917697906494144, 0.03795455932617187, 0.037975040435791016, 0.03828224182128906, 0.038002689361572264, 0.037910526275634765, 0.03793100738525391, 0.03803852844238281, 0.03797708892822266, 0.03793612670898437, 0.0378869743347168, 0.0380313606262207, 0.03825971221923828, 0.037953536987304685, 0.03790643310546875, 0.038029312133789066, 0.03808153533935547, 0.03808870315551758, 0.0379791374206543, 0.03979673767089844, 0.03898777770996094, 0.03879116821289062, 0.037969921112060545, 0.03789926528930664, 0.03830579376220703, 0.038193153381347655, 0.038182910919189454, 0.03849523162841797, 0.03826278305053711, 0.08257740783691406, 0.03822489547729492, 0.03793920135498047, 0.03778559875488281, 0.03784499359130859, 0.037872638702392575, 0.038141952514648435, 0.037921791076660154, 0.03786342239379883, 0.03808256149291992, 0.03808051300048828, 0.0378419189453125, 0.03813273620605469, 0.038711296081542966, 0.038114303588867186, 0.03786240005493164, 0.03781119918823242, 0.03850137710571289, 0.03799347305297852, 0.0378152961730957, 0.0378869743347168, 0.037989376068115234, 0.03785728073120117, 0.037850112915039064, 0.0379607048034668, 0.03787782287597656, 0.03834259033203125, 0.03791462326049805, 0.03792588806152344, 0.03778047943115234, 0.038484992980957033, 0.03798015975952149, 0.03783782577514649, 0.03778559875488281, 0.03781017684936523, 0.0379791374206543, 0.03785625457763672, 0.03782553482055664, 0.0378603515625, 0.037923839569091795, 0.0378152961730957, 0.03780505752563477, 0.03792998504638672, 0.03794124984741211, 0.03778355026245117, 0.03781119918823242, 0.03775385665893555, 0.03779891204833984, 0.03803238296508789, 0.03786956787109375, 0.03816960144042969, 0.037814273834228515, 0.03785625457763672, 0.038043647766113284, 0.03777849578857422, 0.037824447631835935, 0.037713920593261716, 0.03772108840942383, 0.03790233612060547, 0.0377784309387207, 0.037776382446289065, 0.0378152961730957, 0.037921791076660154, 0.07919411468505859, 0.03782758331298828, 0.03773747253417969, 0.03775385665893555, 0.03801804733276367, 0.037800960540771485, 0.03785728073120117, 0.03783679962158203, 0.037719039916992186, 0.038024192810058595, 0.03775590515136719, 0.037768192291259765, 0.037823486328125, 0.037910526275634765, 0.03777433776855469, 0.03780505752563477, 0.03785113525390625, 0.0377968635559082, 0.037874687194824216, 0.03781222534179687, 0.03783475112915039, 0.03781119918823242, 0.037754878997802735, 0.03797094345092773, 0.03784499359130859, 0.03780710220336914, 0.03785728073120117, 0.03789209747314453, 0.0379791374206543, 0.03779891204833984, 0.03774771118164062, 0.03775078582763672, 0.03789926528930664, 0.03783987045288086, 0.037738529205322266, 0.037819358825683595, 0.03775795364379883, 0.037889022827148434, 0.037773311614990236, 0.03777740859985351, 0.03778662490844727, 0.03790950393676758, 0.03822079849243164, 0.04011008071899414, 0.03838054275512695, 0.0379576301574707, 0.03801599884033203, 0.03980595016479492, 0.03857920074462891, 0.03786956787109375, 0.03796384048461914, 0.03790534210205078, 0.03787059020996094, 0.037891071319580076, 0.037838848114013675, 0.03804774475097656, 0.03783270263671875, 0.03785113525390625, 0.038819839477539066, 0.038046718597412106, 0.03789823913574219, 0.03790233612060547, 0.03783679962158203, 0.08027750396728515, 0.038125568389892575, 0.03789619064331055, 0.03792588806152344, 0.03793100738525391, 0.03866726303100586, 0.03827916717529297, 0.03803443145751953, 0.03789619064331055, 0.03781222534179687, 0.03788185501098633, 0.03791257476806641, 0.03783065414428711, 0.03784505462646484, 0.03773846435546875, 0.03788899230957031, 0.03809280014038086, 0.03783475112915039, 0.03798323059082031, 0.03775385665893555, 0.037966846466064456, 0.038214656829833986, 0.03794124984741211, 0.03781119918823242, 0.037868545532226565, 0.03785113525390625, 0.03793100738525391, 0.0380211181640625, 0.03793817520141601, 0.037855232238769534, 0.03802828979492188, 0.0378081283569336, 0.03788185501098633, 0.03851264190673828, 0.03827097702026367, 0.03790643310546875, 0.03789619064331055, 0.03783065414428711, 0.037907455444335936, 0.03800985717773438, 0.03793920135498047, 0.03778867340087891, 0.03791872024536133, 0.037806079864501956, 0.03793920135498047, 0.03792793655395508, 0.037855232238769534, 0.03791974258422852, 0.03787161636352539, 0.03796582412719727, 0.038024192810058595, 0.03782860946655273, 0.03795251083374023, 0.03783782577514649, 0.037956607818603515, 0.03788083267211914, 0.03803964614868164, 0.03777833557128906, 0.037766143798828124, 0.03796275329589844, 0.03784703826904297, 0.03799859237670898, 0.03801804733276367, 0.07892582702636719, 0.03798527908325195, 0.037719039916992186, 0.03776921463012695, 0.03804467010498047, 0.037907455444335936, 0.03788288116455078, 0.03789516830444336, 0.03778867340087891, 0.037966846466064456, 0.03824844741821289, 0.03791155242919922, 0.03777228927612305, 0.03783065414428711, 0.03782758331298828, 0.03794739151000977, 0.03860070419311523, 0.0380948486328125, 0.037868545532226565, 0.03780505752563477, 0.0378152961730957, 0.037822463989257815, 0.03793407821655274, 0.03778771209716797, 0.03782342529296875, 0.037855232238769534, 0.03776921463012695, 0.038046718597412106, 0.037803009033203126, 0.03780198287963867, 0.037838848114013675, 0.03789516830444336, 0.03798015975952149, 0.0378603515625, 0.03887923049926758, 0.03785625457763672, 0.03788390350341797, 0.03847884750366211, 0.03800064086914062, 0.03786342239379883, 0.03780710220336914, 0.037826591491699216, 0.03794736099243164, 0.0378787841796875, 0.03783065414428711, 0.03801702499389648, 0.03784601593017578, 0.03804569625854492, 0.03787571334838867, 0.03788288116455078, 0.03784806442260742, 0.03797094345092773, 0.038019073486328124, 0.03792281723022461, 0.037956607818603515, 0.03784601593017578, 0.03790848159790039, 0.03802316665649414, 0.037868545532226565, 0.037868545532226565, 0.03931238555908203, 0.03824435043334961, 0.038171646118164065, 0.07896473693847657, 0.03796480178833008, 0.038031391143798825, 0.038429664611816405, 0.03807231903076172, 0.03814297485351562, 0.04003839874267578, 0.038354942321777344, 0.038171646118164065, 0.03779481506347656, 0.03799347305297852, 0.03781631851196289, 0.037822463989257815, 0.03803852844238281, 0.037795841217041014, 0.03802009582519531, 0.03805184173583984, 0.037812255859375, 0.038042591094970706, 0.0378787841796875, 0.03813785552978516, 0.03794636917114258, 0.038004737854003906, 0.03808153533935547, 0.03786240005493164, 0.03785318374633789, 0.03793817520141601, 0.0379156494140625, 0.03786137771606445, 0.03794124984741211, 0.037972991943359374, 0.03801599884033203, 0.03807743835449219, 0.038757377624511716, 0.038007808685302735, 0.03791155242919922, 0.03793407821655274, 0.03837849426269531, 0.038100990295410156, 0.037975040435791016, 0.037891071319580076, 0.03786751937866211, 0.03800678253173828, 0.03780505752563477, 0.03791360092163086, 0.037907455444335936, 0.03789209747314453, 0.03788288116455078, 0.037894142150878905, 0.03790335845947266, 0.0380497932434082, 0.0378787841796875, 0.0379504623413086, 0.03782963180541992, 0.037920768737792966, 0.03815731048583984, 0.03894476699829102, 0.03841228866577148, 0.0381102066040039, 0.03795455932617187, 0.03790848159790039, 0.03784396743774414, 0.03802521514892578]",tokens/s,25.828900065262776,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1253.777408,923.271168,0.0,276.824064,221.070848,s,10,0.22682729530334472,0.022682729530334473,0.0009842736621772984,0.022835423469543457,0.023319087219238283,0.02412570343017578,0.02477099639892578,"[0.024932319641113282, 0.02290608024597168, 0.02150435256958008, 0.021476160049438475, 0.02149331283569336, 0.022852031707763672, 0.02280348777770996, 0.02281881523132324, 0.022900896072387697, 0.023139839172363282]",tokens/s,11286.119673456473,kWh,2.6270753218803877e-07,1.439365579438648e-07,5.549522958133342e-07,9.615963859452379e-07,tokens/kWh,266223962.30031067,MB,1253.777408,923.271168,0.0,276.824064,242.38336,s,10,13.654691040039062,1.3654691040039062,0.033806523242069786,1.382609924316406,1.3859093383789063,1.389884356689453,1.3930643713378905,"[1.393859375, 1.3825091552734374, 1.2991619873046876, 1.2997125244140626, 1.36266357421875, 1.3833658447265624, 1.3850260009765625, 1.382710693359375, 1.383292236328125, 1.3823896484375]",tokens/s,46.13799009825108,kWh,1.6105208038547895e-05,8.825535893337172e-06,3.0158516719384448e-05,5.50892606512695e-05,tokens/kWh,1143598.5753885445,,s,629,13.832976406097409,0.02199201336422482,0.0027840527017240512,0.021901311874389647,0.02209239044189453,0.022377881622314452,0.04265828521728516,"[0.02269900894165039, 0.02235699272155762, 0.022237184524536133, 0.021989376068115234, 0.022571008682250978, 0.022978559494018554, 0.022577152252197266, 0.02215936088562012, 0.022391807556152343, 0.02284441566467285, 0.02265395164489746, 0.022572032928466795, 0.022537216186523438, 0.02263654327392578, 0.02216352081298828, 0.02219615936279297, 0.023150592803955077, 0.022541311264038084, 0.022089759826660157, 0.02138313674926758, 0.02108415985107422, 0.02107494354248047, 0.02275328063964844, 0.02267955207824707, 0.022188032150268554, 0.02208358383178711, 0.022177791595458983, 0.021926912307739257, 0.02185420799255371, 0.02188390350341797, 0.021849088668823242, 0.02187161636352539, 0.021998592376708984, 0.022008832931518556, 0.021964799880981444, 0.021884927749633788, 0.021950464248657226, 0.022023168563842774, 0.02191564750671387, 0.022055936813354493, 0.021914623260498048, 0.021881856918334962, 0.02187059211730957, 0.022091775894165038, 0.02186342430114746, 0.021960704803466798, 0.021902336120605468, 0.021934080123901366, 0.022013952255249023, 0.02190438461303711, 0.02185318374633789, 0.02185420799255371, 0.021981184005737304, 0.021910528182983398, 0.02167296028137207, 0.021799936294555664, 0.0218603515625, 0.021884927749633788, 0.022037504196166992, 0.021968896865844727, 0.0219289608001709, 0.021897216796875, 0.046129150390625, 0.021908479690551756, 0.02188697624206543, 0.021855295181274412, 0.021900224685668945, 0.02202828788757324, 0.02191257667541504, 0.021906431198120118, 0.02184499168395996, 0.02193715286254883, 0.0218470401763916, 0.02204569625854492, 0.021917695999145507, 0.021703680038452147, 0.021559295654296876, 0.021514240264892577, 0.02185318374633789, 0.021754880905151368, 0.0218787841796875, 0.021955583572387697, 0.021881856918334962, 0.021825536727905274, 0.021968896865844727, 0.02230169677734375, 0.022837247848510742, 0.022252544403076172, 0.02205695915222168, 0.02204569625854492, 0.021999616622924805, 0.021878816604614257, 0.021904352188110352, 0.02183782386779785, 0.021902336120605468, 0.021964799880981444, 0.02184499168395996, 0.02191257667541504, 0.021898239135742188, 0.02191564750671387, 0.02186649513244629, 0.021926912307739257, 0.021922815322875978, 0.0218470401763916, 0.02186956787109375, 0.021810176849365235, 0.022040576934814454, 0.021841920852661133, 0.021922815322875978, 0.021789695739746092, 0.021768192291259765, 0.021801984786987305, 0.021803007125854493, 0.021910528182983398, 0.0218787841796875, 0.021991424560546875, 0.021993471145629884, 0.021948415756225585, 0.02205695915222168, 0.021951488494873047, 0.022029312133789062, 0.02189619255065918, 0.021993471145629884, 0.02208153533935547, 0.02191155242919922, 0.042651649475097655, 0.02065100860595703, 0.020591615676879883, 0.02070528030395508, 0.020600896835327148, 0.020591552734375, 0.020617216110229493, 0.020668415069580077, 0.020610048294067384, 0.02066534423828125, 0.020610048294067384, 0.020642816543579103, 0.020700159072875975, 0.02067251205444336, 0.02062745666503906, 0.020634624481201173, 0.020692991256713866, 0.020722688674926756, 0.02064384078979492, 0.020572160720825194, 0.020576255798339844, 0.020716543197631835, 0.020610048294067384, 0.020599807739257812, 0.020624383926391602, 0.02062848091125488, 0.02066431999206543, 0.020674560546875, 0.020668479919433595, 0.02061414337158203, 0.020717504501342774, 0.02056604766845703, 0.020622304916381836, 0.020685823440551757, 0.02064076805114746, 0.020737024307250978, 0.02062131118774414, 0.020611072540283205, 0.020626432418823244, 0.020748287200927733, 0.020644863128662108, 0.020479999542236327, 0.020487167358398437, 0.020694015502929687, 0.020617216110229493, 0.020644863128662108, 0.020569087982177735, 0.020710399627685547, 0.020619264602661135, 0.02061414337158203, 0.020551679611206054, 0.020682815551757813, 0.020646848678588868, 0.02062950325012207, 0.020549631118774413, 0.02064896011352539, 0.02068992042541504, 0.02047590446472168, 0.020406272888183592, 0.02044211196899414, 0.020378623962402344, 0.020385791778564453, 0.020351999282836913, 0.04227276611328125, 0.02045337677001953, 0.020575231552124023, 0.02069606399536133, 0.020585472106933594, 0.020588544845581053, 0.02062848091125488, 0.020590591430664062, 0.020601856231689454, 0.02062848091125488, 0.02066431999206543, 0.020692991256713866, 0.020702239990234374, 0.020806623458862306, 0.020595712661743162, 0.020783103942871094, 0.02068172836303711, 0.020697120666503907, 0.02067043113708496, 0.020750335693359375, 0.0207258243560791, 0.020595647811889647, 0.02065715217590332, 0.020560895919799805, 0.0206376953125, 0.020537343978881836, 0.02063052749633789, 0.0206878719329834, 0.020546560287475587, 0.02066739273071289, 0.02064387130737305, 0.020584415435791016, 0.020590591430664062, 0.020554752349853517, 0.02080156707763672, 0.020672479629516603, 0.020684799194335936, 0.020611072540283205, 0.020504575729370117, 0.02066534423828125, 0.020704288482666016, 0.020639711380004883, 0.020595712661743162, 0.02042982482910156, 0.020374528884887694, 0.02046976089477539, 0.020548608779907225, 0.02061414337158203, 0.02062233543395996, 0.02068172836303711, 0.020651071548461915, 0.020612031936645507, 0.02057318305969238, 0.02059878349304199, 0.020533248901367186, 0.02067251205444336, 0.02063155174255371, 0.020593664169311524, 0.02065100860595703, 0.020666368484497072, 0.02066329574584961, 0.020625408172607423, 0.02083020782470703, 0.04266086578369141, 0.02060492706298828, 0.020632575988769532, 0.020582399368286132, 0.020577280044555665, 0.02058956718444824, 0.02064896011352539, 0.02061414337158203, 0.02065203285217285, 0.020633600234985353, 0.020725759506225586, 0.02064384078979492, 0.020574207305908202, 0.020518911361694335, 0.02066534423828125, 0.020592639923095703, 0.020756479263305663, 0.020632575988769532, 0.020969472885131835, 0.022564863204956053, 0.02230271911621094, 0.024410112380981445, 0.022129663467407225, 0.021884927749633788, 0.021947391510009767, 0.021910528182983398, 0.02185830307006836, 0.02205695915222168, 0.021790719985961913, 0.02192793655395508, 0.02187980842590332, 0.02187571144104004, 0.021978111267089845, 0.02191155242919922, 0.021849088668823242, 0.02188595199584961, 0.02186137580871582, 0.021966848373413086, 0.021966848373413086, 0.021916671752929686, 0.021924863815307616, 0.02186240005493164, 0.021974016189575195, 0.021975040435791016, 0.022004800796508787, 0.02199135971069336, 0.02186956787109375, 0.02185113525390625, 0.021974016189575195, 0.02191564750671387, 0.021918720245361328, 0.02194534492492676, 0.0220948486328125, 0.02224127960205078, 0.022979583740234375, 0.022125568389892578, 0.021780479431152345, 0.022021120071411132, 0.02188697624206543, 0.02186854362487793, 0.02191155242919922, 0.022004735946655272, 0.02189926338195801, 0.045532161712646485, 0.021926912307739257, 0.021946367263793946, 0.02205183982849121, 0.021918720245361328, 0.022195199966430663, 0.022026239395141603, 0.021938175201416017, 0.022000640869140626, 0.022005760192871093, 0.021943296432495117, 0.021961727142333985, 0.021897216796875, 0.02202828788757324, 0.021955583572387697, 0.021962751388549806, 0.021881919860839844, 0.021951423645019532, 0.021949440002441405, 0.02188595199584961, 0.021876735687255858, 0.021841920852661133, 0.02190028762817383, 0.021901311874389647, 0.02191974449157715, 0.02190438461303711, 0.022220800399780274, 0.022046720504760742, 0.021909503936767577, 0.021984256744384766, 0.02185215950012207, 0.021974016189575195, 0.0218787841796875, 0.02190336036682129, 0.0218603515625, 0.021952512741088868, 0.02187571144104004, 0.021916671752929686, 0.022009855270385743, 0.021939199447631837, 0.02188902473449707, 0.021925888061523437, 0.021996543884277343, 0.021987327575683592, 0.02204364776611328, 0.02207539176940918, 0.021956607818603514, 0.02191360092163086, 0.02188902473449707, 0.021914623260498048, 0.0219289608001709, 0.02185523223876953, 0.02202009582519531, 0.0219238395690918, 0.02191155242919922, 0.02183884811401367, 0.02185625648498535, 0.021795839309692384, 0.021934080123901366, 0.022091775894165038, 0.02205183982849121, 0.021910528182983398, 0.02191980743408203, 0.045168609619140626, 0.021933023452758788, 0.021878816604614257, 0.021850080490112306, 0.02191974449157715, 0.021975040435791016, 0.021996543884277343, 0.02204979133605957, 0.021960735321044922, 0.022155231475830078, 0.021994495391845705, 0.021870624542236327, 0.022069215774536133, 0.021966848373413086, 0.02191155242919922, 0.02210201644897461, 0.02186956787109375, 0.021831680297851562, 0.022005760192871093, 0.021857280731201172, 0.022000640869140626, 0.02190438461303711, 0.021966848373413086, 0.0218470401763916, 0.02187264060974121, 0.021964799880981444, 0.02188697624206543, 0.02187059211730957, 0.02191155242919922, 0.022090751647949217, 0.02186240005493164, 0.02270310401916504, 0.02314854431152344, 0.0222873592376709, 0.021965824127197265, 0.02186956787109375, 0.022054912567138672, 0.022203392028808593, 0.021961727142333985, 0.02188287925720215, 0.021989376068115234, 0.02185318374633789, 0.02182963180541992, 0.02201907157897949, 0.021866559982299805, 0.021827520370483397, 0.02188390350341797, 0.02188800048828125, 0.021884927749633788, 0.021976064682006836, 0.021831680297851562, 0.021873664855957032, 0.02226380729675293, 0.02209587287902832, 0.021943296432495117, 0.022006784439086914, 0.021968896865844727, 0.021770240783691407, 0.021836799621582033, 0.02186956787109375, 0.021893119812011717, 0.021974016189575195, 0.02186956787109375, 0.04530380630493164, 0.02186956787109375, 0.0217672004699707, 0.021849056243896485, 0.021957632064819335, 0.02191257667541504, 0.021857280731201172, 0.02186956787109375, 0.02183782386779785, 0.02186240005493164, 0.022030336380004883, 0.02189107131958008, 0.02183475112915039, 0.02191974449157715, 0.02191974449157715, 0.021987327575683592, 0.022199296951293947, 0.022040576934814454, 0.021940223693847655, 0.021961727142333985, 0.021812223434448243, 0.021916671752929686, 0.021925888061523437, 0.021931007385253908, 0.022156288146972656, 0.021982208251953125, 0.022746112823486327, 0.022005760192871093, 0.021990400314331054, 0.02185420799255371, 0.021906431198120118, 0.02190336036682129, 0.022071296691894532, 0.021950464248657226, 0.022021120071411132, 0.021984256744384766, 0.021839872360229492, 0.021901311874389647, 0.021894144058227538, 0.021940223693847655, 0.022039552688598633, 0.021929983139038087, 0.02185523223876953, 0.021953535079956055, 0.02188697624206543, 0.021906431198120118, 0.021873664855957032, 0.02183884811401367, 0.021995519638061522, 0.021951488494873047, 0.021901311874389647, 0.021942272186279296, 0.02191974449157715, 0.02184806442260742, 0.021835775375366212, 0.02180607986450195, 0.02210099220275879, 0.021956607818603514, 0.021843967437744142, 0.02208563232421875, 0.02186956787109375, 0.02182143974304199, 0.02187775993347168, 0.045279232025146485, 0.0219289608001709, 0.0218603515625, 0.02188697624206543, 0.021805055618286134, 0.021949440002441405, 0.02203545570373535, 0.02186342430114746, 0.021873664855957032, 0.021843967437744142, 0.021929983139038087, 0.021843967437744142, 0.021936128616333008, 0.022016000747680665, 0.021909503936767577, 0.02185523223876953, 0.02210304069519043, 0.021953535079956055, 0.021850112915039063, 0.0218470401763916, 0.02191155242919922, 0.021910528182983398, 0.02191974449157715, 0.02187775993347168, 0.02206822395324707, 0.02186854362487793, 0.021959680557250977, 0.02185318374633789, 0.02189926338195801, 0.02185113525390625, 0.02184294319152832, 0.02208153533935547, 0.021901311874389647, 0.02202828788757324, 0.021902336120605468, 0.02191257667541504, 0.02190028762817383, 0.02190336036682129, 0.022046720504760742, 0.02268569564819336, 0.022952959060668944, 0.022235136032104492, 0.021976064682006836, 0.021959680557250977, 0.02189926338195801, 0.022008832931518556, 0.021960704803466798, 0.02184601593017578, 0.021908479690551756, 0.02188083267211914, 0.021946367263793946, 0.021958656311035156, 0.022016000747680665, 0.021865503311157226, 0.02190230369567871, 0.02185830307006836, 0.02186342430114746, 0.021795839309692384, 0.021989376068115234, 0.02183782386779785, 0.02185318374633789, 0.02186342430114746, 0.02191974449157715, 0.04557926559448242, 0.02191257667541504, 0.021948415756225585, 0.021535743713378908, 0.021737472534179687, 0.021984256744384766, 0.021947391510009767, 0.021850112915039063, 0.021921791076660157, 0.02188287925720215, 0.02181427192687988, 0.021735424041748046, 0.022176767349243166, 0.02208563232421875, 0.0220897274017334, 0.021932031631469725, 0.021995519638061522, 0.021922815322875978, 0.022038528442382813, 0.02191257667541504, 0.021933055877685546, 0.021971967697143553, 0.022154239654541014, 0.02204569625854492, 0.0220948486328125, 0.021942272186279296, 0.0219238395690918, 0.02188390350341797, 0.02188595199584961, 0.022018047332763673, 0.02199350357055664, 0.022036447525024414, 0.02207744026184082, 0.021951488494873047, 0.021922815322875978, 0.0219289608001709, 0.02188697624206543, 0.021899328231811524, 0.021891008377075194, 0.02191155242919922, 0.021857280731201172, 0.022039552688598633, 0.021966880798339843, 0.022029279708862304, 0.021902336120605468, 0.021962751388549806, 0.021811199188232423, 0.02190336036682129, 0.021944320678710938, 0.021772287368774415, 0.021573631286621094, 0.02203647994995117, 0.021929983139038087, 0.021957632064819335, 0.02190540885925293, 0.02186854362487793, 0.021926912307739257, 0.021916671752929686, 0.02205388832092285, 0.02189107131958008, 0.022130687713623046, 0.021939199447631837, 0.021926912307739257]",tokens/s,45.47105276075973,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, @@ -7251,48 +7167,33 @@ ChildProcessError: Traceback (most recent call last): RuntimeError: shape '[-1, 128, 4672]' is invalid for input of size 21229568 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 86.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -7393,48 +7294,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2_moe,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,mixtral,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1354.99776,1313.34144,0.0,666.894336,583.411712,s,10,0.367520923614502,0.036752092361450196,0.0018078330313943103,0.03598340797424317,0.037524640655517576,0.03974108772277831,0.04151424537658691,"[0.04195753479003906, 0.03595008087158203, 0.035949630737304686, 0.0360167350769043, 0.035681278228759765, 0.0355464973449707, 0.03703209686279297, 0.0357196159362793, 0.036844062805175784, 0.03682339096069336]",tokens/s,6965.589808664122,kWh,4.4016995025242427e-07,2.4119189856989944e-07,1.376317705761802e-06,2.0576795545841257e-06,tokens/kWh,124411986.03041947,MB,1354.99776,1313.34144,0.0,666.894336,644.020224,s,10,22.070285400390624,2.2070285400390626,0.012889624369515166,2.207716430664062,2.217943408203125,2.2274182373046876,2.2349981005859374,"[2.215837890625, 2.19123291015625, 2.190115966796875, 2.2071044921875, 2.196080810546875, 2.203771240234375, 2.21104345703125, 2.209877197265625, 2.208328369140625, 2.23689306640625]",tokens/s,28.54516779329231,kWh,2.7174306134456744e-05,1.489238295772977e-05,7.386624248823808e-05,0.0001159329315804246,tokens/kWh,543417.6393296486,,s,629,22.359374866485588,0.035547495813172646,0.004369599305993048,0.03485388946533203,0.035935641479492186,0.03633192977905273,0.0706462094116211,"[0.035517440795898435, 0.03587686538696289, 0.035958782196044925, 0.036241409301757815, 0.03552358245849609, 0.03649740982055664, 0.036086784362792966, 0.035787841796875, 0.03611334228515625, 0.035814399719238284, 0.036670463562011715, 0.03558399963378906, 0.03552460861206055, 0.03667763137817383, 0.03713945770263672, 0.03656499099731445, 0.035566593170166014, 0.03501260757446289, 0.03496345520019531, 0.03650969696044922, 0.03527679824829102, 0.034767871856689454, 0.03477913665771484, 0.03477196884155274, 0.03492454528808594, 0.0348671989440918, 0.034900993347167966, 0.03502796936035156, 0.034513919830322266, 0.0346429443359375, 0.03457024002075195, 0.0347883529663086, 0.034560001373291016, 0.03474943923950195, 0.03465420913696289, 0.034563072204589845, 0.03450982284545898, 0.0343111686706543, 0.034579456329345705, 0.03442483139038086, 0.034683902740478514, 0.03441356658935547, 0.034772991180419925, 0.034407424926757815, 0.035517440795898435, 0.03611238479614258, 0.03475046539306641, 0.034597888946533206, 0.034735103607177735, 0.03493081665039063, 0.03529203033447265, 0.0354785270690918, 0.034852863311767575, 0.034748416900634765, 0.03462963104248047, 0.034566143035888675, 0.03419750213623047, 0.03429683303833008, 0.03449958419799805, 0.03483647918701172, 0.034385921478271485, 0.034536449432373044, 0.07078399658203124, 0.03447091293334961, 0.03465625762939453, 0.03450470352172851, 0.034754558563232424, 0.03484467315673828, 0.03481702423095703, 0.03438694381713867, 0.034683902740478514, 0.03473715209960938, 0.03455692672729492, 0.03723980712890625, 0.03523583984375, 0.035286014556884765, 0.034473983764648435, 0.0345456657409668, 0.034359294891357424, 0.034477054595947264, 0.03466239929199219, 0.03511808013916016, 0.03498291015625, 0.034813953399658204, 0.03462662506103516, 0.03509958267211914, 0.03489996719360351, 0.03486207962036133, 0.034920448303222655, 0.034597888946533206, 0.03445862579345703, 0.03462246322631836, 0.0344268798828125, 0.03480678558349609, 0.03458969497680664, 0.034900993347167966, 0.034776065826416014, 0.034786304473876956, 0.034699264526367186, 0.03467366409301758, 0.03476889419555664, 0.03478220748901367, 0.03483750534057617, 0.034511871337890625, 0.03454771041870117, 0.03439513778686523, 0.034931713104248044, 0.03477811050415039, 0.034948097229003904, 0.034587646484375, 0.03447296142578125, 0.03463065719604492, 0.034923519134521484, 0.035004417419433595, 0.0352174072265625, 0.03488460922241211, 0.034713600158691404, 0.034785472869873046, 0.03485164642333984, 0.034582527160644534, 0.03461324691772461, 0.034735103607177735, 0.03479142379760742, 0.03497881698608398, 0.03479654312133789, 0.07093452453613282, 0.03465216064453125, 0.0348221435546875, 0.03531161499023437, 0.03479449462890625, 0.034871295928955076, 0.03495219039916992, 0.034723838806152346, 0.03477401733398437, 0.0347064323425293, 0.03492454528808594, 0.034536449432373044, 0.03510784149169922, 0.03465420913696289, 0.03480684661865235, 0.03470124816894531, 0.03458371353149414, 0.03471139144897461, 0.03486003112792969, 0.034661376953125, 0.03493798446655273, 0.03478515243530273, 0.034732032775878906, 0.03485184097290039, 0.03488870239257812, 0.03453235244750977, 0.03489382553100586, 0.03434188842773438, 0.034872318267822264, 0.03492454528808594, 0.034148353576660156, 0.03449651336669922, 0.034813953399658204, 0.03492966461181641, 0.034841598510742186, 0.03460300827026367, 0.034508800506591795, 0.03467264175415039, 0.03484262466430664, 0.034628608703613284, 0.03479859161376953, 0.03498700714111328, 0.034651294708251956, 0.03462236785888672, 0.034640830993652345, 0.03490611267089844, 0.03469823837280273, 0.034902015686035154, 0.034928638458251955, 0.034651294708251956, 0.03462025451660156, 0.03465932846069336, 0.03507712173461914, 0.03481804656982422, 0.03491020965576172, 0.03464089584350586, 0.03476275253295898, 0.035087390899658205, 0.034976734161376956, 0.03449958419799805, 0.03471257781982422, 0.034669567108154296, 0.0345272331237793, 0.071478271484375, 0.034956287384033204, 0.034609153747558595, 0.036604927062988284, 0.037515262603759765, 0.035896320343017575, 0.03503411102294922, 0.03479347229003906, 0.03482828903198242, 0.03454873657226563, 0.03466342544555664, 0.03487334442138672, 0.03464396667480469, 0.034988033294677735, 0.03512319946289062, 0.034746368408203124, 0.03474124908447265, 0.034568286895751955, 0.034476959228515625, 0.03470963287353516, 0.03468991851806641, 0.03481190490722656, 0.034735103607177735, 0.03486617660522461, 0.03487744140625, 0.034931713104248044, 0.035102718353271486, 0.034764801025390625, 0.034582527160644534, 0.03495219039916992, 0.03488870239257812, 0.03491328048706055, 0.03520204925537109, 0.0358809585571289, 0.03488051223754883, 0.03489996719360351, 0.034969600677490234, 0.03485081481933594, 0.03455187225341797, 0.03485996627807617, 0.03504127883911133, 0.036961280822753906, 0.03565055847167969, 0.0350832633972168, 0.034813953399658204, 0.034816001892089846, 0.034685951232910156, 0.034516990661621096, 0.03474534225463867, 0.03615027236938476, 0.03591372680664062, 0.03477196884155274, 0.034917377471923826, 0.034800640106201174, 0.03520204925537109, 0.03498700714111328, 0.03488051223754883, 0.03541401672363281, 0.034797569274902344, 0.034551807403564457, 0.03476377487182617, 0.03484262466430664, 0.03485081481933594, 0.07039282989501953, 0.03467059326171875, 0.03469619369506836, 0.034702335357666016, 0.0346060791015625, 0.0348221435546875, 0.03482217788696289, 0.03479548645019531, 0.03466649627685547, 0.034697216033935545, 0.03539763259887695, 0.035337215423583986, 0.034797569274902344, 0.03480166244506836, 0.03540070343017578, 0.03477503967285156, 0.03475763320922851, 0.034759681701660154, 0.03535257720947266, 0.03613798522949219, 0.035148799896240236, 0.034947071075439456, 0.03492659378051758, 0.03488665771484375, 0.03466239929199219, 0.034715648651123046, 0.034902015686035154, 0.03509145736694336, 0.034800640106201174, 0.03488358306884766, 0.03472281646728516, 0.034729984283447264, 0.03480473709106445, 0.03463679885864258, 0.035009536743164066, 0.03467366409301758, 0.03496448135375976, 0.034678783416748044, 0.034909183502197266, 0.034754558563232424, 0.034869312286376956, 0.03521017456054688, 0.03491123199462891, 0.03483443069458008, 0.034644992828369144, 0.034625537872314455, 0.03480883026123047, 0.035149822235107424, 0.03476070404052734, 0.034825214385986326, 0.035009536743164066, 0.03492454528808594, 0.034746368408203124, 0.034784255981445314, 0.03487948989868164, 0.03450368118286133, 0.03453459167480469, 0.03429254531860351, 0.03462758255004883, 0.034925567626953126, 0.03499724960327148, 0.03464908981323242, 0.035037185668945314, 0.07064678192138672, 0.03441459274291992, 0.034694145202636716, 0.03461734390258789, 0.03445248031616211, 0.03491952133178711, 0.03476166534423828, 0.0344227523803711, 0.034531326293945314, 0.03464191818237305, 0.03439820861816406, 0.03489894485473633, 0.03634380722045898, 0.03576115036010742, 0.03449446487426758, 0.03520307159423828, 0.03449139022827148, 0.03478937530517578, 0.035342334747314456, 0.035163135528564454, 0.034618400573730466, 0.034358238220214846, 0.034702335357666016, 0.03500032043457031, 0.034753631591796875, 0.0347421760559082, 0.03495423889160156, 0.035460094451904296, 0.03466342544555664, 0.03442483139038086, 0.03458457565307617, 0.03460505676269531, 0.03462041473388672, 0.03481292724609375, 0.03470131301879883, 0.03467264175415039, 0.03499728012084961, 0.03446780776977539, 0.03447609710693359, 0.03452105712890625, 0.03455382537841797, 0.034874366760253905, 0.0379422721862793, 0.03952230453491211, 0.036410369873046876, 0.03606425476074219, 0.035922943115234376, 0.03479859161376953, 0.03526860809326172, 0.035707904815673826, 0.034740222930908206, 0.03471155166625976, 0.03473920059204102, 0.034819072723388675, 0.034661376953125, 0.0346951675415039, 0.03463270568847656, 0.034637889862060546, 0.03448416137695313, 0.03467366409301758, 0.034648063659667966, 0.03489279937744141, 0.03479865646362305, 0.07351289367675781, 0.03600896072387695, 0.035197952270507815, 0.03555123138427734, 0.034988094329833984, 0.03467871856689453, 0.03475763320922851, 0.03546121597290039, 0.034708385467529294, 0.03477196884155274, 0.03465932846069336, 0.03476486587524414, 0.03479750442504883, 0.03556249618530274, 0.03582156753540039, 0.03582880020141602, 0.035523521423339845, 0.03475251388549805, 0.03458358383178711, 0.034864097595214846, 0.035884033203125, 0.03502489471435547, 0.03483955383300781, 0.03487334442138672, 0.03500543975830078, 0.035759105682373046, 0.03461324691772461, 0.03523481750488281, 0.034956287384033204, 0.03481190490722656, 0.03480883026123047, 0.034560001373291016, 0.03494911956787109, 0.03489382553100586, 0.03491635131835937, 0.03486617660522461, 0.035081214904785156, 0.034776065826416014, 0.034514942169189454, 0.034531326293945314, 0.03461734390258789, 0.03551846313476562, 0.035914752960205076, 0.035579902648925785, 0.03478220748901367, 0.035248126983642575, 0.03557580947875977, 0.034890750885009765, 0.034685951232910156, 0.0346429443359375, 0.03615334320068359, 0.035279006958007814, 0.034461536407470704, 0.03487846374511719, 0.03565465545654297, 0.03600384140014649, 0.03564851379394531, 0.03565478515625, 0.03515276718139648, 0.03466547012329101, 0.03437875366210937, 0.03461750411987305, 0.0344431037902832, 0.07064473724365235, 0.0347147216796875, 0.034790302276611326, 0.03484377670288086, 0.03461516952514648, 0.034869247436523435, 0.03594137573242188, 0.034958335876464845, 0.03640115356445312, 0.03510476684570313, 0.03486105728149414, 0.03501055908203125, 0.03485388946533203, 0.034310142517089845, 0.03443916702270508, 0.03494911956787109, 0.03459481430053711, 0.03470438385009766, 0.03486617660522461, 0.034344959259033206, 0.03521843338012695, 0.03473715209960938, 0.03472588729858399, 0.03449446487426758, 0.034948097229003904, 0.03617996978759765, 0.03462656021118164, 0.0347064323425293, 0.034874366760253905, 0.03575807952880859, 0.03489177703857422, 0.03488051223754883, 0.034956287384033204, 0.03717324829101563, 0.036507713317871095, 0.03579283142089844, 0.03566592025756836, 0.035716094970703126, 0.03486617660522461, 0.034557952880859374, 0.034680831909179685, 0.035053569793701174, 0.03573657608032227, 0.03503923034667969, 0.03496857452392578, 0.03485184097290039, 0.035326976776123044, 0.03513753509521484, 0.03506380844116211, 0.036201473236083984, 0.034890750885009765, 0.034939903259277344, 0.034902015686035154, 0.035476478576660156, 0.0344002571105957, 0.03499008178710938, 0.03501260757446289, 0.03476688003540039, 0.03490915298461914, 0.0347770881652832, 0.034776065826416014, 0.034857982635498046, 0.03577036666870117, 0.07236412811279297, 0.03441961669921875, 0.034358272552490236, 0.03481497573852539, 0.03492659378051758, 0.0358287353515625, 0.0354703369140625, 0.035351551055908204, 0.03686297607421875, 0.03505152130126953, 0.03466239929199219, 0.0353361930847168, 0.035487743377685545, 0.035620864868164064, 0.03545600128173828, 0.03588608169555664, 0.034941951751708986, 0.03483340835571289, 0.034923519134521484, 0.03481087875366211, 0.034938880920410156, 0.03464089584350586, 0.03494911956787109, 0.034764801025390625, 0.03468288040161133, 0.03474537658691406, 0.034764766693115234, 0.035383296966552735, 0.03480678558349609, 0.03476377487182617, 0.034338878631591796, 0.03467871856689453, 0.03476377487182617, 0.034759681701660154, 0.03488870239257812, 0.034748416900634765, 0.03457740783691406, 0.034948097229003904, 0.03500339126586914, 0.03466035079956055, 0.034597888946533206, 0.03524505615234375, 0.03553484725952148, 0.03542835235595703, 0.03479244613647461, 0.03461529541015625, 0.035593215942382815, 0.03561065673828125, 0.03489891052246094, 0.035195903778076174, 0.03493580627441406, 0.034948097229003904, 0.034941951751708986, 0.03497267150878906, 0.034748416900634765, 0.035342334747314456, 0.036103168487548826, 0.03496550369262695, 0.035093505859375, 0.03604991912841797, 0.034797569274902344, 0.03469823837280273, 0.03488256072998047, 0.07262105560302734, 0.03481190490722656, 0.034570270538330075, 0.03525526428222656, 0.03586457443237305, 0.03491839981079101, 0.03548262405395508, 0.03534131240844727, 0.034533374786376955, 0.03462041473388672, 0.03450582504272461, 0.03459779357910156, 0.03499008178710938, 0.034854911804199216, 0.03489791870117188, 0.034700286865234374, 0.03567718505859375, 0.03510067367553711, 0.034855934143066404, 0.03494604873657227, 0.035383296966552735, 0.03561471939086914, 0.035422206878662106, 0.03482624053955078, 0.03482931137084961, 0.03605408096313477, 0.03592902374267578, 0.035507198333740234, 0.03497369766235352, 0.03558911895751953, 0.03528806304931641, 0.035912704467773435, 0.03631411361694336, 0.03609292984008789, 0.03615129470825195, 0.03599871826171875, 0.035932159423828124, 0.035958782196044925, 0.035934207916259765, 0.03591987228393555, 0.03645542526245117, 0.03650252914428711, 0.03688243103027344, 0.03597107315063477, 0.03595161437988281, 0.035606529235839846, 0.035883007049560545, 0.03639807891845703, 0.035996673583984375, 0.0357314567565918, 0.03626803207397461, 0.0361451530456543, 0.03608473587036133, 0.03609190368652344, 0.0363765754699707, 0.03603148651123047, 0.03598643112182617, 0.03509964752197266, 0.03460403060913086, 0.03459686279296875, 0.034941951751708986, 0.03470131301879883, 0.03467366409301758]",tokens/s,28.131376827659274,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -7987,48 +7873,33 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1271.554048,1017.643008,0.0,371.195904,315.446272,s,10,0.19529705619812013,0.019529705619812012,0.0009140443366206495,0.019363104820251463,0.021000485801696778,0.021201970767974854,0.021363158740997316,"[0.020955711364746093, 0.018462751388549806, 0.01938128089904785, 0.019328096389770507, 0.02140345573425293, 0.018568992614746094, 0.018725887298583984, 0.01934492874145508, 0.019606687545776366, 0.019519264221191407]",tokens/s,13108.236497958242,kWh,2.261678843151771e-07,1.2392898202703704e-07,5.473570344435316e-07,8.974539007857457e-07,tokens/kWh,285251420.4638978,MB,1271.554048,1017.643008,0.0,371.195904,340.902912,s,10,11.830128784179688,1.1830128784179688,0.010220015064036666,1.1847022705078125,1.1948863647460937,1.1950671447753907,1.1952117687988282,"[1.1664869384765626, 1.1741630859375, 1.19484619140625, 1.1935650634765624, 1.1794437255859376, 1.188765869140625, 1.1682054443359375, 1.1879569091796875, 1.1814476318359375, 1.1952479248046874]",tokens/s,53.25385813571976,kWh,1.3598944398914649e-05,7.451885775429064e-06,2.9480500321153233e-05,5.0531330495496944e-05,tokens/kWh,1246751.2606978395,,s,629,11.982314514160153,0.019049784601208517,0.002366055636612425,0.01886207962036133,0.019090431213378906,0.01922559928894043,0.037088911590576173,"[0.020017152786254884, 0.01830297660827637, 0.01788211250305176, 0.017999872207641602, 0.017724416732788087, 0.017986560821533205, 0.01797324752807617, 0.017954816818237306, 0.018149375915527344, 0.01821286392211914, 0.017949695587158202, 0.01780019187927246, 0.017811519622802734, 0.01791276741027832, 0.01782374382019043, 0.018069503784179687, 0.018145280838012694, 0.018121728897094725, 0.017817728042602538, 0.017745792388916017, 0.017811456680297853, 0.017704959869384765, 0.017649663925170898, 0.01785139274597168, 0.017945600509643556, 0.017889280319213868, 0.017904640197753906, 0.017810432434082032, 0.017958911895751953, 0.01818009567260742, 0.018121728897094725, 0.018191360473632814, 0.01867366409301758, 0.019141664505004884, 0.019366880416870118, 0.01925324821472168, 0.019109952926635743, 0.018839487075805662, 0.018568191528320312, 0.019009536743164062, 0.018899967193603515, 0.01883647918701172, 0.0190382080078125, 0.01876479911804199, 0.018717695236206054, 0.018765823364257812, 0.01883545684814453, 0.018886688232421876, 0.018552799224853516, 0.01900748825073242, 0.018927616119384767, 0.018916351318359375, 0.018776063919067384, 0.018911231994628908, 0.018926591873168946, 0.019090431213378906, 0.018936832427978514, 0.018993152618408202, 0.018960384368896483, 0.018997247695922852, 0.019090431213378906, 0.018817119598388672, 0.03665296173095703, 0.017954816818237306, 0.018086912155151368, 0.01783193588256836, 0.01798860740661621, 0.018092031478881835, 0.018083839416503905, 0.018058240890502928, 0.018153472900390624, 0.017795072555541993, 0.0180316162109375, 0.018076671600341796, 0.018059263229370116, 0.01798860740661621, 0.01844428825378418, 0.01783500862121582, 0.01799065589904785, 0.018164863586425783, 0.019538816452026368, 0.019020927429199218, 0.018961280822753907, 0.01883750343322754, 0.01900851249694824, 0.01863577651977539, 0.018566144943237304, 0.018934783935546876, 0.018930688858032226, 0.01903104019165039, 0.019366912841796875, 0.019103744506835937, 0.01902592086791992, 0.018909183502197266, 0.018679807662963867, 0.018682880401611326, 0.01884876823425293, 0.01885593605041504, 0.018932735443115235, 0.018775039672851563, 0.018779136657714843, 0.01899929618835449, 0.018562047958374024, 0.018780160903930664, 0.01865830421447754, 0.018711551666259766, 0.018930688858032226, 0.018962560653686525, 0.018713472366333007, 0.018715648651123046, 0.018820127487182616, 0.01867670440673828, 0.018869247436523438, 0.018967552185058592, 0.0190248966217041, 0.01882316780090332, 0.01884979248046875, 0.018946048736572265, 0.018947071075439453, 0.018792448043823243, 0.018743295669555664, 0.018677759170532226, 0.018784255981445314, 0.019042303085327148, 0.018932895660400392, 0.03863228988647461, 0.01879654312133789, 0.019182592391967773, 0.01887948799133301, 0.01880985641479492, 0.018988031387329102, 0.018840576171875, 0.018990079879760743, 0.01983897590637207, 0.01920102310180664, 0.019095552444458007, 0.019406848907470704, 0.019109888076782225, 0.018985984802246093, 0.01922559928894043, 0.018942975997924806, 0.01883443260192871, 0.019072000503540038, 0.019217439651489258, 0.019035104751586915, 0.018884607315063476, 0.01901260757446289, 0.018982912063598634, 0.01926976013183594, 0.018795391082763672, 0.01906380844116211, 0.018922496795654296, 0.018951168060302736, 0.018964479446411133, 0.018975744247436522, 0.01904742431640625, 0.018957311630249024, 0.018982912063598634, 0.01903206443786621, 0.01886617660522461, 0.01887948799133301, 0.01885491180419922, 0.01879654312133789, 0.018924543380737305, 0.018731008529663085, 0.018837663650512697, 0.018797407150268553, 0.01885081672668457, 0.018994176864624023, 0.01904025650024414, 0.01883750343322754, 0.018844671249389648, 0.01892777633666992, 0.018924383163452147, 0.019001344680786132, 0.01899212837219238, 0.019002368927001953, 0.01883852767944336, 0.018927616119384767, 0.0189040641784668, 0.018844671249389648, 0.01922777557373047, 0.018747264862060548, 0.01879756736755371, 0.01862860870361328, 0.01885491180419922, 0.018626560211181642, 0.018851839065551757, 0.03850870513916015, 0.018795360565185548, 0.01887027168273926, 0.018887744903564454, 0.018889663696289062, 0.018948095321655273, 0.01903206443786621, 0.01909350395202637, 0.01902592086791992, 0.018903039932250978, 0.019079168319702147, 0.019013631820678712, 0.0190515193939209, 0.018990079879760743, 0.01904640007019043, 0.019207296371459962, 0.01885683250427246, 0.018952192306518553, 0.018957311630249024, 0.019281919479370118, 0.019170303344726563, 0.01906483268737793, 0.018900064468383788, 0.018952096939086914, 0.01886310386657715, 0.01868083190917969, 0.01889587211608887, 0.018905088424682616, 0.018905088424682616, 0.018985984802246093, 0.018840576171875, 0.01866035270690918, 0.018889728546142577, 0.01878118324279785, 0.018689023971557618, 0.018696191787719727, 0.018860031127929687, 0.01886207962036133, 0.0188272647857666, 0.01899519920349121, 0.018921472549438476, 0.018953216552734374, 0.018710687637329103, 0.01884553527832031, 0.018985984802246093, 0.019359743118286133, 0.018723840713500976, 0.018914304733276367, 0.01888768005371094, 0.019006464004516603, 0.01885491180419922, 0.018954240798950195, 0.018852863311767578, 0.01925734329223633, 0.019005439758300782, 0.019078144073486326, 0.018884607315063476, 0.018915327072143554, 0.018993152618408202, 0.018742271423339844, 0.019376127243041993, 0.0190515193939209, 0.01882316780090332, 0.04119039916992188, 0.018949119567871094, 0.019070976257324217, 0.018922496795654296, 0.018876415252685547, 0.019170303344726563, 0.019082239151000976, 0.018942975997924806, 0.018954240798950195, 0.018775039672851563, 0.018718719482421875, 0.018762752532958983, 0.018670591354370117, 0.018731039047241212, 0.019002336502075196, 0.018882560729980468, 0.018729984283447267, 0.01904947280883789, 0.018869247436523438, 0.018756704330444338, 0.0187554874420166, 0.018935808181762694, 0.01886617660522461, 0.018966527938842775, 0.018799615859985352, 0.01905766487121582, 0.018744319915771485, 0.01900851249694824, 0.019042303085327148, 0.018704383850097657, 0.018779136657714843, 0.018898944854736328, 0.018880512237548826, 0.019350528717041016, 0.018659328460693358, 0.018634752273559572, 0.018785280227661134, 0.018497535705566406, 0.018531360626220704, 0.018791391372680665, 0.018892799377441406, 0.019111936569213867, 0.018805856704711913, 0.01858550453186035, 0.018798656463623047, 0.018900928497314454, 0.01904844856262207, 0.018962432861328125, 0.018732032775878905, 0.01922559928894043, 0.018960384368896483, 0.018158592224121094, 0.017957887649536132, 0.017901632308959962, 0.017839040756225586, 0.017967103958129883, 0.0178606071472168, 0.01805619239807129, 0.018120704650878908, 0.018111488342285157, 0.018066432952880858, 0.017943552017211914, 0.01808076858520508, 0.036969470977783206, 0.018720767974853517, 0.018741312026977538, 0.018897855758666992, 0.018877439498901367, 0.018736127853393555, 0.01883955192565918, 0.018892799377441406, 0.019025951385498046, 0.01890915107727051, 0.01892972755432129, 0.018944992065429687, 0.018826208114624023, 0.018890752792358398, 0.018911231994628908, 0.018881536483764647, 0.018629728317260744, 0.018756511688232422, 0.018934783935546876, 0.018855968475341798, 0.018990047454833986, 0.01881907272338867, 0.019137535095214844, 0.01884569549560547, 0.01884160041809082, 0.018712608337402344, 0.018834400177001952, 0.018743295669555664, 0.018499584197998048, 0.01885491180419922, 0.018917375564575196, 0.018942975997924806, 0.01904640007019043, 0.018973695755004884, 0.018658336639404298, 0.01889686393737793, 0.018727935791015626, 0.019357696533203125, 0.019466239929199217, 0.01920614433288574, 0.01903001594543457, 0.018939903259277344, 0.01880268859863281, 0.01885081672668457, 0.01865727996826172, 0.0187064323425293, 0.01901568031311035, 0.018894847869873048, 0.01869004821777344, 0.01888768005371094, 0.018707456588745116, 0.018958335876464845, 0.01881507110595703, 0.018725791931152345, 0.018954240798950195, 0.018803712844848632, 0.018970624923706055, 0.018896896362304686, 0.01886617660522461, 0.018924575805664062, 0.018801631927490233, 0.01897881507873535, 0.01885081672668457, 0.03713536071777344, 0.01801420783996582, 0.01782681655883789, 0.01780121612548828, 0.017937408447265626, 0.01788313674926758, 0.018086912155151368, 0.017979488372802735, 0.018089887619018554, 0.01799679946899414, 0.01784217643737793, 0.01804800033569336, 0.018077695846557617, 0.018420736312866212, 0.01813811111450195, 0.018081792831420897, 0.01803264045715332, 0.017977344512939454, 0.01801215934753418, 0.017757247924804688, 0.01795167922973633, 0.01799679946899414, 0.01804083251953125, 0.01806438446044922, 0.01840640068054199, 0.018937856674194335, 0.01873523139953613, 0.018714591979980467, 0.0187554874420166, 0.0188723201751709, 0.018910207748413087, 0.01945292854309082, 0.01900032043457031, 0.018687999725341797, 0.018775039672851563, 0.01882111930847168, 0.018571264266967775, 0.01883558464050293, 0.01885580825805664, 0.018915327072143554, 0.018856960296630858, 0.018940959930419922, 0.01891119956970215, 0.018887712478637696, 0.01914569664001465, 0.01885491180419922, 0.01862246322631836, 0.01860403251647949, 0.018769920349121092, 0.018704383850097657, 0.01893587112426758, 0.019063743591308593, 0.018961408615112304, 0.019014656066894533, 0.018930688858032226, 0.01883033561706543, 0.018955263137817382, 0.018968576431274413, 0.01900748825073242, 0.018897920608520507, 0.019014656066894533, 0.018824192047119142, 0.01883135986328125, 0.038749183654785156, 0.01903923225402832, 0.01925433540344238, 0.018804672241210938, 0.01879449653625488, 0.018709503173828124, 0.018998271942138673, 0.0188538875579834, 0.01887846374511719, 0.018940927505493164, 0.018928640365600585, 0.0188221435546875, 0.018743295669555664, 0.01880575942993164, 0.01882019233703613, 0.019021728515625, 0.01899929618835449, 0.018715648651123046, 0.019109888076782225, 0.01904640007019043, 0.01886412811279297, 0.01902796745300293, 0.01884671974182129, 0.018951168060302736, 0.01845248031616211, 0.01866649627685547, 0.018948095321655273, 0.0188221435546875, 0.018982912063598634, 0.018701311111450195, 0.018494464874267577, 0.018754560470581053, 0.018967552185058592, 0.018783231735229493, 0.018856960296630858, 0.01887539291381836, 0.018757696151733397, 0.018967487335205077, 0.018883583068847656, 0.01875660705566406, 0.018545663833618165, 0.01883647918701172, 0.01883135986328125, 0.01861631965637207, 0.018696191787719727, 0.018667520523071288, 0.018711551666259766, 0.018933759689331055, 0.018709503173828124, 0.018785280227661134, 0.01859891128540039, 0.018540576934814455, 0.01872483253479004, 0.01882521629333496, 0.018964479446411133, 0.0194652156829834, 0.018729984283447267, 0.019041311264038085, 0.01892348861694336, 0.01902592086791992, 0.018899967193603515, 0.018945024490356444, 0.01891744041442871, 0.039, 0.01886207962036133, 0.018724863052368163, 0.018851839065551757, 0.01878835105895996, 0.018738208770751955, 0.01889481544494629, 0.018662399291992187, 0.01781760025024414, 0.017797119140625, 0.017918975830078124, 0.017838079452514647, 0.01800601577758789, 0.018058240890502928, 0.017975296020507812, 0.018059263229370116, 0.01868815994262695, 0.01871446418762207, 0.0186562557220459, 0.01878953552246094, 0.01872265625, 0.018795551300048827, 0.018784223556518556, 0.01888768005371094, 0.018791423797607423, 0.018795520782470702, 0.01904844856262207, 0.0188221435546875, 0.018909183502197266, 0.01878835105895996, 0.01879654312133789, 0.0188723201751709, 0.018998271942138673, 0.01898806381225586, 0.01884873580932617, 0.018703359603881836, 0.018964479446411133, 0.019092479705810548, 0.01896361541748047, 0.01868988800048828, 0.018832384109497072, 0.018837535858154297, 0.018735200881958007, 0.018717567443847657, 0.019216384887695313, 0.01887948799133301, 0.018950143814086915, 0.018972671508789063, 0.019096576690673828, 0.01882009506225586, 0.018949119567871094, 0.01900851249694824, 0.018808832168579103, 0.018893823623657227, 0.018735103607177735, 0.018786304473876952, 0.018913280487060546, 0.018896896362304686, 0.019161088943481445, 0.01902796745300293, 0.01904435157775879, 0.019018943786621095, 0.018677568435668944, 0.038816768646240236, 0.018958335876464845, 0.019017728805541992, 0.018979839324951172, 0.01882521629333496, 0.018888704299926756, 0.01883340835571289, 0.01893174362182617, 0.018801631927490233, 0.018751487731933594, 0.018884607315063476, 0.01901875114440918, 0.01905254364013672, 0.018852863311767578, 0.01902796745300293, 0.01903104019165039, 0.019095552444458007, 0.018968576431274413, 0.019164159774780275, 0.018955263137817382, 0.018951168060302736, 0.018985984802246093, 0.01905254364013672, 0.018993152618408202, 0.018977792739868164, 0.01875660705566406, 0.018876415252685547, 0.018720895767211913, 0.01888140869140625, 0.018778112411499022, 0.018736127853393555, 0.018723840713500976, 0.018611200332641603, 0.019019775390625, 0.019397632598876953, 0.019515392303466796, 0.01984102439880371, 0.01907721519470215, 0.01902275276184082, 0.019108928680419923, 0.018827199935913086, 0.018966527938842775, 0.01899929618835449, 0.018892799377441406, 0.01909760093688965, 0.01885593605041504, 0.018971647262573242, 0.019215360641479492, 0.018757631301879883, 0.019134464263916014, 0.01903923225402832, 0.019009536743164062, 0.01903308868408203, 0.019096576690673828, 0.01880985641479492, 0.018685951232910156, 0.01884671974182129, 0.01868499183654785, 0.019015615463256835, 0.01898700714111328, 0.018881599426269532, 0.018994112014770508, 0.019103744506835937]",tokens/s,52.49403187144489,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,3768.549376,12218.53184,0.0,11572.084736,11067.830272,s,10,6.234430114746093,0.6234430114746093,0.0012077956402805968,0.6233088073730468,0.6253229248046874,0.6256142822265625,0.6258473681640625,"[0.6225289306640625, 0.6219365234375, 0.6234368896484375, 0.6232257080078125, 0.6233919067382813, 0.6224734497070312, 0.623823974609375, 0.6252581787109375, 0.6224489135742187, 0.6259056396484375]",tokens/s,410.6229363201804,kWh,7.3510337207052455e-06,4.028090057331343e-06,3.744761165742497e-05,4.8826735435461556e-05,tokens/kWh,5243029.207602399,MB,3768.549376,12218.53184,0.0,11572.084736,11417.44384,s,10,338.151859375,33.815185937500004,0.002359642213759128,33.815607421875,33.817147265624996,33.8180384765625,33.8187514453125,"[33.813171875, 33.80974609375, 33.815921875, 33.81585546875, 33.8189296875, 33.814140625, 33.815359375, 33.81694921875, 33.8149453125, 33.81683984375]",tokens/s,1.8630682710555475,kWh,0.00039913779823316466,0.00021875981365912164,0.0020517374763559647,0.0026696350882482512,tokens/kWh,23598.73088173225,,s,629,343.2234896850585,0.5456653254134476,0.07413550819300896,0.53669677734375,0.5370257202148437,0.5371780883789062,1.1599311865234374,"[0.5362810668945313, 0.5365596313476563, 0.5364756469726563, 0.5363067016601563, 0.536511474609375, 0.5362237548828125, 0.536647705078125, 0.5367030029296875, 0.5366149291992187, 0.536501220703125, 0.5363087158203125, 0.5366619873046875, 0.5364234008789063, 0.536468505859375, 0.5364039306640624, 0.536290283203125, 0.5366968383789062, 0.5364224243164063, 0.5364838256835938, 0.5366435546875, 0.5367398681640625, 0.5367551879882813, 0.5366241455078125, 0.536642578125, 0.5363783569335937, 0.5365247802734375, 0.5365186767578125, 0.5369231567382813, 0.5367019653320313, 0.5366395263671875, 0.5363342895507812, 0.5365678100585938, 0.536532958984375, 0.536953857421875, 0.5370747680664063, 0.53696728515625, 0.5369107055664063, 0.53680126953125, 0.536784912109375, 0.5369497680664063, 0.5368955078125, 0.536742919921875, 0.5366722412109375, 0.5367408447265625, 0.5370828857421875, 0.5366578979492187, 0.5366333618164062, 0.5367685546875, 0.5373429565429687, 0.5368453369140626, 0.5371207885742187, 0.5368853149414062, 0.5367161865234376, 0.536858642578125, 0.5368402099609375, 0.5368125610351563, 0.5368002319335937, 0.5371084594726563, 0.5369446411132812, 0.5370245361328125, 0.5367992553710937, 0.536958984375, 1.159732177734375, 0.5364080810546875, 0.53661181640625, 0.5363170166015625, 0.5364008178710937, 0.5362647094726563, 0.5362411499023437, 0.5363547973632813, 0.5363763427734375, 0.5366896362304687, 0.5368125610351563, 0.5364735717773438, 0.5364674682617188, 0.5364121704101562, 0.536563720703125, 0.536489990234375, 0.5364398193359375, 0.5366784057617188, 0.537006103515625, 0.5365493774414063, 0.5365432739257813, 0.5365350341796875, 0.536369140625, 0.5363394775390625, 0.5367193603515625, 0.5366928100585937, 0.5367510375976563, 0.5365565185546874, 0.5368657836914063, 0.53655859375, 0.5367664794921875, 0.536964111328125, 0.5367859497070312, 0.5366876220703125, 0.5368555297851563, 0.536141845703125, 0.53671630859375, 0.5368156127929687, 0.5365514526367188, 0.5365862426757813, 0.5367060546875, 0.5366968383789062, 0.5366732788085937, 0.5367880249023438, 0.5367142333984375, 0.5371443481445313, 0.5368463134765625, 0.53663232421875, 0.5368638305664063, 0.53669677734375, 0.5367633666992188, 0.5369375, 0.5367716064453125, 0.5365780639648438, 0.5366773681640625, 0.5368350830078125, 0.5367705688476563, 0.5367490844726562, 0.536859619140625, 0.5367521362304688, 0.537037841796875, 0.536658935546875, 0.5368186645507812, 1.1609200439453125, 0.5362657470703125, 0.5365555419921875, 0.536721435546875, 0.53648486328125, 0.53642138671875, 0.5364653930664063, 0.5365667724609375, 0.5366026000976563, 0.536511474609375, 0.5365575561523438, 0.536310791015625, 0.5365985107421875, 0.5365227661132812, 0.536529052734375, 0.5363894653320312, 0.53654833984375, 0.5364520874023437, 0.5367716064453125, 0.5366333618164062, 0.5367439575195313, 0.5365176391601563, 0.5366046752929687, 0.53838232421875, 0.5368411865234375, 0.536627197265625, 0.5364777221679687, 0.5364521484375, 0.5369036254882813, 0.5366814575195312, 0.5366497192382812, 0.536468505859375, 0.5365791015625, 0.5366394653320312, 0.536933349609375, 0.5364234008789063, 0.536679443359375, 0.5367470703125, 0.5367152099609375, 0.5366619873046875, 0.536711181640625, 0.5366661376953125, 0.5368309936523438, 0.5369999389648438, 0.5369364624023437, 0.536826904296875, 0.536858642578125, 0.5387857666015625, 0.5367183227539063, 0.5369241333007813, 0.5373931274414062, 0.5369784545898437, 0.536806396484375, 0.536595458984375, 0.5368524780273437, 0.536958984375, 0.5367183227539063, 0.5368104858398437, 0.5369609985351562, 0.5369999389648438, 0.536901611328125, 0.53682177734375, 0.53682275390625, 1.160515625, 0.5361018676757813, 0.5365678100585938, 0.53625244140625, 0.5364254760742188, 0.5363978271484375, 0.5362288818359375, 0.53628515625, 0.5363446044921875, 0.5365831909179688, 0.5366722412109375, 0.53650537109375, 0.5376819458007812, 0.536669189453125, 0.5364346923828125, 0.536553466796875, 0.5367654418945312, 0.53661181640625, 0.5366384887695312, 0.53665283203125, 0.537007080078125, 0.5366466674804687, 0.5365309448242187, 0.5365933837890625, 0.536742919921875, 0.5365678100585938, 0.5365565185546874, 0.536458251953125, 0.5372078247070312, 0.5368923950195312, 0.5368361206054687, 0.5366210327148437, 0.5368002319335937, 0.5366077270507813, 0.5368760375976562, 0.5365084228515625, 0.5366435546875, 0.5367418823242187, 0.5365227661132812, 0.5366753540039062, 0.5366435546875, 0.5366548461914062, 0.5368023071289062, 0.5367337036132812, 0.5392015380859375, 0.5369129028320313, 0.5367388305664063, 0.5366497192382812, 0.5369774169921875, 0.53680126953125, 0.5369784545898437, 0.5371361083984375, 0.5369088134765625, 0.5367767333984375, 0.5368411865234375, 0.536658935546875, 0.5367756958007812, 0.5368413696289063, 0.5370243530273437, 0.5370480346679688, 0.5370777587890625, 0.5369784545898437, 0.5370316772460938, 1.1606220703125, 0.5364172973632813, 0.5367091064453124, 0.5370255126953125, 0.5368668212890625, 0.5365166015625, 0.5362349853515626, 0.536364013671875, 0.5366016235351563, 0.5368115234375, 0.5364859008789062, 0.5362606201171874, 0.5366507568359375, 0.5365955810546875, 0.5366537475585937, 0.5365309448242187, 0.5365606689453125, 0.53635693359375, 0.53663330078125, 0.536543212890625, 0.536880126953125, 0.5366712036132812, 0.5365288696289062, 0.536700927734375, 0.536458251953125, 0.5365206909179687, 0.5366599731445313, 0.5366569213867187, 0.5367030029296875, 0.5365924072265625, 0.5366507568359375, 0.5365924072265625, 0.53671630859375, 0.5367152709960937, 0.5367725830078125, 0.5365667724609375, 0.5367265014648438, 0.5366703491210938, 0.5366823120117188, 0.5366497192382812, 0.5367418823242187, 0.536647705078125, 0.5368361206054687, 0.5367510986328125, 0.537059326171875, 0.53715966796875, 0.5364152221679688, 0.5364254760742188, 0.5369794311523437, 0.53715966796875, 0.5369088134765625, 0.5369815063476563, 0.5368207397460938, 0.5366527709960938, 0.5367030029296875, 0.5368309936523438, 0.5368862915039062, 0.5418343505859375, 0.5373899536132812, 0.53707470703125, 0.5371945190429688, 0.537112548828125, 0.53711767578125, 1.15996875, 0.5362606201171874, 0.5365380859375, 0.5366005859375, 0.5365452880859375, 0.53631591796875, 0.536321044921875, 0.5363015747070312, 0.5365176391601563, 0.536543212890625, 0.5366220703125, 0.5364295654296874, 0.5364049682617188, 0.536647705078125, 0.536542236328125, 0.5364705200195312, 0.5366282348632813, 0.5364315795898438, 0.536563720703125, 0.536521728515625, 0.5365892944335937, 0.536700927734375, 0.5367664794921875, 0.5364962158203125, 0.53661376953125, 0.536468505859375, 0.5364418334960938, 0.5364090576171875, 0.5366179809570313, 0.5364705200195312, 0.5369210815429688, 0.5367070922851562, 0.536922119140625, 0.536896484375, 0.536995849609375, 0.5365135498046875, 0.5370050659179687, 0.5370306396484374, 0.5394656982421875, 0.536394775390625, 0.5367521362304688, 0.537101318359375, 0.5368862915039062, 0.5367367553710938, 0.536669189453125, 0.5368678588867187, 0.5367992553710937, 0.5364859008789062, 0.5366539306640625, 0.5368544311523438, 0.5369548950195312, 0.5371525268554688, 0.5368790893554688, 0.5366578979492187, 0.5367459716796875, 0.5368186645507812, 0.5368729858398438, 0.5367808227539063, 0.5369108276367187, 0.5368453369140626, 0.5372252197265625, 0.5366773681640625, 0.5368678588867187, 1.1612681884765625, 0.5363363647460937, 0.5364295654296874, 0.5364602661132812, 0.5364306030273438, 0.5367684936523438, 0.5365238037109376, 0.5364756469726563, 0.5367265014648438, 0.5365872802734375, 0.5365821533203124, 0.53650537109375, 0.5366405029296875, 0.5364459228515625, 0.5365443115234375, 0.5362177124023437, 0.5364324951171875, 0.536721435546875, 0.5367879638671875, 0.5362688598632812, 0.5366260986328125, 0.5364387817382813, 0.5364408569335938, 0.5387745361328125, 0.5367828979492187, 0.5366834716796876, 0.536796142578125, 0.536754150390625, 0.5371105346679688, 0.5366548461914062, 0.536666259765625, 0.53655029296875, 0.536647705078125, 0.5366445922851563, 0.5368330078125, 0.536394775390625, 0.536531982421875, 0.5366241455078125, 0.5366220703125, 0.5366558837890625, 0.5366978759765625, 0.536542236328125, 0.536754150390625, 0.536795166015625, 0.5367787475585938, 0.5369467163085937, 0.536838134765625, 0.5365647583007812, 0.5367982177734375, 0.5368995971679688, 0.5371882934570312, 0.53720166015625, 0.53701220703125, 0.5369548950195312, 0.5367684936523438, 0.5368616943359374, 0.5367019653320313, 0.5368186645507812, 0.5369036865234375, 0.5370368041992187, 0.5373153076171875, 0.5373296508789063, 0.5370501098632813, 1.1625748291015625, 0.5365565185546874, 0.5367439575195313, 0.5362565307617188, 0.5365073852539063, 0.5365032958984375, 0.5364664306640625, 0.53650537109375, 0.5365309448242187, 0.5368842163085937, 0.5364326171875, 0.5366077270507813, 0.5365883178710937, 0.5364326171875, 0.5365944213867188, 0.5365892944335937, 0.5378468017578125, 0.5365228881835937, 0.5367540283203125, 0.5366405029296875, 0.5367450561523438, 0.536545166015625, 0.5364736328125, 0.5366353759765625, 0.5367736206054687, 0.5366937866210938, 0.5369927978515625, 0.5367716064453125, 0.536795166015625, 0.5367532348632813, 0.536639404296875, 0.5366845703125, 0.5370337524414063, 0.536859619140625, 0.5371627807617188, 0.5368104858398437, 0.5368361206054687, 0.5368084716796875, 0.5367183227539063, 0.5369231567382813, 0.5367869262695313, 0.5366549682617188, 0.5366117553710937, 0.53684326171875, 0.5366405029296875, 0.536875, 0.5368248291015625, 0.5365494384765624, 0.5367357177734375, 0.536890380859375, 0.5368504028320312, 0.5369609985351562, 0.5369129028320313, 0.53688525390625, 0.5371054077148437, 0.5369815063476563, 0.5370162963867188, 0.5372426147460938, 0.5372528686523438, 0.5367838745117187, 0.5369313354492188, 0.536858642578125, 0.536848388671875, 1.1598345947265625, 0.5363496704101562, 0.5367756958007812, 0.536563720703125, 0.5363814697265625, 0.5364010009765625, 0.5364909057617188, 0.5365892944335937, 0.5365883178710937, 0.5366558837890625, 0.5366210327148437, 0.536458251953125, 0.5366978759765625, 0.5365155639648438, 0.5365811157226562, 0.5366445922851563, 0.5368104858398437, 0.5383587646484375, 0.5365862426757813, 0.5365698852539063, 0.5368668212890625, 0.5365514526367188, 0.5363937377929687, 0.5366200561523438, 0.5369077758789063, 0.5366599731445313, 0.5366865844726563, 0.5365606689453125, 0.536816650390625, 0.5366128540039062, 0.536658935546875, 0.5365227661132812, 0.536748046875, 0.5365575561523438, 0.53676953125, 0.5365933837890625, 0.5366190185546875, 0.5366876220703125, 0.5366507568359375, 0.5364306030273438, 0.5365791015625, 0.5367398681640625, 0.5368473510742188, 0.53686474609375, 0.5368627319335938, 0.5368157958984375, 0.5369567260742187, 0.5364766845703125, 0.5367859497070312, 0.5371401977539062, 0.5371054077148437, 0.5369528198242187, 0.5369129028320313, 0.5367203979492188, 0.536784912109375, 0.53682177734375, 0.5367337036132812, 0.53671728515625, 0.536933349609375, 0.5369047241210938, 0.5369364624023437, 0.5370839233398438, 0.5373368530273438, 1.1633480224609376, 0.5362933959960937, 0.536469482421875, 0.5363046875, 0.5364182739257812, 0.5366026000976563, 0.5363189697265625, 0.5363978271484375, 0.5365073852539063, 0.5366466674804687, 0.5364121704101562, 0.5363589477539062, 0.5364653930664063, 0.5364387817382813, 0.5365453491210938, 0.5363465576171875, 0.536647705078125, 0.5366210327148437, 0.536754150390625, 0.536722412109375, 0.5366098022460938, 0.5365985107421875, 0.5363824462890625, 0.5365166015625, 0.5368186645507812, 0.5366896362304687, 0.5384898681640625, 0.5365606689453125, 0.5365924072265625, 0.5365964965820312, 0.5368607177734375, 0.5365575561523438, 0.536784912109375, 0.5367736206054687, 0.536806396484375, 0.5364295654296874, 0.5370234985351563, 0.536875, 0.5366773681640625, 0.5367664794921875, 0.5367080688476562, 0.5366456298828125, 0.5365166015625, 0.5366886596679687, 0.5369712524414062, 0.5371290893554688, 0.53671923828125, 0.5365913696289063, 0.5369273071289062, 0.5369098510742187, 0.5368984375, 0.5369497680664063, 0.5370265502929688, 0.5368084716796875, 0.5368350830078125, 0.537069580078125, 0.5368176879882812, 0.5368914794921875, 0.5369465942382813, 0.5369600219726562, 0.5369692993164062, 0.53684423828125, 0.5369835815429688]",tokens/s,1.8326251521338754,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,5067.636736,5954.338816,0.0,5307.891712,5047.542272,s,10,3.4323795471191403,0.34323795471191404,0.00048827040489550147,0.3433291473388672,0.34376642456054685,0.34387233276367185,0.3439570593261719,"[0.34300979614257815, 0.34262857055664064, 0.34346331787109374, 0.34354571533203127, 0.34363656616210936, 0.34266415405273437, 0.34397824096679686, 0.34251531982421873, 0.34374288940429687, 0.34319497680664063]",tokens/s,745.8382631806134,kWh,4.0470568118272e-06,2.2176103339916154e-06,1.9958756707734912e-05,2.6223423853553725e-05,tokens/kWh,9762264.509380898,MB,5067.636736,5956.435968,0.0,5309.988864,5098.334208,s,10,188.4459453125,18.844594531250003,0.000986841295957757,18.8447451171875,18.845630078125,18.8459078125,18.84613,"[18.845328125, 18.843513671875, 18.843123046875, 18.843962890625, 18.8434375, 18.8453359375, 18.844580078125, 18.84491015625, 18.845568359375, 18.846185546875]",tokens/s,3.3431337509293213,kWh,0.000222482685468815,0.00012193904832238328,0.0010750333692852647,0.0014194551030764628,tokens/kWh,44383.22836943321,,s,629,191.23617190551752,0.30403206980209474,0.04079414021479381,0.2991022033691406,0.2992777221679687,0.299380126953125,0.6420626220703125,"[0.2995671081542969, 0.2989332580566406, 0.2989557800292969, 0.2990223388671875, 0.29910324096679686, 0.2990162048339844, 0.2989967346191406, 0.2990715026855469, 0.29976065063476565, 0.29905307006835935, 0.299109375, 0.29900799560546876, 0.2995210266113281, 0.2990540771484375, 0.299109375, 0.2990162048339844, 0.29891070556640625, 0.29890765380859374, 0.2993131408691406, 0.29908990478515624, 0.2989967346191406, 0.2990448608398438, 0.29937152099609377, 0.2990192565917969, 0.29921075439453126, 0.29905511474609375, 0.2990469055175781, 0.29901824951171874, 0.29900698852539065, 0.2991349792480469, 0.29893734741210937, 0.2990162048339844, 0.2990223388671875, 0.29918310546875, 0.2991749267578125, 0.29907354736328123, 0.2990458984375, 0.2990807189941406, 0.2992537536621094, 0.29905511474609375, 0.29910528564453126, 0.2991902770996094, 0.29912472534179685, 0.2991769714355469, 0.29928549194335935, 0.29916876220703126, 0.29912371826171874, 0.2990745544433594, 0.299061279296875, 0.2990888671875, 0.29916159057617187, 0.2990950317382812, 0.299109375, 0.2991073303222656, 0.2991851501464844, 0.2991708068847656, 0.29912576293945314, 0.2992127990722656, 0.29917593383789065, 0.29935308837890623, 0.29931417846679687, 0.29923123168945315, 0.6421053466796875, 0.2988984375, 0.29912371826171874, 0.2989240417480469, 0.2988410949707031, 0.2988554382324219, 0.29901412963867186, 0.299104248046875, 0.29892913818359373, 0.2989690856933594, 0.29887387084960937, 0.298956787109375, 0.29899365234375, 0.2990878601074219, 0.29906124877929685, 0.299030517578125, 0.2991513671875, 0.2989742126464844, 0.29914727783203127, 0.29903359985351563, 0.29900286865234377, 0.2990387268066406, 0.2991329345703125, 0.2989742126464844, 0.29904281616210937, 0.2989813842773438, 0.29913088989257813, 0.29914111328125, 0.2990909423828125, 0.29912677001953125, 0.29905612182617186, 0.2990448608398438, 0.2990315551757812, 0.29912371826171874, 0.2989854736328125, 0.2990592041015625, 0.29906942749023435, 0.2989537353515625, 0.29910629272460937, 0.29900799560546876, 0.29908377075195314, 0.2993008728027344, 0.29918719482421874, 0.29916775512695315, 0.29908377075195314, 0.29916775512695315, 0.29922201538085935, 0.2990929870605469, 0.299261962890625, 0.2989619140625, 0.2993121337890625, 0.29969100952148436, 0.2992056274414063, 0.29919540405273437, 0.29918206787109375, 0.2991441955566406, 0.2991749267578125, 0.299177978515625, 0.2993008728027344, 0.2992701416015625, 0.29937152099609377, 0.29927935791015625, 0.2991646728515625, 0.6426142578125, 0.298935302734375, 0.2990387268066406, 0.2990909423828125, 0.2989056091308594, 0.29897830200195313, 0.2988984375, 0.2990858154296875, 0.2988769226074219, 0.2989690856933594, 0.29888101196289063, 0.29904177856445313, 0.2989834289550781, 0.29909811401367187, 0.29904998779296876, 0.29913088989257813, 0.2990315551757812, 0.2989383544921875, 0.2991421508789063, 0.2991912841796875, 0.29896600341796875, 0.29906842041015624, 0.298998779296875, 0.29906942749023435, 0.29893939208984377, 0.29915341186523436, 0.29893121337890627, 0.2990878601074219, 0.2988963928222656, 0.2990950317382812, 0.29900390625, 0.29905612182617186, 0.2990878601074219, 0.29924453735351564, 0.29900390625, 0.298925048828125, 0.2989803466796875, 0.29914727783203127, 0.2991441955566406, 0.29915545654296877, 0.2991697998046875, 0.29916775512695315, 0.2991298522949219, 0.29921075439453126, 0.2991329345703125, 0.29903973388671873, 0.29923635864257814, 0.29909197998046877, 0.2995701904296875, 0.2990745544433594, 0.29925070190429687, 0.29932440185546877, 0.2994759826660156, 0.29927218627929686, 0.299219970703125, 0.29913702392578123, 0.2991421508789063, 0.29938177490234374, 0.2992404479980469, 0.299146240234375, 0.29913088989257813, 0.29912884521484373, 0.29908172607421873, 0.6427955322265625, 0.2988380126953125, 0.29899468994140627, 0.29905816650390626, 0.2989322204589844, 0.2990950317382812, 0.29887591552734377, 0.29903359985351563, 0.29886669921875, 0.29894451904296876, 0.298935302734375, 0.29905511474609375, 0.298903564453125, 0.299040771484375, 0.29913189697265624, 0.2990602111816406, 0.2990325622558594, 0.29894757080078127, 0.29899777221679685, 0.29923019409179685, 0.29894757080078127, 0.29910629272460937, 0.29928857421875, 0.29942681884765626, 0.2991175537109375, 0.299082763671875, 0.2993438720703125, 0.29921588134765625, 0.29901312255859375, 0.299472900390625, 0.29896395874023435, 0.29901516723632815, 0.29911141967773436, 0.2992486267089844, 0.2990223388671875, 0.29911141967773436, 0.29921484375, 0.29914520263671873, 0.2990929870605469, 0.29911346435546876, 0.29921075439453126, 0.2991697998046875, 0.29918719482421874, 0.2990745544433594, 0.2990663757324219, 0.29914520263671873, 0.2990960693359375, 0.29916058349609376, 0.29912472534179685, 0.29911962890625, 0.29925784301757813, 0.29921588134765625, 0.29918719482421874, 0.29902029418945314, 0.29905307006835935, 0.2990909423828125, 0.2991769714355469, 0.2992261047363281, 0.29937765502929686, 0.2991749267578125, 0.2991800231933594, 0.2990868530273437, 0.29933670043945315, 0.6428313598632812, 0.2993582153320313, 0.29901516723632815, 0.2991769714355469, 0.2989690856933594, 0.2989834289550781, 0.2989854736328125, 0.2989537353515625, 0.29893121337890627, 0.29903564453125, 0.2991441955566406, 0.2989906005859375, 0.2990571594238281, 0.2990295104980469, 0.298919921875, 0.29927423095703126, 0.29909197998046877, 0.29910015869140627, 0.2989260864257812, 0.2991984558105469, 0.29919232177734373, 0.2990110778808594, 0.2990120849609375, 0.2991964111328125, 0.2992414855957031, 0.2988717956542969, 0.2990673828125, 0.29917388916015625, 0.29899468994140627, 0.29915853881835935, 0.29908990478515624, 0.2990458984375, 0.2990366821289063, 0.29912472534179685, 0.2990295104980469, 0.2990243835449219, 0.299082763671875, 0.29906942749023435, 0.29905307006835935, 0.29905612182617186, 0.29910833740234377, 0.2990868530273437, 0.29914520263671873, 0.299109375, 0.29907763671875, 0.2991851501464844, 0.29910833740234377, 0.29907763671875, 0.2990909423828125, 0.2991482849121094, 0.299399169921875, 0.29915240478515626, 0.29914212036132815, 0.29911859130859375, 0.2991523742675781, 0.29915341186523436, 0.29911962890625, 0.2990715026855469, 0.2992179260253906, 0.2994176025390625, 0.2992680969238281, 0.2991011962890625, 0.2991575012207031, 0.6419527587890625, 0.29895269775390626, 0.29895782470703125, 0.2991626281738281, 0.29960601806640624, 0.2990438537597656, 0.2988912658691406, 0.29905099487304687, 0.2989742126464844, 0.2992076721191406, 0.299104248046875, 0.29909811401367187, 0.2988840942382813, 0.299114501953125, 0.2989219970703125, 0.29906124877929685, 0.2989906005859375, 0.2991278076171875, 0.2990848083496094, 0.2991226806640625, 0.29885952758789064, 0.2991206359863281, 0.2990315551757812, 0.2990325622558594, 0.2990643310546875, 0.299082763671875, 0.29917593383789065, 0.29901412963867186, 0.2991011962890625, 0.29909912109375, 0.2991206359863281, 0.299030517578125, 0.29902130126953125, 0.2991994934082031, 0.29918206787109375, 0.2990592041015625, 0.2989906005859375, 0.2991226806640625, 0.299156494140625, 0.29920870971679686, 0.2990950317382812, 0.2991298522949219, 0.29918106079101564, 0.2991216735839844, 0.2989168701171875, 0.2990592041015625, 0.2990653381347656, 0.29903973388671873, 0.29919232177734373, 0.2990807189941406, 0.2991513671875, 0.2993039245605469, 0.2991728515625, 0.29925784301757813, 0.2992261047363281, 0.29921484375, 0.29934490966796873, 0.29938482666015626, 0.29939712524414064, 0.2992824401855469, 0.2998896789550781, 0.29921588134765625, 0.2993233947753906, 0.64334130859375, 0.29923019409179685, 0.2991206359863281, 0.299082763671875, 0.29902542114257813, 0.29899365234375, 0.2990960693359375, 0.29924453735351564, 0.2990448608398438, 0.29908172607421873, 0.29901824951171874, 0.29904998779296876, 0.298967041015625, 0.29910015869140627, 0.2989404296875, 0.2989742126464844, 0.29897012329101563, 0.2990848083496094, 0.29909707641601563, 0.299030517578125, 0.298998779296875, 0.29902029418945314, 0.29907354736328123, 0.29908172607421873, 0.2989875183105469, 0.298998779296875, 0.29901824951171874, 0.2990960693359375, 0.2989967346191406, 0.2990295104980469, 0.2991595458984375, 0.2995118103027344, 0.2993684387207031, 0.29906942749023435, 0.29918923950195314, 0.299104248046875, 0.2993070068359375, 0.29917901611328124, 0.2990950317382812, 0.29909811401367187, 0.2991493225097656, 0.2991974487304688, 0.2990489501953125, 0.2992056274414063, 0.2990796813964844, 0.2991964111328125, 0.2990807189941406, 0.29902542114257813, 0.2990858154296875, 0.29915545654296877, 0.2991216735839844, 0.2991503295898438, 0.29911346435546876, 0.29918719482421874, 0.2990653381347656, 0.2991708068847656, 0.2992547912597656, 0.2992435302734375, 0.2993602600097656, 0.29922918701171874, 0.2991728515625, 0.29920254516601563, 0.2992015380859375, 0.6418012084960938, 0.2991349792480469, 0.29889434814453125, 0.29896295166015624, 0.29901516723632815, 0.29908990478515624, 0.2990458984375, 0.2989803466796875, 0.2989537353515625, 0.298956787109375, 0.29917593383789065, 0.2989956970214844, 0.29910015869140627, 0.29907659912109374, 0.2990540771484375, 0.29910324096679686, 0.298956787109375, 0.2991964111328125, 0.2990489501953125, 0.2991544189453125, 0.2989342651367187, 0.29900286865234377, 0.2990376892089844, 0.29903564453125, 0.29907763671875, 0.29900390625, 0.2990120849609375, 0.29911962890625, 0.29990194702148437, 0.29911141967773436, 0.2991841430664062, 0.2991912841796875, 0.29942169189453127, 0.29902334594726565, 0.2992752685546875, 0.2991329345703125, 0.2989834289550781, 0.29915545654296877, 0.2990858154296875, 0.29927117919921875, 0.29918206787109375, 0.29913906860351563, 0.2991155090332031, 0.29912677001953125, 0.2991073303222656, 0.2992117614746094, 0.29907659912109374, 0.29912576293945314, 0.299040771484375, 0.2991349792480469, 0.299109375, 0.2991441955566406, 0.2991912841796875, 0.29917388916015625, 0.29917388916015625, 0.2992650146484375, 0.2991380615234375, 0.29926705932617187, 0.2992127990722656, 0.2992332763671875, 0.29924249267578124, 0.2991441955566406, 0.29926092529296877, 0.6430504760742187, 0.2989834289550781, 0.29909811401367187, 0.2994985046386719, 0.2990489501953125, 0.298977294921875, 0.29900286865234377, 0.29916058349609376, 0.2991421508789063, 0.29897830200195313, 0.298919921875, 0.2990295104980469, 0.2990366821289063, 0.29935617065429687, 0.299404296875, 0.2989803466796875, 0.29897625732421873, 0.29904794311523436, 0.2991646728515625, 0.29921383666992185, 0.2990223388671875, 0.299072509765625, 0.29908990478515624, 0.29908990478515624, 0.29904998779296876, 0.29908377075195314, 0.2990643310546875, 0.299283447265625, 0.29897830200195313, 0.29912371826171874, 0.2990192565917969, 0.29912677001953125, 0.299040771484375, 0.29923019409179685, 0.2990571594238281, 0.29902642822265624, 0.2989854736328125, 0.29928448486328124, 0.2991708068847656, 0.2990325622558594, 0.29913702392578123, 0.29911654663085935, 0.2992414855957031, 0.299104248046875, 0.2991493225097656, 0.2990100402832031, 0.2993510437011719, 0.29903973388671873, 0.29935308837890623, 0.29907046508789065, 0.29934695434570313, 0.29927731323242185, 0.299261962890625, 0.29923019409179685, 0.2991708068847656, 0.2992281494140625, 0.299177978515625, 0.29919436645507813, 0.299261962890625, 0.29917593383789065, 0.2992384033203125, 0.2992486267089844, 0.2990745544433594, 0.6430740356445312, 0.29913906860351563, 0.2991749267578125, 0.29905511474609375, 0.2995701904296875, 0.29907046508789065, 0.29899777221679685, 0.2992384033203125, 0.29895782470703125, 0.29899777221679685, 0.2990120849609375, 0.29908990478515624, 0.299114501953125, 0.29909811401367187, 0.29896295166015624, 0.2990878601074219, 0.2992056274414063, 0.2990807189941406, 0.29908172607421873, 0.29921075439453126, 0.2990643310546875, 0.2990796813964844, 0.2988984375, 0.29907659912109374, 0.2990366821289063, 0.2989609069824219, 0.29907559204101564, 0.29899365234375, 0.2989875183105469, 0.2994810791015625, 0.2991155090332031, 0.29918106079101564, 0.29911041259765625, 0.2991493225097656, 0.298998779296875, 0.29911859130859375, 0.2991513671875, 0.2991329345703125, 0.2992261047363281, 0.2991902770996094, 0.299114501953125, 0.29907354736328123, 0.2991902770996094, 0.2991994934082031, 0.2991011962890625, 0.2990940246582031, 0.29914727783203127, 0.2992117614746094, 0.2990940246582031, 0.2996756591796875, 0.2992486267089844, 0.2991349792480469, 0.2990848083496094, 0.29916363525390627, 0.29900698852539065, 0.2991022033691406, 0.29899981689453126, 0.29923941040039065, 0.29935000610351564, 0.29913088989257813, 0.29920254516601563, 0.29925784301757813, 0.299298828125]",tokens/s,3.2891267051233624,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,MB,1786.404864,7019.692032,0.0,6371.147776,6308.565504,s,10,1.3137465667724608,0.1313746566772461,8.258479129390438e-05,0.13135326385498047,0.1314678756713867,0.13150939712524415,0.13154261428833008,"[0.13141754150390625, 0.13145864868164062, 0.13155091857910156, 0.13134719848632812, 0.1313323516845703, 0.1313593292236328, 0.13122665405273437, 0.13133062744140625, 0.1313363494873047, 0.13138694763183595]",tokens/s,1948.6254539102354,kWh,1.5525963674963728e-06,8.507484807100274e-07,6.44209389825974e-06,8.845438746466142e-06,tokens/kWh,28941469.986695126,MB,1849.270272,7021.789184,0.0,6373.244928,6309.223424,s,10,19.932973754882813,1.9932973754882812,0.01668513659128983,1.9877265014648438,2.011113586425781,2.024606719970703,2.035401226806641,"[2.0081151123046874, 2.038099853515625, 1.9893792724609376, 1.990716552734375, 1.9883814697265625, 1.978673828125, 1.9832659912109376, 1.98658740234375, 1.987071533203125, 1.9826827392578126]",tokens/s,31.605921311448785,kWh,2.357618249382328e-05,1.29200524256593e-05,7.239874695214015e-05,0.00010889498187162272,tokens/kWh,578539.0558609144,,s,630,19.93104792213441,0.031636584003387945,0.0005797264447420896,0.03146854400634766,0.0326434814453125,0.032836249351501465,0.03354303352355958,"[0.0318023681640625, 0.03161497688293457, 0.03158732795715332, 0.031547391891479495, 0.03312639999389649, 0.031491071701049804, 0.031470592498779294, 0.0316866569519043, 0.031306751251220705, 0.031546367645263675, 0.031426559448242186, 0.03139891242980957, 0.03079884719848633, 0.032379905700683595, 0.03144396781921387, 0.03152383995056152, 0.03141529655456543, 0.03141734313964844, 0.031579135894775394, 0.0313702392578125, 0.03143987274169922, 0.03153305625915527, 0.03143168067932129, 0.03292364883422851, 0.03278335952758789, 0.03273830413818359, 0.03268198394775391, 0.03266252899169922, 0.03283456039428711, 0.03270655822753906, 0.031575040817260744, 0.031720447540283206, 0.03143168067932129, 0.03129548835754394, 0.03159040069580078, 0.03145113563537598, 0.0312729606628418, 0.03137638473510742, 0.03168767929077149, 0.03160780715942383, 0.03148185539245606, 0.03146854400634766, 0.031662080764770506, 0.031409151077270506, 0.031322111129760744, 0.03163955116271973, 0.03177574348449707, 0.031631359100341795, 0.03124838447570801, 0.031524864196777344, 0.03146137619018555, 0.0313384952545166, 0.031388671875, 0.03189555168151856, 0.03297484970092773, 0.033056766510009765, 0.032742401123046876, 0.033159168243408206, 0.0329543685913086, 0.03284275054931641, 0.032494590759277346, 0.032911361694335936, 0.032863231658935545, 0.032868350982666016, 0.03268710327148437, 0.032514049530029294, 0.03277004623413086, 0.03245363235473633, 0.03278847885131836, 0.03273420715332031, 0.03259494400024414, 0.032814079284667966, 0.03097395133972168, 0.03095961570739746, 0.03159654426574707, 0.0311592960357666, 0.03091967964172363, 0.031319040298461914, 0.031627264022827145, 0.031352832794189454, 0.031356927871704104, 0.03194572830200195, 0.03132928085327148, 0.031304704666137696, 0.031221759796142577, 0.03165593528747559, 0.03271475219726563, 0.03278540802001953, 0.03274649429321289, 0.03297587203979492, 0.032656383514404294, 0.03260927963256836, 0.03284275054931641, 0.03259699249267578, 0.03282227325439453, 0.03283967971801758, 0.032553985595703126, 0.03261644744873047, 0.03222118377685547, 0.032963584899902344, 0.0326901741027832, 0.03283148956298828, 0.0331253776550293, 0.03241164779663086, 0.0330250244140625, 0.0328089599609375, 0.03279564666748047, 0.0327086067199707, 0.03283763122558594, 0.031604736328125, 0.03162009620666504, 0.03261030578613281, 0.033188865661621096, 0.03255295944213867, 0.032696319580078126, 0.03297177505493164, 0.032642047882080076, 0.03274137496948242, 0.032486400604248046, 0.03290828704833984, 0.03540070343017578, 0.031238143920898437, 0.031037439346313478, 0.031252479553222655, 0.031021055221557618, 0.03378688049316406, 0.03139174461364746, 0.03151769638061523, 0.031218687057495118, 0.0314071044921875, 0.031422464370727536, 0.03147980880737305, 0.03125760078430176, 0.031558656692504884, 0.03146137619018555, 0.03156582450866699, 0.03160166358947754, 0.03137740707397461, 0.03141939163208008, 0.03147980880737305, 0.03136819267272949, 0.031306751251220705, 0.031115264892578126, 0.03100569534301758, 0.03120947265625, 0.03457535934448242, 0.032553985595703126, 0.03222630310058594, 0.03144396781921387, 0.03161087989807129, 0.03139481544494629, 0.03134464073181152, 0.032008190155029294, 0.03266969680786133, 0.031645696640014646, 0.03285094451904297, 0.032551937103271485, 0.031591424942016604, 0.03136716842651367, 0.031456256866455076, 0.031470592498779294, 0.03148287963867188, 0.031682559967041016, 0.03156377601623535, 0.031507455825805664, 0.03172249603271484, 0.03163955116271973, 0.031546367645263675, 0.031475711822509765, 0.03168972778320313, 0.03161292839050293, 0.031153152465820313, 0.031091712951660157, 0.031339519500732424, 0.03146751976013184, 0.03139993667602539, 0.03102720069885254, 0.031235071182250978, 0.03139072036743164, 0.031222784042358398, 0.031178752899169923, 0.03131084823608398, 0.0315361270904541, 0.03138764762878418, 0.031071231842041015, 0.03144704055786133, 0.03160371208190918, 0.03157094383239746, 0.03189657592773437, 0.03177676773071289, 0.03170508766174317, 0.031052799224853517, 0.03120844841003418, 0.031202304840087892, 0.032099327087402346, 0.03176140785217285, 0.032421886444091795, 0.03347251129150391, 0.03279872131347656, 0.031488000869750975, 0.031045631408691408, 0.03138047981262207, 0.03134771156311035, 0.03171327972412109, 0.031406080245971676, 0.03140812873840332, 0.032707584381103515, 0.033463294982910154, 0.032683006286621095, 0.032054271697998044, 0.03135078430175781, 0.03134873580932617, 0.03144908714294434, 0.03154431915283203, 0.03141119956970215, 0.0314204158782959, 0.03153203201293946, 0.032077823638916016, 0.031318016052246093, 0.03148185539245606, 0.031710208892822264, 0.03144396781921387, 0.031341567993164066, 0.03151872062683105, 0.03153817558288574, 0.03155046463012695, 0.03132825660705566, 0.03146854400634766, 0.03138047981262207, 0.031099903106689454, 0.03144908714294434, 0.03159859275817871, 0.03145113563537598, 0.031340543746948245, 0.03156787109375, 0.031438848495483396, 0.03122790336608887, 0.03118284797668457, 0.03158118438720703, 0.031735807418823245, 0.031110143661499022, 0.030918655395507814, 0.031092735290527345, 0.03146342468261719, 0.03159347152709961, 0.031510528564453126, 0.03153203201293946, 0.03080601692199707, 0.031128576278686523, 0.03150643157958984, 0.03227238464355469, 0.03150643157958984, 0.03129446411132813, 0.031926271438598636, 0.03153305625915527, 0.03137843132019043, 0.0315996150970459, 0.03240959930419922, 0.031525888442993165, 0.031733760833740236, 0.032797695159912106, 0.03136819267272949, 0.03159756851196289, 0.031494144439697266, 0.03250175857543945, 0.031543296813964845, 0.03161292839050293, 0.03136614418029785, 0.03139379119873047, 0.0315863037109375, 0.031491071701049804, 0.03160063934326172, 0.03151667213439941, 0.03143475151062012, 0.031524864196777344, 0.03139891242980957, 0.03139993667602539, 0.03140096092224121, 0.03288063812255859, 0.0326901741027832, 0.03183206367492676, 0.031817728042602536, 0.031132671356201173, 0.031558656692504884, 0.03144704055786133, 0.031562751770019534, 0.031528959274291994, 0.03180441665649414, 0.031455232620239255, 0.03137433624267578, 0.031356927871704104, 0.03155251121520996, 0.03160371208190918, 0.0315545597076416, 0.03139993667602539, 0.03146342468261719, 0.03159244728088379, 0.031507455825805664, 0.03094121551513672, 0.03108348846435547, 0.031247360229492187, 0.031525888442993165, 0.03133030319213867, 0.03126169586181641, 0.03100262451171875, 0.031095808029174804, 0.03160780715942383, 0.031144960403442383, 0.030943231582641603, 0.03130572891235352, 0.03180646324157715, 0.032110591888427735, 0.03140096092224121, 0.03181056022644043, 0.031031295776367186, 0.031122432708740235, 0.031437824249267575, 0.03153305625915527, 0.031194112777709962, 0.031275007247924806, 0.03170918464660644, 0.03141529655456543, 0.03143987274169922, 0.031178752899169923, 0.03158323287963867, 0.031106048583984375, 0.031015935897827147, 0.03119206428527832, 0.03113167953491211, 0.030952415466308593, 0.03100467109680176, 0.031096832275390625, 0.031422464370727536, 0.03127910423278808, 0.031118335723876952, 0.03120748710632324, 0.03171219253540039, 0.03134771156311035, 0.032368640899658206, 0.031898624420166014, 0.031753215789794925, 0.03226828765869141, 0.03136000061035156, 0.0314521598815918, 0.031558656692504884, 0.031369216918945314, 0.03131289672851562, 0.03115007972717285, 0.03122585678100586, 0.031682559967041016, 0.03100979232788086, 0.031319040298461914, 0.031525888442993165, 0.03158732795715332, 0.031648767471313476, 0.031066112518310547, 0.031204351425170897, 0.03155148887634277, 0.031127552032470703, 0.031122432708740235, 0.03149007987976074, 0.030943199157714842, 0.03115519905090332, 0.03138662338256836, 0.03171737670898438, 0.03134464073181152, 0.03157606315612793, 0.031130624771118165, 0.030947328567504883, 0.03178598403930664, 0.0325591049194336, 0.031663103103637694, 0.03169587135314941, 0.031817728042602536, 0.03169484710693359, 0.031389696121215824, 0.030938112258911132, 0.031213567733764647, 0.03122483253479004, 0.03129548835754394, 0.031409151077270506, 0.03141529655456543, 0.031425535202026365, 0.03152179145812988, 0.031611904144287106, 0.03132620811462403, 0.03124019241333008, 0.03146240043640137, 0.031340543746948245, 0.031306751251220705, 0.031524864196777344, 0.0315043830871582, 0.03134976005554199, 0.03158425521850586, 0.03144908714294434, 0.031102975845336913, 0.03124224090576172, 0.031780864715576174, 0.03154022407531738, 0.030934015274047853, 0.03117363166809082, 0.03196313667297363, 0.03169075202941894, 0.031237119674682616, 0.031252479553222655, 0.0314521598815918, 0.031302656173706055, 0.03133337593078613, 0.03140096092224121, 0.03155558395385742, 0.031369216918945314, 0.03158323287963867, 0.031318016052246093, 0.031474687576293944, 0.031545343399047854, 0.03141734313964844, 0.031438848495483396, 0.031541248321533204, 0.03141734313964844, 0.03123302459716797, 0.030991359710693358, 0.03232563018798828, 0.03140300750732422, 0.03201228713989258, 0.031340543746948245, 0.03151667213439941, 0.031514623641967776, 0.0314071044921875, 0.03160268783569336, 0.03240447998046875, 0.03149516868591309, 0.03184435272216797, 0.03150643157958984, 0.031066112518310547, 0.0321341438293457, 0.03172863960266113, 0.03140505599975586, 0.032266239166259765, 0.03134771156311035, 0.031471616744995115, 0.03100262451171875, 0.03130982398986817, 0.03373158264160156, 0.03321139144897461, 0.03181363105773926, 0.03153305625915527, 0.03147776031494141, 0.031218687057495118, 0.0312801284790039, 0.03144601631164551, 0.03162521553039551, 0.031513599395751955, 0.031494144439697266, 0.031628288269042966, 0.031649791717529296, 0.03155046463012695, 0.031475711822509765, 0.0314654712677002, 0.0311592960357666, 0.03139276885986328, 0.031355903625488284, 0.03139686393737793, 0.03196313667297363, 0.031751167297363284, 0.03150540733337402, 0.03167436790466309, 0.031307775497436525, 0.031301631927490234, 0.031643648147583005, 0.03142758369445801, 0.03099750328063965, 0.03098931121826172, 0.031498239517211916, 0.03132620811462403, 0.0313118724822998, 0.03136204719543457, 0.03175424003601074, 0.031265792846679685, 0.031082496643066407, 0.031194112777709962, 0.0314839038848877, 0.03161087989807129, 0.031473663330078124, 0.0315156478881836, 0.030925823211669923, 0.031421440124511715, 0.030680063247680665, 0.030955520629882813, 0.0311910400390625, 0.03137740707397461, 0.03120128059387207, 0.031473663330078124, 0.031437824249267575, 0.03137228775024414, 0.03136511993408203, 0.031732736587524416, 0.03201945495605469, 0.031438848495483396, 0.03357183837890625, 0.032467967987060545, 0.03166924858093262, 0.03139583969116211, 0.03123200035095215, 0.03130060768127441, 0.034353153228759765, 0.03251814270019531, 0.031180799484252928, 0.031628288269042966, 0.03148287963867188, 0.031268863677978515, 0.03142451286315918, 0.030852096557617188, 0.03099238395690918, 0.030857215881347655, 0.031205375671386718, 0.03094937515258789, 0.03135078430175781, 0.0314521598815918, 0.030911487579345705, 0.03120639991760254, 0.03113369560241699, 0.03143168067932129, 0.031325183868408206, 0.031202304840087892, 0.03194470405578613, 0.03170611190795898, 0.031510528564453126, 0.03246284866333008, 0.031389696121215824, 0.031339519500732424, 0.031230976104736328, 0.03126067161560059, 0.03141836738586426, 0.031321088790893556, 0.03134566307067871, 0.03162419128417969, 0.03197747230529785, 0.031071231842041015, 0.031303680419921875, 0.03138150405883789, 0.03139583969116211, 0.03144908714294434, 0.0317071361541748, 0.03143577575683594, 0.03139174461364746, 0.0314654712677002, 0.03186278343200684, 0.031494144439697266, 0.0314654712677002, 0.031698944091796875, 0.031486976623535154, 0.03148492813110351, 0.03141734313964844, 0.03144806480407715, 0.03135897636413574, 0.031093759536743162, 0.03281817626953125, 0.031987712860107424, 0.031245311737060546, 0.031488000869750975, 0.031006719589233397, 0.0313702392578125, 0.031324159622192385, 0.03155763244628906, 0.03137638473510742, 0.032361473083496094, 0.03366195297241211, 0.031883264541625975, 0.031456256866455076, 0.03141939163208008, 0.03156070327758789, 0.03124019241333008, 0.031083520889282228, 0.03144601631164551, 0.031680511474609374, 0.0314521598815918, 0.03148287963867188, 0.032102401733398435, 0.0315043830871582, 0.03148287963867188, 0.03201740646362305, 0.03141119956970215, 0.031718399047851564, 0.031083520889282228, 0.03095756721496582, 0.03177267265319824, 0.03237785720825195, 0.03146342468261719, 0.03131391906738281, 0.03156172752380371, 0.03145113563537598, 0.03146444892883301, 0.031308799743652346, 0.03144601631164551, 0.03131699180603027, 0.03129855918884277, 0.03170816040039062, 0.03140812873840332, 0.03140096092224121, 0.03149004745483398, 0.031486976623535154, 0.03099443244934082, 0.03166720008850098, 0.031528959274291994, 0.031247360229492187, 0.031060991287231447, 0.031054847717285155, 0.03125964736938477, 0.03232665634155273, 0.031437824249267575, 0.03122585678100586, 0.03130060768127441, 0.031090688705444337, 0.031369216918945314, 0.03148492813110351, 0.031290367126464845, 0.031354976654052735, 0.0310086727142334, 0.032043006896972655, 0.03168870353698731, 0.031221759796142577, 0.03147878456115723, 0.03158323287963867, 0.03152076721191406, 0.03118387222290039, 0.03143475151062012, 0.0315043830871582, 0.03144704055786133, 0.032200702667236326, 0.03123302459716797]",tokens/s,31.608975226051918,,,,,,gemma,MB,4750.483456,6422.003712,0.0,5792.333824,5786.358272,s,1,11.078771484375,11.078771484375,0.0,11.078771484375,11.078771484375,11.078771484375,11.078771484375,[11.078771484375],,kWh,5.032034448611411e-05,2.7563363017127772e-05,7.23628356679995e-05,0.0001502465431712414, @@ -8063,48 +7934,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v2-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,4111.491072,5287.44448,0.0,4640.997376,4330.799616,s,10,3.226784301757813,0.3226784301757813,0.00036527946974059694,0.3228235473632812,0.3229854919433594,0.3231664367675781,0.3233111926269531,"[0.3222513732910156, 0.3229452819824219, 0.3228663024902344, 0.32282205200195313, 0.3228476257324219, 0.32209072875976563, 0.3233473815917969, 0.32249490356445315, 0.3228250427246094, 0.3222936096191406]",tokens/s,793.3595061205122,kWh,3.802613670834237e-06,2.0836708957028095e-06,1.8872671348126036e-05,2.475895591466308e-05,tokens/kWh,10339692.872444117,MB,4111.491072,5287.44448,0.0,4640.997376,4515.66848,s,10,175.44622851562502,17.544622851562497,0.0012475331034878638,17.5441357421875,17.546382031249998,17.546624609374998,17.546818671875,"[17.5441171875, 17.545755859375, 17.543529296875, 17.545173828125, 17.546328125, 17.54321484375, 17.543701171875, 17.5468671875, 17.544154296875, 17.54338671875]",tokens/s,3.5908437891777947,kWh,0.00020715286491128305,0.0001135369823019525,0.0010055265370594742,0.0013262163842727099,tokens/kWh,47503.560314215894,,s,629,178.07182846069333,0.28310306591525175,0.03837524534756707,0.278476806640625,0.2786437072753906,0.2787579833984375,0.6014762182617188,"[0.27898776245117185, 0.2789980163574219, 0.27833856201171875, 0.2784163818359375, 0.27828939819335935, 0.2783426513671875, 0.2782156677246094, 0.2783068237304688, 0.278192138671875, 0.27826995849609376, 0.2781388854980469, 0.2783426513671875, 0.27819931030273437, 0.27850750732421875, 0.27829043579101564, 0.2783999938964844, 0.2782494812011719, 0.2784798583984375, 0.2784337768554688, 0.2785310668945313, 0.2782771301269531, 0.27843585205078125, 0.2784327697753906, 0.2784798583984375, 0.2783631286621094, 0.27856793212890624, 0.27837542724609377, 0.2785208435058594, 0.27838668823242185, 0.2784635009765625, 0.2784471130371094, 0.2785535888671875, 0.2784409484863281, 0.2785791931152344, 0.278408203125, 0.27850958251953123, 0.2783365173339844, 0.27850958251953123, 0.2784204711914062, 0.27854437255859377, 0.27840203857421875, 0.27851263427734374, 0.2784604187011719, 0.2785433654785156, 0.27850546264648435, 0.27857510375976563, 0.2786293640136719, 0.278572021484375, 0.2785208435058594, 0.27859967041015626, 0.27844403076171875, 0.27852291870117185, 0.27841943359375, 0.2788905029296875, 0.2784184265136719, 0.2785710144042969, 0.27854232788085936, 0.2785515441894531, 0.27840408325195315, 0.27864984130859377, 0.27850958251953123, 0.2785853576660156, 0.6017566528320313, 0.27841537475585937, 0.27831195068359377, 0.27905227661132814, 0.27840103149414064, 0.27819110107421874, 0.27843685913085936, 0.2782791748046875, 0.2782801818847656, 0.2781614074707031, 0.27826483154296877, 0.2782443542480469, 0.2783846435546875, 0.2782863464355469, 0.27843585205078125, 0.27829351806640623, 0.278255615234375, 0.2783743896484375, 0.2784132995605469, 0.2783170471191406, 0.2789324951171875, 0.2784429931640625, 0.2784604187011719, 0.2784757690429687, 0.2785218505859375, 0.27847271728515627, 0.2788792419433594, 0.27871539306640625, 0.27856484985351565, 0.27845632934570314, 0.278508544921875, 0.2784757690429687, 0.2786160583496094, 0.27841741943359377, 0.2785157165527344, 0.2784747619628906, 0.2785228881835938, 0.2784194641113281, 0.27849420166015626, 0.2787901306152344, 0.27852801513671877, 0.2783682861328125, 0.2785689392089844, 0.27856692504882813, 0.27852593994140623, 0.27843072509765626, 0.278687744140625, 0.27848703002929687, 0.2785269775390625, 0.278518798828125, 0.27858740234375, 0.27840103149414064, 0.2785638427734375, 0.2785321044921875, 0.27859661865234375, 0.2784675903320312, 0.2790205383300781, 0.2785515441894531, 0.2786007080078125, 0.278513671875, 0.2786293640136719, 0.2785740661621094, 0.27877886962890625, 0.6017741088867188, 0.27854849243164065, 0.278319091796875, 0.27832830810546877, 0.2783969421386719, 0.2782054443359375, 0.27833242797851565, 0.2782443542480469, 0.2784460754394531, 0.27819827270507813, 0.27826174926757813, 0.27834878540039065, 0.2784337768554688, 0.2783303527832031, 0.27842559814453127, 0.2782597045898437, 0.278403076171875, 0.2785167236328125, 0.27843685913085936, 0.2784327697753906, 0.2784460754394531, 0.27831195068359377, 0.27852593994140623, 0.278476806640625, 0.27842251586914063, 0.2785218505859375, 0.2786344909667969, 0.2785024108886719, 0.27854949951171876, 0.27850137329101565, 0.27848397827148436, 0.27858123779296873, 0.27854132080078126, 0.27842559814453127, 0.27854132080078126, 0.27832830810546877, 0.27854745483398435, 0.2783006591796875, 0.27839794921875, 0.27840615844726563, 0.27851467895507814, 0.27842251586914063, 0.2784471130371094, 0.27854849243164065, 0.2784737243652344, 0.2784901123046875, 0.27870513916015627, 0.27853927612304685, 0.2786344909667969, 0.2785372009277344, 0.2786078796386719, 0.278498291015625, 0.27864474487304686, 0.2784655456542969, 0.27859457397460935, 0.2785228881835938, 0.27856280517578125, 0.2785730590820312, 0.27854541015625, 0.2784122924804687, 0.27856484985351565, 0.2784122924804687, 0.27866827392578125, 0.601596923828125, 0.27889459228515623, 0.2783109130859375, 0.27825048828125, 0.27840716552734374, 0.2782853088378906, 0.2783498229980469, 0.2783273010253906, 0.2783160400390625, 0.2782740478515625, 0.27838156127929686, 0.2782013549804688, 0.2784880676269531, 0.27822900390625, 0.27846142578125, 0.2783918151855469, 0.2784266357421875, 0.27835186767578124, 0.2784952392578125, 0.2784378967285156, 0.27846450805664064, 0.27835391235351564, 0.27842868041992186, 0.27840203857421875, 0.2784296875, 0.27854745483398435, 0.27856793212890624, 0.2784204711914062, 0.27891302490234376, 0.278434814453125, 0.27845428466796873, 0.2784532470703125, 0.27840615844726563, 0.2785607604980469, 0.2784849853515625, 0.2785587158203125, 0.27897445678710936, 0.2788044738769531, 0.278498291015625, 0.2784757690429687, 0.278513671875, 0.27842251586914063, 0.2785238952636719, 0.27860275268554685, 0.27848089599609377, 0.27847271728515627, 0.27856484985351565, 0.2786160583496094, 0.27856484985351565, 0.27850546264648435, 0.27857510375976563, 0.2784552917480469, 0.27852291870117185, 0.27853717041015624, 0.27864678955078126, 0.27850546264648435, 0.2786437072753906, 0.27860479736328125, 0.27857510375976563, 0.2783743896484375, 0.27860888671875, 0.27841741943359377, 0.27879220581054687, 0.6017177734375, 0.27843173217773437, 0.27856484985351565, 0.27820645141601563, 0.2784532470703125, 0.27828121948242185, 0.2783703002929688, 0.27863961791992187, 0.2783273010253906, 0.27872869873046874, 0.27829144287109375, 0.27829248046875, 0.27839385986328125, 0.278645751953125, 0.2785607604980469, 0.2783549499511719, 0.27834573364257814, 0.27826278686523437, 0.2784409484863281, 0.2784675903320312, 0.27854949951171876, 0.27830374145507814, 0.278476806640625, 0.27845428466796873, 0.2783498229980469, 0.27835699462890623, 0.2788239440917969, 0.27835699462890623, 0.2785310668945313, 0.278550537109375, 0.27856484985351565, 0.27840103149414064, 0.27856793212890624, 0.2785863647460938, 0.2785269775390625, 0.27840716552734374, 0.2784952392578125, 0.27840615844726563, 0.27848703002929687, 0.27833856201171875, 0.27859762573242186, 0.27845632934570314, 0.2785884094238281, 0.27898675537109374, 0.278550537109375, 0.27880551147460936, 0.27859762573242186, 0.27865396118164065, 0.27867034912109373, 0.27876043701171876, 0.27874508666992187, 0.27836517333984373, 0.27855975341796874, 0.2786437072753906, 0.2785218811035156, 0.2784992980957031, 0.2785587158203125, 0.27855563354492185, 0.27860992431640624, 0.27846142578125, 0.2785710144042969, 0.2784337768554688, 0.27867544555664064, 0.6009231567382812, 0.27844915771484374, 0.27832626342773437, 0.27822796630859375, 0.278319091796875, 0.27819314575195314, 0.2783150024414062, 0.2781388854980469, 0.27829556274414063, 0.2782443542480469, 0.27844198608398435, 0.2782054443359375, 0.2784204711914062, 0.2782515258789062, 0.27840103149414064, 0.2783088684082031, 0.27841433715820313, 0.2784051208496094, 0.27846450805664064, 0.27850137329101565, 0.2785863647460938, 0.2783273010253906, 0.2784471130371094, 0.27842251586914063, 0.2783897705078125, 0.2784798583984375, 0.2786458129882812, 0.2784869689941406, 0.2785003662109375, 0.2784901123046875, 0.27853927612304685, 0.27848703002929687, 0.27861810302734374, 0.2784532470703125, 0.27848907470703127, 0.27840203857421875, 0.2785525817871094, 0.27850137329101565, 0.2785290222167969, 0.27842251586914063, 0.27850445556640624, 0.27838772583007815, 0.27854232788085936, 0.2785658874511719, 0.2785986633300781, 0.2783918151855469, 0.2785904541015625, 0.2784696350097656, 0.27860275268554685, 0.27858944702148436, 0.27858944702148436, 0.2784378967285156, 0.2785228881835938, 0.27860479736328125, 0.2785290222167969, 0.27843685913085936, 0.2785802307128906, 0.2785003662109375, 0.2785658874511719, 0.27848397827148436, 0.2786344909667969, 0.2784583740234375, 0.2787481689453125, 0.602113037109375, 0.2783365173339844, 0.2783211669921875, 0.27824127197265625, 0.2783375244140625, 0.2782105712890625, 0.27836721801757813, 0.27829144287109375, 0.2783723449707031, 0.27835903930664063, 0.27832421875, 0.2782996520996094, 0.2784122924804687, 0.27834878540039065, 0.2786160583496094, 0.2783375244140625, 0.27824127197265625, 0.278371337890625, 0.2785658874511719, 0.27838772583007815, 0.27851980590820313, 0.27827096557617187, 0.27854949951171876, 0.2785321044921875, 0.2784276428222656, 0.27848703002929687, 0.2785791931152344, 0.2783416442871094, 0.2785105895996094, 0.2784778137207031, 0.27854232788085936, 0.2784430541992188, 0.27862832641601565, 0.27843585205078125, 0.2785218505859375, 0.27845733642578124, 0.27856179809570314, 0.27849932861328125, 0.2785587158203125, 0.2783477783203125, 0.27854541015625, 0.27844915771484374, 0.2785433654785156, 0.2784583740234375, 0.27859457397460935, 0.27845016479492185, 0.27858944702148436, 0.27850958251953123, 0.27855667114257815, 0.2786160583496094, 0.27856997680664064, 0.27856179809570314, 0.27869696044921877, 0.27845120239257815, 0.2784655456542969, 0.27840103149414064, 0.278550537109375, 0.27846142578125, 0.2785904541015625, 0.2785884094238281, 0.27869696044921877, 0.27850137329101565, 0.2786795654296875, 0.6017638549804688, 0.2783743896484375, 0.27835699462890623, 0.2782576599121094, 0.2783549499511719, 0.2782853088378906, 0.2783774719238281, 0.2782945251464844, 0.27829043579101564, 0.27835391235351564, 0.278423583984375, 0.27823715209960936, 0.27836517333984373, 0.2787543029785156, 0.2783446960449219, 0.27829144287109375, 0.27843685913085936, 0.27847064208984373, 0.27850958251953123, 0.2783897705078125, 0.27845733642578124, 0.2784163818359375, 0.2784163818359375, 0.27838156127929686, 0.2784635009765625, 0.2784245910644531, 0.27858432006835937, 0.27845016479492185, 0.27850650024414064, 0.27848602294921876, 0.27913421630859375, 0.27850445556640624, 0.2785167236328125, 0.2786078796386719, 0.2784952392578125, 0.27845733642578124, 0.2789427185058594, 0.27852593994140623, 0.27844915771484374, 0.2784583740234375, 0.27856692504882813, 0.278466552734375, 0.27858328247070313, 0.2785433654785156, 0.27850546264648435, 0.27851776123046873, 0.27864678955078126, 0.27859967041015626, 0.27850650024414064, 0.2791065673828125, 0.27865396118164065, 0.2785607604980469, 0.27861810302734374, 0.2784532470703125, 0.27860992431640624, 0.2784962463378906, 0.2786344909667969, 0.2785310668945313, 0.278687744140625, 0.2785935363769531, 0.27867340087890624, 0.27854745483398435, 0.2786990051269531, 0.601702392578125, 0.278392822265625, 0.2786713562011719, 0.2782300109863281, 0.2784583740234375, 0.2782576599121094, 0.2783621215820313, 0.2783365173339844, 0.2783354797363281, 0.27830374145507814, 0.27835699462890623, 0.2782668762207031, 0.27834060668945315, 0.27823513793945315, 0.27846142578125, 0.27838668823242185, 0.2784194641113281, 0.27830374145507814, 0.27834878540039065, 0.2784184265136719, 0.27842355346679687, 0.27863348388671877, 0.27850546264648435, 0.2784122924804687, 0.27835391235351564, 0.2784122924804687, 0.27847271728515627, 0.27833856201171875, 0.27849932861328125, 0.27848602294921876, 0.2785218505859375, 0.27842251586914063, 0.27849420166015626, 0.2784583740234375, 0.2783856506347656, 0.27835186767578124, 0.27848501586914065, 0.27843887329101563, 0.2786611328125, 0.27842355346679687, 0.27854849243164065, 0.2784798583984375, 0.27856280517578125, 0.2785218505859375, 0.27852801513671877, 0.2784655456542969, 0.2786488342285156, 0.2786211853027344, 0.27856280517578125, 0.27854437255859377, 0.2785904541015625, 0.27843072509765626, 0.2785986633300781, 0.2788648986816406, 0.27861709594726564, 0.27849114990234375, 0.27876455688476565, 0.2785853576660156, 0.27858944702148436, 0.27855462646484375, 0.2785802307128906, 0.2784532470703125, 0.2786324462890625, 0.6011658325195313, 0.2784409484863281, 0.2784972839355469, 0.278302734375, 0.27838668823242185, 0.27827096557617187, 0.27833447265625, 0.27829556274414063, 0.2782597045898437, 0.27825357055664063, 0.2784184265136719, 0.27841537475585937, 0.27831808471679687, 0.278408203125, 0.2784194641113281, 0.27828121948242185, 0.2784747619628906, 0.2783160400390625, 0.2783692932128906, 0.2782843017578125, 0.2784716796875, 0.27823410034179685, 0.278403076171875, 0.27845120239257815, 0.2784215087890625, 0.27848907470703127, 0.27865805053710935, 0.2783918151855469, 0.27849319458007815, 0.2785710144042969, 0.27845733642578124, 0.27843685913085936, 0.2785300598144531, 0.2784215087890625, 0.27843072509765626, 0.27834878540039065, 0.2785382385253906, 0.2784409484863281, 0.2785535888671875, 0.27840615844726563, 0.2786293640136719, 0.27847064208984373, 0.27848193359375, 0.27861505126953123, 0.2784409484863281, 0.27836007690429687, 0.27853515625, 0.2786437072753906, 0.2786375732421875, 0.27863143920898437, 0.27857818603515627, 0.278403076171875, 0.2786017150878906, 0.2784962463378906, 0.27854849243164065, 0.27848193359375, 0.27860479736328125, 0.27847885131835937, 0.2786693115234375, 0.2785024108886719, 0.2786283569335937, 0.27850546264648435, 0.2787266540527344]",tokens/s,3.532282480823979,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch @@ -8962,6 +8818,33 @@ ChildProcessError: Traceback (most recent call last): OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,2123.460608,5971.116032,0.0,5324.668928,5080.876032,s,10,3.2550746459960944,0.3255074645996094,0.0008337963032000422,0.325276611328125,0.3257435028076172,0.32685052642822265,0.3277361453247071,"[0.32795755004882815, 0.32484524536132814, 0.32533059692382815, 0.3251536560058594, 0.32514990234375, 0.32549749755859375, 0.32518228149414063, 0.3252226257324219, 0.32538031005859375, 0.32535498046875]",tokens/s,786.4642991056838,kWh,3.8430350602314035e-06,2.105488517026383e-06,1.9587694881612944e-05,2.553621845887073e-05,tokens/kWh,10024976.89359605,MB,2123.460608,5971.116032,0.0,5324.668928,5290.118656,s,10,177.20532812500002,17.720532812500004,0.0052544305035837425,17.7196923828125,17.7276015625,17.7277421875,17.7278546875,"[17.71912890625, 17.71456640625, 17.720630859375, 17.7278828125, 17.71237890625, 17.727517578125, 17.716267578125, 17.719486328125, 17.7275703125, 17.7198984375]",tokens/s,3.5551978412048655,kWh,0.0002091554838338847,0.00011463415165342726,0.0010539822473069871,0.0013777718827942989,tokens/kWh,45726.00209566469,,s,629,179.85040585327167,0.2859306929304793,0.03863957620976894,0.2812538757324219,0.28160614013671875,0.2817001525878906,0.6063846411132813,"[0.28122930908203125, 0.28129177856445314, 0.2810245056152344, 0.2810921020507812, 0.2811351013183594, 0.28118630981445314, 0.28117401123046876, 0.281101318359375, 0.28111154174804687, 0.28105831909179685, 0.2810572814941406, 0.28108389282226565, 0.28105010986328127, 0.2811955261230469, 0.28121395874023436, 0.2812866516113281, 0.2815754089355469, 0.28128973388671874, 0.2816645202636719, 0.281312255859375, 0.28124160766601564, 0.28118630981445314, 0.2813624267578125, 0.28113613891601563, 0.28115045166015623, 0.28137777709960937, 0.2815621032714844, 0.28139520263671874, 0.28123544311523435, 0.281302001953125, 0.2815447082519531, 0.28124261474609374, 0.28110540771484377, 0.28135833740234373, 0.28128460693359375, 0.2810368041992187, 0.2815948791503906, 0.2811248779296875, 0.28113201904296875, 0.28118118286132815, 0.28119448852539064, 0.28112896728515624, 0.28108901977539064, 0.28126412963867187, 0.28096820068359374, 0.2810245056152344, 0.281270263671875, 0.28120883178710937, 0.28115045166015623, 0.2812313537597656, 0.2814156799316406, 0.28137677001953126, 0.281375732421875, 0.28129995727539064, 0.281628662109375, 0.2811781005859375, 0.28127743530273436, 0.28120779418945313, 0.281449462890625, 0.2811463623046875, 0.28092312622070315, 0.2812456970214844, 0.6060492553710938, 0.280764404296875, 0.2809722900390625, 0.2808719482421875, 0.281027587890625, 0.2810429382324219, 0.28127435302734377, 0.2810992736816406, 0.28099072265625, 0.2813102111816406, 0.28106854248046875, 0.28104193115234377, 0.28099172973632813, 0.28096408081054686, 0.28089447021484376, 0.281059326171875, 0.2810460205078125, 0.28098968505859373, 0.2810777587890625, 0.2809354248046875, 0.281017333984375, 0.2809866333007813, 0.2811197509765625, 0.2810849304199219, 0.2814791564941406, 0.28121600341796876, 0.28174029541015627, 0.2812938232421875, 0.28124774169921873, 0.28133477783203126, 0.2811064453125, 0.28119143676757813, 0.2810798034667969, 0.2810624084472656, 0.28098458862304687, 0.28118731689453125, 0.28110848999023436, 0.2811555786132813, 0.28107571411132815, 0.28158770751953127, 0.281459716796875, 0.2814412841796875, 0.2813644714355469, 0.2810142822265625, 0.2813040771484375, 0.28120574951171873, 0.28168295288085937, 0.28170138549804685, 0.2815539245605469, 0.2811822204589844, 0.2811094970703125, 0.28119143676757813, 0.28096307373046875, 0.2810716247558594, 0.28109722900390627, 0.2813061218261719, 0.2811678771972656, 0.2812630920410156, 0.2813839416503906, 0.2814310302734375, 0.281312255859375, 0.2810705871582031, 0.28164608764648436, 0.6065008544921875, 0.28089752197265627, 0.2808965148925781, 0.28090573120117185, 0.281059326171875, 0.28105831909179685, 0.2810101623535156, 0.28111566162109375, 0.28095489501953125, 0.2809876403808594, 0.2810449829101562, 0.2810992736816406, 0.28105831909179685, 0.28147406005859377, 0.2811781005859375, 0.28130303955078123, 0.28129791259765624, 0.2810439758300781, 0.2809374694824219, 0.2809610290527344, 0.2811955261230469, 0.28140542602539065, 0.281196533203125, 0.2814361572265625, 0.28143307495117187, 0.2812866516113281, 0.2814289855957031, 0.2812600402832031, 0.2813726806640625, 0.2812508239746094, 0.2812313537597656, 0.28151092529296873, 0.2815119323730469, 0.28151092529296873, 0.28148532104492185, 0.2814044189453125, 0.2812886962890625, 0.28121395874023436, 0.2811412353515625, 0.28146585083007813, 0.28120166015625, 0.2812876892089844, 0.28140032958984373, 0.28124261474609374, 0.2812620849609375, 0.2812119140625, 0.281275390625, 0.2814259338378906, 0.2816952209472656, 0.2817484741210938, 0.2815324096679688, 0.28149554443359376, 0.281486328125, 0.2813368225097656, 0.28129995727539064, 0.2814289855957031, 0.2814341125488281, 0.2812600402832031, 0.28163482666015627, 0.2814218139648437, 0.2814986267089844, 0.2813286437988281, 0.2812938232421875, 0.6064302368164063, 0.28105831909179685, 0.28143719482421875, 0.28149249267578125, 0.28150885009765625, 0.2818365478515625, 0.28147198486328123, 0.28145150756835935, 0.2814064636230469, 0.281459716796875, 0.28177203369140624, 0.2814310302734375, 0.281238525390625, 0.2812333984375, 0.28132659912109376, 0.28122213745117186, 0.2811781005859375, 0.28126412963867187, 0.28113409423828123, 0.2814412841796875, 0.28169830322265627, 0.2810572814941406, 0.28097537231445313, 0.28115045166015623, 0.281069580078125, 0.2811566162109375, 0.281143310546875, 0.2815662231445312, 0.2811402282714844, 0.28129995727539064, 0.28122113037109375, 0.28142385864257813, 0.2810705871582031, 0.28127435302734377, 0.2812467346191406, 0.2814709777832031, 0.28154061889648435, 0.2812989501953125, 0.281712646484375, 0.28118630981445314, 0.28149658203125, 0.2824161376953125, 0.28164913940429687, 0.28166656494140624, 0.28172903442382813, 0.28148532104492185, 0.28164813232421876, 0.2819051513671875, 0.2816860046386719, 0.28129074096679685, 0.2810828857421875, 0.28118014526367185, 0.28128564453125, 0.28141876220703127, 0.28150067138671875, 0.2815682678222656, 0.28134912109375, 0.28145562744140623, 0.2814617614746094, 0.2815119323730469, 0.28112896728515624, 0.28120065307617187, 0.2814361572265625, 0.6062673950195312, 0.2807910461425781, 0.2809088134765625, 0.2809866333007813, 0.28096511840820315, 0.2809466857910156, 0.28097024536132814, 0.2810449829101562, 0.28104702758789063, 0.2812518310546875, 0.2811535339355469, 0.2812518310546875, 0.28147406005859377, 0.28112896728515624, 0.28104806518554687, 0.2808985595703125, 0.28109722900390627, 0.2809774169921875, 0.2810019836425781, 0.2810368041992187, 0.28092007446289063, 0.28094259643554687, 0.280816650390625, 0.2810726318359375, 0.28107467651367185, 0.28106854248046875, 0.2810071105957031, 0.2813572998046875, 0.281196533203125, 0.28106341552734376, 0.28121908569335935, 0.2811473999023438, 0.28106137084960936, 0.2811832275390625, 0.28110540771484377, 0.28103372192382814, 0.2810624084472656, 0.281280517578125, 0.28099993896484377, 0.2810992736816406, 0.28094464111328127, 0.2814136352539062, 0.2813040771484375, 0.2811217956542969, 0.2810163269042969, 0.281491455078125, 0.28120574951171873, 0.28158462524414063, 0.2816952209472656, 0.28116070556640627, 0.2811617431640625, 0.28148736572265626, 0.28130712890625, 0.28105523681640626, 0.281343994140625, 0.28129278564453125, 0.28119244384765624, 0.2813644714355469, 0.2811852722167969, 0.28122213745117186, 0.2812518310546875, 0.28121600341796876, 0.2815703125, 0.6071398315429688, 0.2814771118164062, 0.28115966796875, 0.28129177856445314, 0.28113919067382814, 0.281417724609375, 0.28138290405273436, 0.28137777709960937, 0.28117913818359375, 0.28141671752929687, 0.28105831909179685, 0.2811975708007812, 0.2810941467285156, 0.28144537353515625, 0.28139620971679685, 0.28142385864257813, 0.2814464111328125, 0.2810378112792969, 0.2813409423828125, 0.2816624755859375, 0.28126412963867187, 0.2811781005859375, 0.2808565673828125, 0.2814422912597656, 0.28143820190429686, 0.28139111328125, 0.2813706359863281, 0.2814730224609375, 0.28143206787109376, 0.2816470947265625, 0.28146484375, 0.281670654296875, 0.2815037536621094, 0.2814761047363281, 0.2816030578613281, 0.2814474182128906, 0.28148019409179686, 0.2816225280761719, 0.28140338134765625, 0.28146380615234373, 0.2813716430664063, 0.28143719482421875, 0.2815867004394531, 0.28133273315429685, 0.28145150756835935, 0.28143002319335936, 0.2815672302246094, 0.2816778259277344, 0.28152935791015626, 0.28127435302734377, 0.28112896728515624, 0.28127435302734377, 0.2815170593261719, 0.281143310546875, 0.28146585083007813, 0.2816501770019531, 0.28121395874023436, 0.2810716247558594, 0.2812989501953125, 0.2811627502441406, 0.2813183898925781, 0.2815201416015625, 0.28156414794921875, 0.6065889282226562, 0.2809610290527344, 0.2809169921875, 0.28118014526367185, 0.2810624084472656, 0.28140237426757814, 0.28091595458984375, 0.2810408935546875, 0.2810378112792969, 0.28120269775390627, 0.2810142822265625, 0.2811064453125, 0.2812733459472656, 0.2811217956542969, 0.2809190368652344, 0.2811146240234375, 0.28121395874023436, 0.28103988647460937, 0.28125799560546877, 0.28095590209960936, 0.2810921020507812, 0.2812047424316406, 0.2810859375, 0.28157644653320313, 0.2811074523925781, 0.2813736877441406, 0.28132965087890627, 0.28149554443359376, 0.28099072265625, 0.2809395141601562, 0.2811760559082031, 0.2810900573730469, 0.2810798034667969, 0.28106036376953125, 0.28111770629882815, 0.2811985778808594, 0.2812600402832031, 0.28122930908203125, 0.28111358642578127, 0.2813102111816406, 0.2814115905761719, 0.28146484375, 0.28112384033203125, 0.28095281982421877, 0.28117300415039065, 0.2812620849609375, 0.28116378784179685, 0.28120065307617187, 0.28133377075195315, 0.2811207580566406, 0.28124160766601564, 0.28134808349609375, 0.2812538757324219, 0.2812600402832031, 0.281343994140625, 0.28151910400390623, 0.28132965087890627, 0.2814218139648437, 0.28138906860351565, 0.28175668334960935, 0.28138290405273436, 0.2815150146484375, 0.281712646484375, 0.6064854736328125, 0.28096517944335936, 0.280976318359375, 0.281059326171875, 0.28107366943359374, 0.28107366943359374, 0.2810449829101562, 0.28102859497070315, 0.28117196655273435, 0.2813102111816406, 0.2813982849121094, 0.28133477783203126, 0.2816522216796875, 0.28102041625976565, 0.281122802734375, 0.28104806518554687, 0.28105010986328127, 0.2810296325683594, 0.28088525390625, 0.2810900573730469, 0.28102554321289064, 0.28089547729492187, 0.28095489501953125, 0.2809866333007813, 0.28106137084960936, 0.2808616943359375, 0.281101318359375, 0.2810357666015625, 0.2813009948730469, 0.2810091552734375, 0.281481201171875, 0.2811463623046875, 0.28136038208007813, 0.28138290405273436, 0.28177407836914065, 0.2813921203613281, 0.281101318359375, 0.281275390625, 0.2812200927734375, 0.28148223876953127, 0.281270263671875, 0.28124774169921873, 0.2812549133300781, 0.281275390625, 0.28143515014648435, 0.28106341552734376, 0.2812938232421875, 0.2816450500488281, 0.28184884643554686, 0.2816512145996094, 0.2815047607421875, 0.28159896850585936, 0.2812528686523437, 0.281175048828125, 0.2812651672363281, 0.28133172607421875, 0.28128256225585935, 0.28144332885742185, 0.2814535827636719, 0.28138290405273436, 0.28144436645507814, 0.28169830322265627, 0.2818529357910156, 0.6073692016601563, 0.2816573486328125, 0.2816153564453125, 0.28132760620117186, 0.2814689331054688, 0.28154367065429686, 0.281027587890625, 0.2809866333007813, 0.28119143676757813, 0.2810900573730469, 0.280995849609375, 0.2812333984375, 0.2813255615234375, 0.28138494873046876, 0.2815784912109375, 0.28136959838867187, 0.2814412841796875, 0.28131329345703127, 0.2810378112792969, 0.2817525634765625, 0.2815733642578125, 0.28126617431640627, 0.2815447082519531, 0.2814975891113281, 0.28120883178710937, 0.2812549133300781, 0.2811269226074219, 0.28102041625976565, 0.2809354248046875, 0.2811125793457031, 0.28105523681640626, 0.28127743530273436, 0.2812518310546875, 0.28106854248046875, 0.28128256225585935, 0.2811463623046875, 0.28116378784179685, 0.2817177734375, 0.28164096069335937, 0.28217242431640627, 0.28151602172851564, 0.28213861083984376, 0.2816296997070312, 0.28145458984375, 0.28150271606445315, 0.28149249267578125, 0.28160614013671875, 0.28172491455078125, 0.2816174011230469, 0.28166656494140624, 0.28144024658203126, 0.2816296997070312, 0.2815784912109375, 0.2815047607421875, 0.28180685424804686, 0.281459716796875, 0.2812569580078125, 0.2812108764648438, 0.2812108764648438, 0.2811453552246094, 0.28124774169921873, 0.2810777587890625, 0.28122930908203125, 0.606845947265625, 0.2807613830566406, 0.28094561767578125, 0.2809241638183594, 0.28097637939453124, 0.2810091552734375, 0.2809774169921875, 0.2810460205078125, 0.28142385864257813, 0.28146585083007813, 0.2812518310546875, 0.28173208618164064, 0.28189797973632813, 0.2814361572265625, 0.28138189697265625, 0.28133273315429685, 0.2814044189453125, 0.28123544311523435, 0.28123544311523435, 0.28124774169921873, 0.2811883544921875, 0.28133377075195315, 0.2814422912597656, 0.28136651611328123, 0.28155084228515626, 0.28144436645507814, 0.28169012451171876, 0.2812231750488281, 0.28088116455078127, 0.2811125793457031, 0.2812108764648438, 0.2812108764648438, 0.28103884887695313, 0.2812651672363281, 0.28122213745117186, 0.28098968505859373, 0.2810654602050781, 0.28117300415039065, 0.28104806518554687, 0.2814894104003906, 0.281248779296875, 0.2813368225097656, 0.28116683959960936, 0.2811555786132813, 0.28160513305664064, 0.2811412353515625, 0.28160614013671875, 0.2813726806640625, 0.28156414794921875, 0.2810163269042969, 0.28116888427734377, 0.2811412353515625, 0.2812467346191406, 0.2811412353515625, 0.2812590026855469, 0.28136651611328123, 0.28124160766601564, 0.28138494873046876, 0.2813511657714844, 0.28120269775390627, 0.28121701049804687, 0.28180685424804686, 0.28130712890625]",tokens/s,3.4973510180074943,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,1506.42688,1579.679744,0.0,933.23264,856.69632,s,10,0.43564921951293945,0.043564921951293946,0.0031757046407324095,0.04331248092651367,0.04437063980102539,0.0485523281097412,0.051897678756713875,"[0.052734016418457035, 0.04333430480957031, 0.04164771270751953, 0.043380638122558594, 0.0414692497253418, 0.04140153503417969, 0.04329619216918945, 0.04344137573242188, 0.04332876968383789, 0.04161542510986328]",tokens/s,5876.287355368403,kWh,4.969496102560134e-07,2.72305043992154e-07,1.379101243335609e-06,2.1483558975837764e-06,tokens/kWh,119160889.63095888,MB,1506.42688,1579.679744,0.0,933.23264,899.62496,s,10,24.820336181640624,2.4820336181640625,0.03505314468873128,2.4815734863281254,2.524392333984375,2.52884619140625,2.53240927734375,"[2.533300048828125, 2.512392822265625, 2.510242431640625, 2.523402587890625, 2.427984130859375, 2.441885986328125, 2.461017822265625, 2.4780390625, 2.48510791015625, 2.44696337890625]",tokens/s,25.382412042670286,kWh,2.9558689719627776e-05,1.6199240900285592e-05,7.306973609346903e-05,0.00011882766671338238,tokens/kWh,530179.5595461687,,s,629,25.16415895462036,0.04000661201052522,0.005123259677494279,0.03999129486083984,0.040338432312011716,0.04051681213378906,0.07998427032470704,"[0.04113715362548828, 0.041027584075927735, 0.041578495025634765, 0.04084121704101563, 0.04151398468017578, 0.041063423156738284, 0.04110131072998047, 0.0394700813293457, 0.039803905487060545, 0.03899084854125977, 0.03885260772705078, 0.04075110244750976, 0.04097536087036133, 0.04030361557006836, 0.04015513610839844, 0.04008652877807617, 0.0400261116027832, 0.04022272109985352, 0.04007628631591797, 0.040078399658203125, 0.04018272018432617, 0.040068096160888675, 0.040422401428222655, 0.04001484680175781, 0.04015411376953125, 0.04017356872558594, 0.04025958251953125, 0.04034048080444336, 0.04012134552001953, 0.04011212921142578, 0.04016537475585937, 0.040008705139160154, 0.040130561828613284, 0.04007731246948242, 0.04001484680175781, 0.04027699279785156, 0.03999129486083984, 0.04007731246948242, 0.040084480285644535, 0.0402503662109375, 0.04009983825683594, 0.04006399917602539, 0.04001279830932617, 0.040207359313964845, 0.04010092926025391, 0.04016838455200195, 0.04013158416748047, 0.040172542572021484, 0.040498241424560544, 0.04003936004638672, 0.040325119018554685, 0.04010291290283203, 0.040225791931152347, 0.04005376052856445, 0.04014796829223633, 0.039678016662597654, 0.04023392105102539, 0.04024422454833984, 0.03986227035522461, 0.04005376052856445, 0.040235008239746094, 0.04025446319580078, 0.08356147003173828, 0.04013158416748047, 0.040151039123535154, 0.04020537567138672, 0.04016428756713867, 0.03999334335327148, 0.040237056732177735, 0.0403394546508789, 0.04013260650634766, 0.04007219314575195, 0.04012236785888672, 0.04005785751342773, 0.03998822402954102, 0.040136703491210936, 0.040052734375, 0.040161312103271486, 0.03975164794921875, 0.040303680419921876, 0.04009260940551758, 0.04006707382202149, 0.04007628631591797, 0.04026265716552734, 0.04016128158569336, 0.040051712036132815, 0.040156158447265625, 0.04031795120239258, 0.04012748718261719, 0.04016230392456055, 0.03998617553710938, 0.04028518295288086, 0.04020633697509766, 0.04005068969726563, 0.0399288330078125, 0.040256511688232424, 0.039979007720947264, 0.04008550262451172, 0.04009164810180664, 0.04038860702514648, 0.040223743438720705, 0.04016844940185547, 0.04011724853515625, 0.04027699279785156, 0.04012441635131836, 0.04024729537963867, 0.040101886749267575, 0.04006399917602539, 0.04013158416748047, 0.04007526397705078, 0.039964672088623046, 0.04025958251953125, 0.04006707382202149, 0.040025089263916014, 0.04005068969726563, 0.03987968063354492, 0.03824639892578125, 0.038312961578369144, 0.03836006546020508, 0.038322208404541015, 0.03826992034912109, 0.03886899185180664, 0.03874611282348633, 0.0382751350402832, 0.038287296295166015, 0.08017305755615234, 0.03826483154296875, 0.038435840606689455, 0.03826073455810547, 0.03853107070922852, 0.04071014404296875, 0.03881062316894531, 0.03902771377563476, 0.040190975189208986, 0.04033331298828125, 0.04007526397705078, 0.04009471893310547, 0.0400711669921875, 0.04029747009277344, 0.03996364974975586, 0.040139774322509765, 0.04016128158569336, 0.04015206527709961, 0.04011315155029297, 0.040158206939697266, 0.039951358795166016, 0.04007526397705078, 0.04027699279785156, 0.03998720169067383, 0.04024729537963867, 0.03996672058105469, 0.0404398078918457, 0.040035327911376956, 0.040134654998779294, 0.04028108978271484, 0.03998822402954102, 0.039959552764892575, 0.040130561828613284, 0.04015411376953125, 0.04040499114990234, 0.04033433532714844, 0.04011929702758789, 0.040313854217529296, 0.04006614303588867, 0.039992225646972655, 0.04012543869018555, 0.04003942489624023, 0.04003635025024414, 0.04030976104736328, 0.04017561721801758, 0.03996364974975586, 0.04004048156738281, 0.04014076614379883, 0.040002559661865236, 0.04034969711303711, 0.04014284896850586, 0.04008345413208008, 0.04016844940185547, 0.03999948883056641, 0.039951358795166016, 0.03837644958496094, 0.03825766372680664, 0.03821158218383789, 0.038324222564697266, 0.039932926177978514, 0.040046592712402344, 0.04010291290283203, 0.04031283187866211, 0.08356454467773437, 0.040120319366455076, 0.040197120666503904, 0.04001484680175781, 0.04011520004272461, 0.0399738883972168, 0.04001587295532227, 0.040022014617919925, 0.040030208587646485, 0.03988684844970703, 0.040529918670654294, 0.04023910522460938, 0.039923713684082034, 0.0401797103881836, 0.04036505508422852, 0.040136703491210936, 0.039977985382080077, 0.04037222290039062, 0.040443904876708986, 0.04031999969482422, 0.04006092834472656, 0.04047359848022461, 0.040342529296875, 0.04001279830932617, 0.040022014617919925, 0.040134654998779294, 0.04007219314575195, 0.04011212921142578, 0.040000511169433595, 0.04015209579467773, 0.04024316787719726, 0.0400711669921875, 0.04026265716552734, 0.04001792144775391, 0.04007526397705078, 0.0404398078918457, 0.04011724853515625, 0.04022784042358398, 0.040081409454345705, 0.04112998580932617, 0.04065484619140625, 0.04024524688720703, 0.04050841522216797, 0.04051763153076172, 0.04026163101196289, 0.040182785034179686, 0.04026572799682617, 0.04014080047607422, 0.040338432312011716, 0.04013363265991211, 0.04012236785888672, 0.039981056213378906, 0.04031488037109375, 0.04002304077148437, 0.04011520004272461, 0.040033279418945314, 0.04029849624633789, 0.040041473388671874, 0.03832729721069336, 0.038545406341552735, 0.03859763336181641, 0.03841331100463867, 0.03827199935913086, 0.07982694244384765, 0.03820544052124023, 0.0380579833984375, 0.03835903930664063, 0.038199295043945314, 0.03845939254760742, 0.04019404983520508, 0.04046745681762695, 0.03990630340576172, 0.04013363265991211, 0.03996364974975586, 0.03836006546020508, 0.03817062377929688, 0.038171646118164065, 0.0381921272277832, 0.03843379211425781, 0.03821670532226563, 0.03819007873535156, 0.03816755294799805, 0.038141952514648435, 0.038299648284912106, 0.038242305755615234, 0.038491134643554685, 0.038255615234375, 0.03834982299804687, 0.038201343536376955, 0.03825151824951172, 0.038152191162109376, 0.03830886459350586, 0.03816755294799805, 0.038171646118164065, 0.03821263885498047, 0.03831600189208984, 0.03889561462402344, 0.03994112014770508, 0.04019302368164063, 0.03997183990478516, 0.04004761505126953, 0.03841843032836914, 0.03847679901123047, 0.038371326446533204, 0.038234111785888675, 0.03837849426269531, 0.038258689880371094, 0.03819007873535156, 0.03830886459350586, 0.03834982299804687, 0.03825254440307617, 0.03839590454101562, 0.038346752166748044, 0.03817779159545898, 0.03827199935913086, 0.03827609634399414, 0.03823311996459961, 0.03817987060546875, 0.03817059326171875, 0.0383221435546875, 0.038166526794433595, 0.03820646286010742, 0.03827609634399414, 0.03827507019042969, 0.038204414367675785, 0.03846553421020508, 0.07984639739990235, 0.03822284698486328, 0.03814604949951172, 0.038381568908691405, 0.038329345703125, 0.03812659072875976, 0.0381030387878418, 0.0381921272277832, 0.03803033447265625, 0.03826483154296875, 0.03832831954956055, 0.038163455963134765, 0.038128639221191404, 0.03808256149291992, 0.03830476760864258, 0.038152191162109376, 0.03814912033081055, 0.03819724655151367, 0.03816150283813476, 0.03834255981445313, 0.038214656829833986, 0.0388587532043457, 0.03990937423706055, 0.03986841583251953, 0.04010291290283203, 0.040019966125488284, 0.03821670532226563, 0.038198272705078126, 0.03840409469604492, 0.038588417053222655, 0.03825664138793945, 0.038371326446533204, 0.03809075164794922, 0.03829862213134766, 0.03820236968994141, 0.038201343536376955, 0.038217727661132815, 0.038234111785888675, 0.038209537506103515, 0.03817574310302734, 0.038247425079345705, 0.0382371826171875, 0.03834470367431641, 0.03826892852783203, 0.03811840057373047, 0.03825459289550781, 0.03821363067626953, 0.039229438781738284, 0.04001279830932617, 0.040089599609375, 0.0400076789855957, 0.040134654998779294, 0.040187904357910156, 0.040088607788085935, 0.039887840270996094, 0.04031283187866211, 0.04015411376953125, 0.04027084732055664, 0.04039273452758789, 0.03996566390991211, 0.03829145431518555, 0.03822079849243164, 0.04009471893310547, 0.08352470397949219, 0.0401376953125, 0.040349632263183596, 0.04077977752685547, 0.04024524688720703, 0.04009676742553711, 0.040515583038330076, 0.040248321533203124, 0.04014796829223633, 0.04009574508666992, 0.04011929702758789, 0.040210430145263674, 0.040223743438720705, 0.0399738883972168, 0.04003839874267578, 0.0382371826171875, 0.038629375457763675, 0.038491134643554685, 0.03825766372680664, 0.03826483154296875, 0.03846656036376953, 0.038354942321777344, 0.038373374938964845, 0.03835289764404297, 0.03834265518188477, 0.03841535949707031, 0.03825664138793945, 0.038163455963134765, 0.038193153381347655, 0.038416385650634766, 0.03832217788696289, 0.03832320022583008, 0.0391464958190918, 0.040323070526123043, 0.04012851333618164, 0.04029849624633789, 0.04007526397705078, 0.040306751251220706, 0.03839788818359375, 0.0381399040222168, 0.038373374938964845, 0.03852492904663086, 0.038386688232421876, 0.038383617401123046, 0.03821670532226563, 0.03838054275512695, 0.03832012939453125, 0.038214656829833986, 0.03820032119750977, 0.038507518768310545, 0.03825664138793945, 0.03829145431518555, 0.038245376586914064, 0.03831193542480469, 0.038317054748535154, 0.0382740478515625, 0.038214656829833986, 0.038345729827880856, 0.03892124938964844, 0.03984790420532226, 0.04002918243408203, 0.040172542572021484, 0.04021247863769531, 0.08194969940185547, 0.03841331100463867, 0.038258689880371094, 0.038373374938964845, 0.038317054748535154, 0.038228992462158204, 0.03824127960205078, 0.038424575805664066, 0.03839487838745117, 0.038250495910644534, 0.038561790466308594, 0.03838873672485352, 0.04093439865112305, 0.040123390197753905, 0.04016742324829101, 0.040235008239746094, 0.04011212921142578, 0.04035276794433594, 0.04093132781982422, 0.03868569564819336, 0.038564895629882814, 0.0386354866027832, 0.03827199935913086, 0.038193153381347655, 0.038217727661132815, 0.038247425079345705, 0.03840518569946289, 0.038314945220947264, 0.03830988693237305, 0.03851468658447266, 0.03837952041625976, 0.03823308944702149, 0.03839184188842774, 0.0384859848022461, 0.0382105598449707, 0.0382740478515625, 0.03916595077514649, 0.04033740615844727, 0.038247425079345705, 0.03826176071166992, 0.039144447326660156, 0.041452545166015625, 0.040622081756591794, 0.04063641738891602, 0.040891391754150394, 0.04014591979980469, 0.040700927734375, 0.04041318511962891, 0.040308734893798825, 0.04019404983520508, 0.03999129486083984, 0.04006707382202149, 0.040828929901123044, 0.04010905456542969, 0.04021145629882812, 0.04004249572753906, 0.04000460815429688, 0.04015718460083008, 0.04007014465332031, 0.04011315155029297, 0.03992473602294922, 0.040338432312011716, 0.04014899063110351, 0.08364339447021485, 0.03997491073608399, 0.04009267044067383, 0.03998207855224609, 0.040376319885253906, 0.04017766571044922, 0.04027494430541992, 0.04038451385498047, 0.0402503662109375, 0.04045414352416992, 0.04019814300537109, 0.0397916145324707, 0.03818700790405274, 0.0382105598449707, 0.03836108779907227, 0.03888435363769531, 0.04011929702758789, 0.040153087615966795, 0.039880702972412106, 0.03999846267700195, 0.04047872161865235, 0.040325119018554685, 0.040395774841308595, 0.040190975189208986, 0.040041473388671874, 0.040049663543701174, 0.03856281661987305, 0.0383375358581543, 0.038228992462158204, 0.038348800659179685, 0.03831398391723633, 0.03893657684326172, 0.040150016784667966, 0.04014796829223633, 0.03996876907348633, 0.040226814270019534, 0.03987865447998047, 0.04001792144775391, 0.040079360961914064, 0.04005376052856445, 0.040172542572021484, 0.04024422454833984, 0.04010291290283203, 0.04010905456542969, 0.0401162223815918, 0.040218624114990234, 0.03831808090209961, 0.03854848098754883, 0.038324222564697266, 0.038430721282958984, 0.038348800659179685, 0.03827302551269531, 0.03822796630859375, 0.039934974670410156, 0.04017766571044922, 0.038391807556152346, 0.038354942321777344, 0.03833446502685547, 0.03852799987792969, 0.03838771057128906, 0.038245376586914064, 0.03821875381469726, 0.03834163284301758, 0.08003788757324219, 0.0383631362915039, 0.03848601531982422, 0.03860172653198242, 0.038424575805664066, 0.03824127960205078, 0.03828224182128906, 0.03827814483642578, 0.03832217788696289, 0.038422527313232424, 0.03840921783447265, 0.0383559684753418, 0.03847782516479492, 0.038231040954589846, 0.03817375946044922, 0.038246337890625, 0.03848294448852539, 0.038414337158203124, 0.038351871490478515, 0.03842764663696289, 0.038779937744140625, 0.0384532470703125, 0.038394847869873044, 0.038340606689453126, 0.03857612609863281, 0.0383375358581543, 0.03840716934204102, 0.03837855911254883, 0.03835078430175781, 0.038474750518798825, 0.03840921783447265, 0.038712318420410154, 0.0397844467163086, 0.04009881591796875, 0.04036812973022461, 0.0402606086730957, 0.04012441635131836, 0.04048793411254883, 0.04021350479125976, 0.04022886276245117, 0.04016025543212891, 0.03828326416015625, 0.03829350280761719, 0.038520832061767575, 0.03837747192382813, 0.038302719116210936, 0.03847884750366211, 0.03830476760864258, 0.038267902374267575, 0.038665214538574216, 0.0383744010925293, 0.038373374938964845, 0.038384639739990234, 0.038225921630859375, 0.03869081497192383, 0.038553600311279294, 0.03964313507080078, 0.04035276794433594, 0.03975884628295898, 0.04010700988769531, 0.040389633178710936, 0.03976704025268555, 0.04012748718261719]",tokens/s,24.995868176413264,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1252.757504,923.271168,0.0,276.824064,221.070848,s,10,0.2186687355041504,0.02186687355041504,0.001250916400749165,0.02145625591278076,0.022023866271972656,0.023818652915954586,0.025254482231140137,"[0.025613439559936522, 0.02144879913330078, 0.021386432647705077, 0.02149398422241211, 0.021476255416870118, 0.021359647750854492, 0.021378175735473633, 0.021423263549804686, 0.02146371269226074, 0.021625024795532227]",tokens/s,11707.206309570536,kWh,2.5635838659449237e-07,1.404279138013745e-07,5.56265207604686e-07,9.530515080005529e-07,tokens/kWh,268610875.541316,MB,1252.757504,923.271168,0.0,276.824064,242.38336,s,10,13.0665751953125,1.30665751953125,0.0262609862113034,1.2976937866210938,1.3140078857421875,1.349322119140625,1.377573505859375,"[1.3846363525390626, 1.293762939453125, 1.29631689453125, 1.2978114013671875, 1.302518310546875, 1.297576171875, 1.299317626953125, 1.3061602783203126, 1.2932352294921876, 1.295239990234375]",tokens/s,48.214623233944735,kWh,1.5841335387840763e-05,8.680926422376672e-06,3.0770176520203394e-05,5.529243833042082e-05,tokens/kWh,1139396.3062999635,,s,629,13.234465784072873,0.021040486143200123,0.0027021366626035735,0.020561920166015626,0.021090917587280274,0.022328319931030273,0.04242145309448242,"[0.023242752075195314, 0.022978559494018554, 0.023216127395629883, 0.021728256225585937, 0.02227609634399414, 0.021982208251953125, 0.021679103851318358, 0.02128486442565918, 0.02167500877380371, 0.02414489555358887, 0.02450227165222168, 0.023540735244750977, 0.02306355285644531, 0.023653375625610353, 0.023060480117797853, 0.023163904190063478, 0.02403023910522461, 0.02313520050048828, 0.02290380859375, 0.022996992111206056, 0.022766592025756836, 0.022813695907592774, 0.02304204750061035, 0.02249318313598633, 0.022312959671020507, 0.02226688003540039, 0.022164480209350586, 0.022289407730102538, 0.022160383224487306, 0.022168575286865236, 0.02223308753967285, 0.022338560104370117, 0.022204416275024414, 0.02213382339477539, 0.022255552291870116, 0.022162431716918944, 0.02239897537231445, 0.022441984176635742, 0.02234880065917969, 0.020775999069213867, 0.02058028793334961, 0.020695039749145508, 0.020529151916503906, 0.02046668815612793, 0.021187583923339845, 0.022338560104370117, 0.020694015502929687, 0.02049843215942383, 0.02069196891784668, 0.020485120773315428, 0.02060697555541992, 0.020568063735961914, 0.020392959594726562, 0.020539392471313478, 0.020485120773315428, 0.02064384078979492, 0.020553728103637696, 0.020525056838989256, 0.020534271240234374, 0.02058956718444824, 0.02087014389038086, 0.02067148780822754, 0.042489887237548825, 0.02053430366516113, 0.0206232967376709, 0.020575231552124023, 0.020535295486450195, 0.0206878719329834, 0.020593664169311524, 0.020530176162719727, 0.020584447860717774, 0.020536319732666015, 0.020551679611206054, 0.02069196891784668, 0.02062950325012207, 0.02060492706298828, 0.020594688415527345, 0.020602880477905275, 0.020567039489746093, 0.020600831985473633, 0.020560895919799805, 0.020493312835693358, 0.020563968658447264, 0.020485120773315428, 0.0204902400970459, 0.020620288848876952, 0.020557823181152343, 0.020495391845703124, 0.02051888084411621, 0.020427776336669923, 0.0202608642578125, 0.020364288330078126, 0.020299776077270508, 0.020521984100341797, 0.020535295486450195, 0.020510719299316405, 0.020353023529052734, 0.02057318305969238, 0.02050048065185547, 0.020557823181152343, 0.020505599975585938, 0.020539392471313478, 0.020512767791748047, 0.020601856231689454, 0.020538400650024415, 0.02051171112060547, 0.02050764846801758, 0.02051481628417969, 0.020541439056396483, 0.020520959854125977, 0.02050764846801758, 0.020496383666992187, 0.020496383666992187, 0.020590591430664062, 0.020517887115478514, 0.020625408172607423, 0.020518911361694335, 0.02049126434326172, 0.020518911361694335, 0.020594688415527345, 0.02050764846801758, 0.020351999282836913, 0.02037555122375488, 0.020494335174560546, 0.02064691162109375, 0.042352638244628905, 0.020494335174560546, 0.020503551483154296, 0.02063974380493164, 0.020539392471313478, 0.02052403259277344, 0.020561920166015626, 0.020761600494384767, 0.02068889617919922, 0.0208721923828125, 0.020967424392700194, 0.020716543197631835, 0.020562944412231447, 0.020682752609252928, 0.020610048294067384, 0.02059878349304199, 0.02062950325012207, 0.020531200408935548, 0.020674560546875, 0.020593664169311524, 0.020536319732666015, 0.020540416717529295, 0.020337663650512695, 0.0206376953125, 0.02059775924682617, 0.020510719299316405, 0.02053222465515137, 0.02062950325012207, 0.020717567443847656, 0.020404224395751954, 0.020495359420776366, 0.020584447860717774, 0.020616191864013672, 0.020570112228393556, 0.020521984100341797, 0.020533248901367186, 0.020523008346557618, 0.020590591430664062, 0.02050764846801758, 0.020562944412231447, 0.020509695053100584, 0.020625408172607423, 0.020527103424072265, 0.020494367599487303, 0.02051273536682129, 0.020579328536987306, 0.020549631118774413, 0.02053222465515137, 0.02048102378845215, 0.020518911361694335, 0.020322303771972656, 0.020533248901367186, 0.020410367965698242, 0.020560895919799805, 0.020486143112182616, 0.020587520599365236, 0.020567039489746093, 0.020531200408935548, 0.020552736282348633, 0.02054652786254883, 0.020722688674926756, 0.020563968658447264, 0.020551679611206054, 0.042552318572998044, 0.020709375381469726, 0.020546560287475587, 0.020601856231689454, 0.020523008346557618, 0.020569087982177735, 0.02048409652709961, 0.020641792297363282, 0.020539392471313478, 0.020744224548339844, 0.02050249671936035, 0.020703231811523438, 0.020577280044555665, 0.02069811248779297, 0.020571136474609376, 0.02062950325012207, 0.02062950325012207, 0.02064384078979492, 0.020580352783203124, 0.020594688415527345, 0.020528127670288086, 0.020545568466186524, 0.02054140853881836, 0.02062335968017578, 0.020397056579589845, 0.020380672454833985, 0.02050764846801758, 0.020599807739257812, 0.020564992904663085, 0.020497407913208008, 0.020572160720825194, 0.02057935905456543, 0.02091414451599121, 0.02101862335205078, 0.020585472106933594, 0.02066739273071289, 0.020513792037963868, 0.020563968658447264, 0.02063564872741699, 0.020547584533691408, 0.020536319732666015, 0.020591615676879883, 0.02062851142883301, 0.020594655990600588, 0.020576255798339844, 0.020628543853759767, 0.020568000793457032, 0.020570112228393556, 0.021001279830932616, 0.020595647811889647, 0.020566015243530272, 0.020588544845581053, 0.020571136474609376, 0.020582399368286132, 0.02065100860595703, 0.020513792037963868, 0.020549631118774413, 0.020580352783203124, 0.020611072540283205, 0.020583423614501953, 0.020381696701049806, 0.02044927978515625, 0.02059878349304199, 0.04248064041137695, 0.020577280044555665, 0.02065510368347168, 0.02061414337158203, 0.020553728103637696, 0.020611072540283205, 0.020556800842285155, 0.02046668815612793, 0.020591615676879883, 0.021213184356689452, 0.021947391510009767, 0.021411840438842773, 0.021089279174804687, 0.020612096786499022, 0.020584447860717774, 0.020591615676879883, 0.020405248641967775, 0.020489215850830078, 0.02067251205444336, 0.021012479782104493, 0.020358144760131838, 0.020561920166015626, 0.02050048065185547, 0.020580352783203124, 0.020579328536987306, 0.02066739273071289, 0.020556800842285155, 0.020529151916503906, 0.021124095916748048, 0.02055379295349121, 0.020545536041259766, 0.020809663772583007, 0.020587520599365236, 0.020545536041259766, 0.02060697555541992, 0.02087731170654297, 0.020470783233642577, 0.020610048294067384, 0.020619264602661135, 0.02062950325012207, 0.020578304290771485, 0.02047488021850586, 0.020858879089355468, 0.020595712661743162, 0.02063974380493164, 0.02050764846801758, 0.020478975296020507, 0.020806655883789063, 0.021513216018676756, 0.020787200927734374, 0.020588544845581053, 0.020602880477905275, 0.02066431999206543, 0.020719615936279297, 0.02061414337158203, 0.020509695053100584, 0.02023628807067871, 0.02044927978515625, 0.020331520080566406, 0.020452352523803712, 0.020840448379516603, 0.020552703857421875, 0.020767744064331056, 0.043447296142578126, 0.02059878349304199, 0.020603904724121092, 0.020523008346557618, 0.020747264862060546, 0.020539424896240235, 0.02051068878173828, 0.02050048065185547, 0.020538368225097657, 0.020745216369628908, 0.0206059513092041, 0.020575231552124023, 0.021097471237182617, 0.020747264862060546, 0.020771839141845702, 0.020595712661743162, 0.020606016159057616, 0.02056697654724121, 0.02046566390991211, 0.02066739273071289, 0.020583423614501953, 0.02050764846801758, 0.020566015243530272, 0.020515871047973634, 0.020701152801513672, 0.020546560287475587, 0.020494335174560546, 0.020369407653808593, 0.0206059513092041, 0.02062851142883301, 0.02028028869628906, 0.020420608520507814, 0.020525056838989256, 0.02056403160095215, 0.020686784744262696, 0.020556800842285155, 0.020468736648559572, 0.020571136474609376, 0.020611072540283205, 0.02065203285217285, 0.02048409652709961, 0.02049126434326172, 0.02085785675048828, 0.020570175170898437, 0.020587520599365236, 0.020465599060058594, 0.02060492706298828, 0.020534271240234374, 0.020523008346557618, 0.020501504898071288, 0.02047488021850586, 0.020641792297363282, 0.020533248901367186, 0.02065100860595703, 0.020568063735961914, 0.020550655364990233, 0.020247552871704103, 0.02027724838256836, 0.020558847427368163, 0.020603904724121092, 0.020521984100341797, 0.02083635139465332, 0.020580352783203124, 0.042395648956298826, 0.020556800842285155, 0.02047590446472168, 0.020356096267700196, 0.020523008346557618, 0.020510719299316405, 0.020341791152954102, 0.02055369567871094, 0.020486143112182616, 0.020486143112182616, 0.020486143112182616, 0.020566015243530272, 0.020582399368286132, 0.020536319732666015, 0.02146201515197754, 0.020786176681518553, 0.020455423355102538, 0.020503551483154296, 0.02048409652709961, 0.020644863128662108, 0.020470783233642577, 0.020447231292724608, 0.020805631637573242, 0.020953088760375976, 0.020516864776611327, 0.02041347122192383, 0.02047792053222656, 0.020509695053100584, 0.020539392471313478, 0.02085683250427246, 0.020497407913208008, 0.020527103424072265, 0.020557823181152343, 0.020508672714233397, 0.020575231552124023, 0.02048307228088379, 0.02091007995605469, 0.020774944305419922, 0.020592607498168946, 0.020666368484497072, 0.020610048294067384, 0.020616191864013672, 0.02069811248779297, 0.020561920166015626, 0.020557823181152343, 0.020783103942871094, 0.021014528274536134, 0.02146099281311035, 0.021011455535888672, 0.020527103424072265, 0.020526079177856444, 0.020627519607543946, 0.020879295349121092, 0.020744192123413087, 0.020495359420776366, 0.020609024047851563, 0.020487167358398437, 0.020555776596069338, 0.020533248901367186, 0.020497407913208008, 0.02048307228088379, 0.02059775924682617, 0.020574207305908202, 0.045780990600585936, 0.021638143539428712, 0.02102681541442871, 0.020553728103637696, 0.02049228858947754, 0.020720640182495118, 0.020853759765625, 0.022123519897460937, 0.020996095657348633, 0.020619264602661135, 0.020547584533691408, 0.020780031204223632, 0.02050662422180176, 0.02049126434326172, 0.020526144027709962, 0.02056083106994629, 0.021545984268188476, 0.02124799919128418, 0.02043084716796875, 0.020514848709106446, 0.02052092742919922, 0.022012928009033202, 0.02085478401184082, 0.02046156883239746, 0.020520959854125977, 0.020323328018188477, 0.02045644760131836, 0.020510719299316405, 0.020523008346557618, 0.02069196891784668, 0.02062950325012207, 0.020541439056396483, 0.020528127670288086, 0.02045849609375, 0.02049843215942383, 0.020503551483154296, 0.020410367965698242, 0.02046566390991211, 0.020601856231689454, 0.020535295486450195, 0.02045644760131836, 0.020550655364990233, 0.02044927978515625, 0.02102579116821289, 0.020771839141845702, 0.020747264862060546, 0.020593664169311524, 0.020584447860717774, 0.02064793586730957, 0.02061516761779785, 0.02047488021850586, 0.02050764846801758, 0.020477951049804686, 0.020580352783203124, 0.02051584053039551, 0.020676607131958007, 0.020502527236938475, 0.020486143112182616, 0.020540416717529295, 0.02058137512207031, 0.020513792037963868, 0.020727807998657227, 0.020579328536987306, 0.042431488037109374, 0.020518911361694335, 0.020339712142944336, 0.02053228759765625, 0.020612031936645507, 0.020486143112182616, 0.020549631118774413, 0.02045644760131836, 0.020229120254516602, 0.020518911361694335, 0.020516864776611327, 0.020511743545532226, 0.02033459281921387, 0.02058137512207031, 0.02086604881286621, 0.02064793586730957, 0.020518911361694335, 0.020547584533691408, 0.02048204803466797, 0.020517887115478514, 0.020511743545532226, 0.02050048065185547, 0.020535295486450195, 0.02062848091125488, 0.020544511795043945, 0.02044927978515625, 0.02051481628417969, 0.02046976089477539, 0.020519935607910156, 0.020634624481201173, 0.020563968658447264, 0.02050662422180176, 0.020558847427368163, 0.020495359420776366, 0.020512767791748047, 0.020600831985473633, 0.02048307228088379, 0.020540416717529295, 0.020595712661743162, 0.020552703857421875, 0.020435968399047853, 0.020527103424072265, 0.02070528030395508, 0.020520959854125977, 0.020529151916503906, 0.020473920822143554, 0.020646848678588868, 0.020419584274291993, 0.020462591171264647, 0.02042982482910156, 0.02060492706298828, 0.02048204803466797, 0.020488256454467772, 0.020568000793457032, 0.020580352783203124, 0.02041651153564453, 0.02043187141418457, 0.020575231552124023, 0.020538368225097657, 0.02044211196899414, 0.020553728103637696, 0.02045849609375, 0.020525056838989256, 0.04266294479370117, 0.020557792663574218, 0.020547584533691408, 0.020497407913208008, 0.020550655364990233, 0.020546560287475587, 0.02102681541442871, 0.020626432418823244, 0.02062233543395996, 0.020593664169311524, 0.02050662422180176, 0.020501504898071288, 0.020551679611206054, 0.020669439315795898, 0.020591615676879883, 0.020568063735961914, 0.02061516761779785, 0.020502527236938475, 0.020520959854125977, 0.020561920166015626, 0.020361215591430663, 0.020413440704345705, 0.020508672714233397, 0.020519935607910156, 0.020547584533691408, 0.02048102378845215, 0.02056399917602539, 0.02051171112060547, 0.020494335174560546, 0.020504575729370117, 0.02082713508605957, 0.02069606399536133, 0.020711423873901368, 0.020574207305908202, 0.020533248901367186, 0.020487167358398437, 0.020395008087158203, 0.020505599975585938, 0.020632575988769532, 0.020602880477905275, 0.02043801689147949, 0.02039193534851074, 0.020512800216674804, 0.020453407287597657, 0.020487104415893555, 0.020625408172607423, 0.020685823440551757, 0.020545536041259766, 0.020653055191040038, 0.020521984100341797, 0.020478975296020507, 0.020774911880493165, 0.020530176162719727, 0.020578304290771485, 0.020487199783325194, 0.020478944778442382, 0.02046156883239746, 0.020590591430664062, 0.020546560287475587, 0.02046463966369629, 0.020502527236938475, 0.02049228858947754, 0.02049843215942383]",tokens/s,47.52741895762616,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, @@ -9156,48 +9039,33 @@ ChildProcessError: Traceback (most recent call last): RuntimeError: shape '[-1, 128, 4672]' is invalid for input of size 21229568 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 86.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -9298,48 +9166,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2_moe,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,mixtral,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1355.640832,1313.34144,0.0,666.894336,583.411712,s,10,0.3753043823242187,0.03753043823242187,0.0017061726371702826,0.03732486343383789,0.038417790222167965,0.04030316772460937,0.0418114697265625,"[0.04218854522705078, 0.03567526245117188, 0.03605964660644531, 0.03759164810180664, 0.03732332611083984, 0.03643900680541992, 0.03732640075683594, 0.037700160980224606, 0.03799881744384766, 0.03700156784057617]",tokens/s,6821.130049551251,kWh,4.375063655553041e-07,2.397299215873419e-07,1.3726420446149002e-06,2.0498783317575464e-06,tokens/kWh,124885460.77781504,MB,1355.640832,1313.34144,0.0,666.894336,644.020224,s,10,23.046878906249997,2.3046878906250003,0.010530589688708434,2.310258544921875,2.3126857177734372,2.313240197753906,2.3136837817382814,"[2.311134521484375, 2.282283935546875, 2.28828759765625, 2.3125625, 2.311003662109375, 2.313794677734375, 2.311868896484375, 2.307242431640625, 2.309513427734375, 2.299187255859375]",tokens/s,27.335588587188372,kWh,2.6983760501499532e-05,1.4787944706284982e-05,7.386716814719235e-05,0.00011563887335497687,tokens/kWh,544799.4966762499,,s,629,23.341892585754387,0.037109527163361515,0.004422100805265259,0.03662847900390625,0.03704197158813476,0.03741410675048828,0.07203065734863284,"[0.035879936218261715, 0.03652710342407227, 0.036673534393310545, 0.03687936019897461, 0.03612268829345703, 0.03798419189453125, 0.03842559814453125, 0.037560382843017576, 0.038442943572998045, 0.036365310668945314, 0.037185535430908204, 0.036550655364990234, 0.03583078384399414, 0.03694694519042969, 0.03779174423217774, 0.037651454925537106, 0.037125118255615236, 0.03663359832763672, 0.03691929626464844, 0.0362608642578125, 0.036383743286132815, 0.03656806564331055, 0.03676160049438477, 0.03685171127319336, 0.03665203094482422, 0.036334590911865236, 0.0366376953125, 0.03631923294067383, 0.036729854583740236, 0.03656294250488281, 0.03668172836303711, 0.03625983810424805, 0.03639603042602539, 0.036574207305908206, 0.036863998413085936, 0.036544513702392575, 0.036560897827148435, 0.036261886596679685, 0.0364288330078125, 0.03603152084350586, 0.036614078521728516, 0.03667967987060547, 0.036686847686767575, 0.03632844924926758, 0.03665407943725586, 0.036160511016845705, 0.03701862335205078, 0.036318206787109376, 0.03648409652709961, 0.03649433517456055, 0.03591987228393555, 0.036713470458984376, 0.03633868789672851, 0.03641856002807617, 0.03654870223999023, 0.036613025665283204, 0.036326400756835936, 0.036843521118164066, 0.036278270721435545, 0.036585472106933595, 0.03630387115478516, 0.03623219299316406, 0.07103282928466798, 0.03485388946533203, 0.0347248649597168, 0.035108863830566404, 0.03509862518310547, 0.03478015899658203, 0.03513241577148438, 0.03518463897705078, 0.03530956649780274, 0.03479449462890625, 0.03498495864868164, 0.035130367279052735, 0.03459481430053711, 0.03486822509765625, 0.03480473709106445, 0.0363059196472168, 0.035458206176757816, 0.03472880172729492, 0.037623809814453124, 0.037416961669921874, 0.036972545623779295, 0.03635200119018555, 0.03666329574584961, 0.03677801513671875, 0.03643286514282226, 0.03608063888549805, 0.03639910507202149, 0.03655379104614258, 0.03633145523071289, 0.03641139221191406, 0.03648409652709961, 0.03645753479003906, 0.03661203384399414, 0.03655372619628906, 0.036703262329101566, 0.03689878463745117, 0.036653057098388675, 0.03629363250732422, 0.0367646713256836, 0.03653222274780273, 0.03689267349243164, 0.03652403259277344, 0.036391937255859375, 0.036729854583740236, 0.036588542938232424, 0.03617996978759765, 0.036708351135253905, 0.03844095993041992, 0.03687833786010742, 0.0367534065246582, 0.03680153656005859, 0.03652710342407227, 0.03646771240234375, 0.03656294250488281, 0.03664896011352539, 0.03679743957519531, 0.03657932662963867, 0.036597824096679686, 0.036764606475830075, 0.03685068893432617, 0.036675582885742186, 0.03703910446166992, 0.03671244812011719, 0.07131033325195313, 0.0348171501159668, 0.034966400146484375, 0.03515084838867188, 0.03504947280883789, 0.035148799896240236, 0.03495116806030273, 0.03530752182006836, 0.03618815994262695, 0.03583180618286133, 0.036429824829101565, 0.036746238708496096, 0.03651686477661133, 0.03638886260986328, 0.03627315139770508, 0.03682099151611328, 0.03655680084228516, 0.03652505493164063, 0.03661209487915039, 0.03641548919677735, 0.036465663909912106, 0.03628851318359375, 0.03663564682006836, 0.03638579177856445, 0.03676774215698242, 0.03656512069702148, 0.0366242561340332, 0.03668377685546875, 0.03650252914428711, 0.036977664947509765, 0.036348926544189454, 0.03617484664916992, 0.0358737907409668, 0.03630499267578125, 0.03663759994506836, 0.036116481781005856, 0.03590655899047852, 0.03667763137817383, 0.036582401275634766, 0.03663564682006836, 0.03637964630126953, 0.03645542526245117, 0.0366110725402832, 0.036580352783203124, 0.03642060852050781, 0.03667251205444336, 0.036450302124023434, 0.036241409301757815, 0.03650559997558594, 0.035737598419189456, 0.03617587280273438, 0.036582401275634766, 0.036708351135253905, 0.03638681411743164, 0.03687014389038086, 0.03681792068481445, 0.037059585571289064, 0.037009471893310546, 0.03683116912841797, 0.03663257598876953, 0.03653836822509766, 0.03680460739135742, 0.03676671981811523, 0.07476121520996094, 0.036734977722167966, 0.03669094467163086, 0.03636019134521484, 0.03656294250488281, 0.03658342361450195, 0.036819969177246094, 0.03642572784423828, 0.036560897827148435, 0.03648819351196289, 0.03648716735839844, 0.036776065826416016, 0.036683647155761716, 0.03653017425537109, 0.03601919937133789, 0.036030494689941406, 0.03646358489990234, 0.03705344009399414, 0.0367564811706543, 0.037028865814208986, 0.03644723129272461, 0.03841535949707031, 0.03753881454467774, 0.03693056106567383, 0.03657932662963867, 0.036241409301757815, 0.036604927062988284, 0.036397056579589845, 0.036435966491699216, 0.03657318496704102, 0.036587520599365236, 0.036606975555419925, 0.03623731231689453, 0.03642367935180664, 0.03631411361694336, 0.036944896697998046, 0.03647999954223633, 0.036850753784179686, 0.036499393463134765, 0.03657011032104492, 0.03669510269165039, 0.03658438491821289, 0.03695001602172852, 0.03663270568847656, 0.03635494232177734, 0.036596736907958984, 0.03721011352539062, 0.03652710342407227, 0.03677798461914063, 0.03775590515136719, 0.037362686157226564, 0.037087230682373046, 0.036805633544921876, 0.036760574340820314, 0.0365865592956543, 0.036950977325439456, 0.03668172836303711, 0.036603904724121096, 0.03699302291870117, 0.036676769256591794, 0.03656585693359375, 0.03657932662963867, 0.03695001602172852, 0.07426457977294922, 0.03659571075439453, 0.03641139221191406, 0.03670943832397461, 0.036706241607666015, 0.03663052749633789, 0.03669401550292969, 0.03667865753173828, 0.036647937774658204, 0.03653222274780273, 0.03715891265869141, 0.03657625579833984, 0.03671142578125, 0.03600588989257812, 0.036898815155029296, 0.036296798706054685, 0.036683681488037106, 0.036462593078613284, 0.03666124725341797, 0.03696844863891602, 0.03650048065185547, 0.03650457763671875, 0.03650559997558594, 0.03646156692504883, 0.03740982437133789, 0.0369837760925293, 0.03661004638671875, 0.03638988876342773, 0.03660595321655274, 0.03637152099609375, 0.0366069107055664, 0.036760574340820314, 0.036805633544921876, 0.03635200119018555, 0.03690496063232422, 0.036939777374267575, 0.036633663177490235, 0.03697657775878906, 0.03657011032104492, 0.03658860778808594, 0.03654137420654297, 0.03666124725341797, 0.037074081420898436, 0.036469600677490235, 0.03672371292114258, 0.03642367935180664, 0.0369172477722168, 0.03655372619628906, 0.03650867080688477, 0.036498432159423826, 0.037064704895019535, 0.036811775207519534, 0.036483070373535154, 0.03680361557006836, 0.03691523361206055, 0.03660486221313477, 0.03713433456420898, 0.0365588493347168, 0.036877311706542966, 0.03642879867553711, 0.036311038970947264, 0.03691929626464844, 0.037266433715820314, 0.07231078338623047, 0.03545600128173828, 0.0361973762512207, 0.036348926544189454, 0.03648614501953125, 0.03696332931518555, 0.03681792068481445, 0.0367534065246582, 0.03653222274780273, 0.03669708633422852, 0.03669305419921875, 0.036628414154052734, 0.03664588928222656, 0.036825088500976565, 0.0366192626953125, 0.03630284881591797, 0.03681075286865235, 0.036746238708496096, 0.03674524688720703, 0.0365700798034668, 0.03599871826171875, 0.036734977722167966, 0.036803585052490234, 0.03670425415039062, 0.036705280303955076, 0.03671654510498047, 0.037028865814208986, 0.03733196640014649, 0.036705280303955076, 0.03640524673461914, 0.03664691162109375, 0.03685273742675781, 0.03743231964111328, 0.03712102508544922, 0.03689574432373047, 0.03719680023193359, 0.03839692687988281, 0.03721420669555664, 0.036805633544921876, 0.036775936126708986, 0.03689267349243164, 0.036429824829101565, 0.03629584121704101, 0.036797279357910155, 0.03661619186401367, 0.03638886260986328, 0.03702783966064453, 0.036763713836669924, 0.03662022399902344, 0.03649228668212891, 0.036782081604003904, 0.03670943832397461, 0.03677484893798828, 0.037125118255615236, 0.036915199279785156, 0.03764534378051758, 0.036789215087890625, 0.03653030395507813, 0.036683647155761716, 0.036675582885742186, 0.03653529739379883, 0.03642060852050781, 0.0367749137878418, 0.07448268890380859, 0.037364734649658206, 0.03663974380493164, 0.036640766143798825, 0.036689918518066404, 0.036332542419433594, 0.036566017150878906, 0.03655475234985352, 0.03644211196899414, 0.03668787384033203, 0.03665206527709961, 0.03657827377319336, 0.03700940704345703, 0.03664486312866211, 0.037817344665527344, 0.036803585052490234, 0.03686707305908203, 0.036827136993408206, 0.03720191955566406, 0.03696640014648438, 0.03661721420288086, 0.0366561279296875, 0.036947967529296875, 0.03726243209838867, 0.036662174224853517, 0.03674726486206055, 0.03662847900390625, 0.03632128143310547, 0.036722686767578124, 0.036751361846923826, 0.0367749137878418, 0.03662847900390625, 0.036789249420166016, 0.03681484985351562, 0.03640729522705078, 0.03652608108520508, 0.036170753479003906, 0.0367534065246582, 0.036201473236083984, 0.03642367935180664, 0.03683430480957031, 0.036587520599365236, 0.03666739273071289, 0.03632844924926758, 0.03636019134521484, 0.0369244155883789, 0.03647180938720703, 0.03628646469116211, 0.03655987167358398, 0.036741119384765625, 0.036441089630126954, 0.03647283172607422, 0.03648614501953125, 0.03655168151855469, 0.03670425415039062, 0.03662335968017578, 0.03659775924682617, 0.03648921585083008, 0.03693056106567383, 0.036877311706542966, 0.03659366226196289, 0.03682918548583984, 0.03728076934814453, 0.07459532928466797, 0.03663872146606445, 0.03683020782470703, 0.036759552001953126, 0.036125694274902344, 0.03650876617431641, 0.036785057067871094, 0.03703705596923828, 0.037082111358642575, 0.0363397102355957, 0.03700838470458984, 0.036347904205322266, 0.03688550567626953, 0.03657011032104492, 0.036395008087158204, 0.03655475234985352, 0.03666124725341797, 0.036416511535644534, 0.036863998413085936, 0.036557823181152346, 0.03657628631591797, 0.036175838470458986, 0.03641139221191406, 0.03650668716430664, 0.03668268966674805, 0.036604927062988284, 0.03665817642211914, 0.03649228668212891, 0.036647937774658204, 0.035955711364746096, 0.036209663391113284, 0.03589324951171875, 0.036765697479248044, 0.03657318496704102, 0.03697049713134765, 0.03668070220947266, 0.036997119903564454, 0.03711795043945312, 0.03787571334838867, 0.037082111358642575, 0.036744190216064454, 0.037059585571289064, 0.0366192626953125, 0.03680255889892578, 0.03654246520996094, 0.03559833526611328, 0.035659774780273434, 0.036397056579589845, 0.03662335968017578, 0.03649228668212891, 0.03666534423828125, 0.03685887908935547, 0.03707392120361328, 0.036880382537841795, 0.036691967010498046, 0.035915775299072264, 0.03681689453125, 0.03593318557739258, 0.03673907089233398, 0.03676160049438477, 0.03672576141357422, 0.03676979064941406, 0.036751361846923826, 0.07499673461914062, 0.03750912094116211, 0.03742105484008789, 0.036563968658447264, 0.036923393249511716, 0.036736000061035154, 0.03690598297119141, 0.03644518280029297, 0.03664281463623047, 0.03659980773925781, 0.03663564682006836, 0.036657150268554685, 0.03683327865600586, 0.036310016632080076, 0.03660083389282227, 0.03654860687255859, 0.0364769287109375, 0.03663872146606445, 0.03670220947265625, 0.03670425415039062, 0.03665817642211914, 0.03661414337158203, 0.036468734741210936, 0.036523006439208985, 0.03629568099975586, 0.03698995208740234, 0.03718860626220703, 0.036592735290527346, 0.036276126861572264, 0.03672780990600586, 0.036506622314453126, 0.03644825744628906, 0.03649945449829101, 0.036478977203369144, 0.036232223510742186, 0.03654038238525391, 0.036674560546875, 0.036603904724121096, 0.036805633544921876, 0.03661312103271484, 0.0366929931640625, 0.03685068893432617, 0.03621478271484375, 0.03703910446166992, 0.03681075286865235, 0.036674560546875, 0.036751361846923826, 0.03650457763671875, 0.03661209487915039, 0.03638988876342773, 0.0363694076538086, 0.036770816802978515, 0.036536319732666016, 0.03663359832763672, 0.03691110229492187, 0.036640766143798825, 0.03666841506958008, 0.036380672454833986, 0.03698995208740234, 0.03657011032104492, 0.0366376953125, 0.03661414337158203, 0.03671039962768555, 0.0741048355102539, 0.03667251205444336, 0.03667660903930664, 0.03663359832763672, 0.036534271240234374, 0.03634483337402344, 0.036893695831298826, 0.036549633026123046, 0.03674009704589844, 0.03668070220947266, 0.03635302352905274, 0.0366561279296875, 0.036705280303955076, 0.036689918518066404, 0.03700121688842774, 0.036915199279785156, 0.03671449661254883, 0.036563968658447264, 0.034473983764648435, 0.03458355331420898, 0.034985984802246094, 0.03528499221801758, 0.03512841415405273, 0.034746273040771485, 0.03503308868408203, 0.03523379135131836, 0.03520409774780273, 0.03676876831054687, 0.036634624481201174, 0.03648614501953125, 0.03609600067138672, 0.03650867080688477, 0.03644416046142578, 0.03657734298706055, 0.03598944091796875, 0.03677798461914063, 0.03691417694091797, 0.03661619186401367, 0.036721664428710936, 0.036603904724121096, 0.03657625579833984, 0.036746238708496096, 0.03682515335083008, 0.03706156921386719, 0.03660083389282227, 0.037147647857666014, 0.0368721923828125, 0.03668172836303711, 0.03661004638671875, 0.03752755355834961, 0.03676979064941406, 0.03687526321411133, 0.03750400161743164, 0.037838848114013675, 0.03719379043579102, 0.03667859268188477, 0.037438465118408204, 0.03693363189697266, 0.03649433517456055, 0.03671551895141602, 0.03680255889892578, 0.03664588928222656, 0.0364400634765625]",tokens/s,26.947257926457944,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -9892,48 +9745,33 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1280.995328,1017.643008,0.0,371.195904,315.446272,s,10,0.19319747543334959,0.01931974754333496,0.0010467105269549052,0.01894753646850586,0.019827580642700193,0.021101390457153316,0.022120438308715823,"[0.022375200271606447, 0.01901312065124512, 0.018929792404174806, 0.018751968383789064, 0.019248479843139647, 0.018965280532836915, 0.019544511795043944, 0.018915008544921875, 0.018772640228271485, 0.018681472778320312]",tokens/s,13250.690746645725,kWh,2.25981147787691e-07,1.238256752321744e-07,5.593843438318166e-07,9.091911668516819e-07,tokens/kWh,281568947.58059365,MB,1281.323008,1017.643008,0.0,371.195904,340.902912,s,10,11.583821777343749,1.1583821777343748,0.009627584898594657,1.1564221801757812,1.16910693359375,1.1740381469726562,1.1779831176757813,"[1.1789693603515625, 1.1680111083984375, 1.1652188720703125, 1.14978955078125, 1.1497177734375, 1.1605277099609375, 1.14636669921875, 1.15310693359375, 1.1523763427734375, 1.1597374267578124]",tokens/s,54.38619586086756,kWh,1.370763023194169e-05,7.511351751950323e-06,3.022333472675601e-05,5.144231671064802e-05,tokens/kWh,1224672.6824991468,,s,629,11.732324380874624,0.018652344007749816,0.002332081824858889,0.018243648529052733,0.01894564514160156,0.01906585578918457,0.03724599380493164,"[0.021345279693603517, 0.02056822395324707, 0.018622304916381834, 0.018869312286376953, 0.018895904541015626, 0.01879747200012207, 0.01884569549560547, 0.018906112670898437, 0.01901055908203125, 0.01845964813232422, 0.018405376434326173, 0.018534400939941405, 0.018283519744873047, 0.018355199813842774, 0.018320383071899413, 0.018226175308227538, 0.01817398452758789, 0.017952735900878907, 0.018078720092773438, 0.018127872467041017, 0.018137088775634767, 0.018027519226074217, 0.018348031997680665, 0.01796403121948242, 0.01798041534423828, 0.017889280319213868, 0.01796505546569824, 0.018174976348876954, 0.01842995262145996, 0.01819647979736328, 0.018330623626708984, 0.018799615859985352, 0.018984960556030273, 0.019307519912719725, 0.019362815856933592, 0.018956287384033203, 0.01905971145629883, 0.018949119567871094, 0.018910207748413087, 0.01904128074645996, 0.018909183502197266, 0.018541568756103514, 0.019268608093261717, 0.02164851188659668, 0.019189632415771485, 0.0190699520111084, 0.018994176864624023, 0.019017728805541992, 0.01874028778076172, 0.018768831253051756, 0.01828659248352051, 0.01800294494628906, 0.018103391647338866, 0.018309024810791014, 0.018288639068603514, 0.018249727249145507, 0.018296831130981444, 0.01827020835876465, 0.018159616470336915, 0.018174976348876954, 0.018149375915527344, 0.018314239501953124, 0.03768937683105469, 0.018380767822265626, 0.01881395149230957, 0.018351104736328124, 0.018108415603637695, 0.018318336486816408, 0.018051071166992186, 0.01820979118347168, 0.018351104736328124, 0.018297855377197265, 0.01827436828613281, 0.018240447998046874, 0.018114559173583983, 0.018911359786987304, 0.018926464080810546, 0.018877439498901367, 0.018922496795654296, 0.01882624053955078, 0.018942975997924806, 0.019012672424316406, 0.019114944458007814, 0.018929664611816405, 0.018882560729980468, 0.018909183502197266, 0.018893823623657227, 0.01883545684814453, 0.018930688858032226, 0.019140640258789064, 0.018999263763427733, 0.018945024490356444, 0.01903104019165039, 0.01900032043457031, 0.01904332733154297, 0.018966527938842775, 0.019555328369140625, 0.019763200759887696, 0.019224576950073242, 0.018750463485717773, 0.018890752792358398, 0.018163711547851562, 0.018000896453857423, 0.017994752883911135, 0.01820159912109375, 0.018122751235961913, 0.018229248046875, 0.018174976348876954, 0.018365440368652345, 0.018112512588500978, 0.01814630317687988, 0.018116607666015624, 0.01814630317687988, 0.01823846435546875, 0.01822003173828125, 0.01822822380065918, 0.01809715270996094, 0.01807257652282715, 0.01807360076904297, 0.01816985511779785, 0.018189311981201172, 0.018229248046875, 0.0181790714263916, 0.018077695846557617, 0.018066432952880858, 0.03762995147705078, 0.018239488601684572, 0.018173952102661133, 0.018126911163330078, 0.018182079315185548, 0.01829478454589844, 0.018241535186767577, 0.01802444839477539, 0.018283519744873047, 0.01821696090698242, 0.01819443130493164, 0.018512895584106445, 0.01808076858520508, 0.018274303436279296, 0.01821183967590332, 0.018199552536010744, 0.018395135879516602, 0.018157567977905274, 0.018331743240356444, 0.018150304794311522, 0.01825382423400879, 0.01842790412902832, 0.018110464096069336, 0.018165760040283203, 0.018052095413208007, 0.018263200759887695, 0.018966367721557617, 0.018345983505249023, 0.018093055725097656, 0.018099199295043944, 0.017991680145263672, 0.018292736053466797, 0.018103296279907227, 0.018127872467041017, 0.01846272087097168, 0.018257919311523436, 0.018172927856445312, 0.018128896713256838, 0.018699264526367186, 0.019384319305419923, 0.018778112411499022, 0.019094528198242186, 0.018982912063598634, 0.018769920349121092, 0.018692096710205077, 0.018780160903930664, 0.018891775131225585, 0.018884607315063476, 0.019056640625, 0.018760704040527345, 0.01886412811279297, 0.018914304733276367, 0.018719743728637696, 0.018747392654418944, 0.01879449653625488, 0.018683904647827147, 0.018913280487060546, 0.01885798454284668, 0.019105791091918945, 0.018896896362304686, 0.018958335876464845, 0.018868288040161132, 0.018834367752075195, 0.03720499038696289, 0.017994752883911135, 0.0180633602142334, 0.01810246467590332, 0.018297664642333983, 0.018044927597045898, 0.01820979118347168, 0.01815449523925781, 0.018136064529418947, 0.018220064163208007, 0.01806947135925293, 0.018532352447509767, 0.018326528549194337, 0.01820467185974121, 0.01804902458190918, 0.018113567352294923, 0.018221023559570313, 0.018256895065307616, 0.01828556823730469, 0.018092031478881835, 0.02020966339111328, 0.019123199462890626, 0.01904025650024414, 0.018780160903930664, 0.01865830421447754, 0.018556928634643553, 0.018364416122436524, 0.01815260887145996, 0.01814313507080078, 0.018269119262695314, 0.018164735794067383, 0.018215936660766603, 0.017910911560058595, 0.01787481689453125, 0.01821286392211914, 0.01820057678222656, 0.018172927856445312, 0.018020351409912108, 0.01806540870666504, 0.018712575912475587, 0.018587648391723634, 0.017889280319213868, 0.01821696090698242, 0.01816985511779785, 0.018085887908935547, 0.018095104217529297, 0.01798761558532715, 0.018116575241088867, 0.01822003173828125, 0.0180849609375, 0.017928096771240236, 0.01814227294921875, 0.018178016662597656, 0.01807561683654785, 0.01817087936401367, 0.0182476806640625, 0.018051071166992186, 0.017951744079589844, 0.018018304824829103, 0.01820476722717285, 0.017925024032592773, 0.01824665641784668, 0.018471935272216796, 0.038201343536376955, 0.019075071334838867, 0.019009536743164062, 0.018894975662231445, 0.018753408432006836, 0.018868223190307617, 0.01863065528869629, 0.018965503692626954, 0.018865152359008788, 0.018891775131225585, 0.01866444778442383, 0.018408447265625, 0.018060287475585936, 0.017885183334350584, 0.017976320266723633, 0.017949695587158202, 0.01811155128479004, 0.01834592056274414, 0.018148351669311523, 0.017930240631103517, 0.017978368759155275, 0.0182609920501709, 0.01817190361022949, 0.018254911422729492, 0.018199487686157225, 0.018199583053588868, 0.017978336334228514, 0.01819340705871582, 0.018198528289794923, 0.01797340774536133, 0.0180928955078125, 0.018215936660766603, 0.01821286392211914, 0.018172927856445312, 0.018160640716552736, 0.018123775482177733, 0.01808076858520508, 0.018069503784179687, 0.018173952102661133, 0.01807257652282715, 0.01820979118347168, 0.018010112762451173, 0.01825382423400879, 0.01825279998779297, 0.01800601577758789, 0.018143232345581056, 0.018050048828125, 0.01795686340332031, 0.01807257652282715, 0.018137088775634767, 0.018118656158447266, 0.018051071166992186, 0.017949695587158202, 0.018066495895385743, 0.018194368362426758, 0.01843404769897461, 0.01803775978088379, 0.017875999450683595, 0.01807766342163086, 0.01803059196472168, 0.01820467185974121, 0.018163711547851562, 0.018373632431030275, 0.037416961669921874, 0.018315263748168945, 0.01838489532470703, 0.018367584228515626, 0.018121631622314453, 0.018227231979370116, 0.01852310371398926, 0.018521087646484375, 0.019975168228149414, 0.01922355270385742, 0.018580480575561522, 0.018808832168579103, 0.018928640365600585, 0.01879347229003906, 0.018755712509155274, 0.01847488021850586, 0.018259967803955078, 0.018225151062011717, 0.01827737617492676, 0.018280448913574218, 0.018578432083129884, 0.018981887817382814, 0.01879449653625488, 0.018777088165283205, 0.018954240798950195, 0.018790399551391602, 0.01884262466430664, 0.018937856674194335, 0.018939903259277344, 0.018922496795654296, 0.01882009506225586, 0.018137088775634767, 0.01798963165283203, 0.018159616470336915, 0.018094079971313477, 0.018282495498657226, 0.01803264045715332, 0.017977344512939454, 0.01803980827331543, 0.01808076858520508, 0.018320383071899413, 0.01829478454589844, 0.018257919311523436, 0.018310144424438478, 0.017977344512939454, 0.018115583419799804, 0.01820876884460449, 0.018452640533447265, 0.017978208541870117, 0.018363391876220703, 0.018113536834716795, 0.018053119659423827, 0.018121728897094725, 0.01816985511779785, 0.018192384719848635, 0.018215936660766603, 0.018172927856445312, 0.01824460792541504, 0.018326528549194337, 0.01824460792541504, 0.018246816635131835, 0.018262880325317383, 0.018293760299682618, 0.03905228805541992, 0.018197504043579102, 0.018062335968017578, 0.01804083251953125, 0.01804902458190918, 0.018176000595092775, 0.018295808792114256, 0.017986560821533205, 0.018333696365356447, 0.018181119918823242, 0.01824358367919922, 0.018193536758422852, 0.018217920303344726, 0.018208703994750976, 0.018136064529418947, 0.018336864471435548, 0.018207647323608397, 0.018132991790771484, 0.017987583160400392, 0.01828659248352051, 0.01802342414855957, 0.01806540870666504, 0.01798041534423828, 0.01827123260498047, 0.018158592224121094, 0.01825279998779297, 0.018174976348876954, 0.01807155227661133, 0.018310144424438478, 0.01829478454589844, 0.01805414390563965, 0.01807155227661133, 0.018274303436279296, 0.01821900749206543, 0.01820569610595703, 0.018152448654174806, 0.018167903900146484, 0.018721696853637695, 0.01816268730163574, 0.018124799728393554, 0.017999872207641602, 0.018158592224121094, 0.01822412872314453, 0.01817190361022949, 0.018019327163696287, 0.017936544418334963, 0.017980255126953126, 0.018255903244018556, 0.017989599227905273, 0.018241632461547853, 0.01825987243652344, 0.018539520263671876, 0.018529279708862305, 0.018309247970581054, 0.01804889678955078, 0.018128896713256838, 0.01839411163330078, 0.01815350341796875, 0.01822319984436035, 0.01798643112182617, 0.01789030456542969, 0.01817094421386719, 0.018261951446533205, 0.0376627197265625, 0.01943654441833496, 0.01903308868408203, 0.018584575653076172, 0.018528255462646484, 0.018967552185058592, 0.01887539291381836, 0.01881804847717285, 0.018651136398315428, 0.01829478454589844, 0.01781760025024414, 0.01783296012878418, 0.017933311462402343, 0.018304000854492186, 0.018899967193603515, 0.018166784286499024, 0.018104352951049806, 0.017926111221313476, 0.01844633674621582, 0.01827737617492676, 0.018300928115844727, 0.018282495498657226, 0.018314239501953124, 0.018314239501953124, 0.01830297660827637, 0.018214975357055664, 0.018388927459716795, 0.018646015167236327, 0.01819443130493164, 0.017946624755859376, 0.018058240890502928, 0.017987583160400392, 0.018267135620117187, 0.01808896064758301, 0.01830297660827637, 0.01825382423400879, 0.01807257652282715, 0.018144256591796876, 0.01823846435546875, 0.018239488601684572, 0.01830297660827637, 0.01802239990234375, 0.018226175308227538, 0.01816166305541992, 0.01818435287475586, 0.018342784881591797, 0.018327520370483397, 0.018268320083618166, 0.018170719146728517, 0.018379776000976563, 0.018341888427734376, 0.018197568893432617, 0.018314176559448243, 0.018314239501953124, 0.018317312240600587, 0.018132991790771484, 0.01810758399963379, 0.018072383880615234, 0.018077695846557617, 0.018281471252441405, 0.018387968063354493, 0.018068479537963866, 0.017932287216186525, 0.037222400665283206, 0.018480127334594726, 0.018143232345581056, 0.018041856765747072, 0.018266111373901366, 0.01818726348876953, 0.01836953544616699, 0.018258943557739257, 0.01836953544616699, 0.018282495498657226, 0.018149375915527344, 0.018415615081787108, 0.018137088775634767, 0.01807360076904297, 0.01810534477233887, 0.018094079971313477, 0.018118656158447266, 0.01843097686767578, 0.01861529541015625, 0.01822719955444336, 0.01824051284790039, 0.01816985511779785, 0.01846681594848633, 0.018869247436523438, 0.018137088775634767, 0.018048063278198242, 0.018235328674316407, 0.018192384719848635, 0.01823232078552246, 0.018298879623413086, 0.018140159606933593, 0.01807360076904297, 0.01808793640136719, 0.018168832778930662, 0.018542591094970702, 0.018525184631347655, 0.018155519485473632, 0.01807161521911621, 0.017998783111572266, 0.018185216903686522, 0.017945600509643556, 0.018742271423339844, 0.018907136917114258, 0.01886310386657715, 0.01881088066101074, 0.01864816093444824, 0.01880668830871582, 0.01886412811279297, 0.018655231475830078, 0.017969215393066406, 0.018292671203613282, 0.017846368789672853, 0.018097055435180663, 0.018155519485473632, 0.018289663314819335, 0.01828659248352051, 0.018322431564331054, 0.018144256591796876, 0.018168832778930662, 0.018215936660766603, 0.01809715270996094, 0.018126848220825196, 0.018100223541259765, 0.03725516891479492, 0.018487295150756835, 0.018094079971313477, 0.018165760040283203, 0.018243648529052733, 0.018313152313232422, 0.018050048828125, 0.017924095153808595, 0.01808291244506836, 0.018131872177124024, 0.01863680076599121, 0.0184453125, 0.01882009506225586, 0.019468448638916017, 0.01891004753112793, 0.018972671508789063, 0.018988031387329102, 0.01880678367614746, 0.018761728286743166, 0.018769920349121092, 0.018964479446411133, 0.018885631561279297, 0.01905356788635254, 0.01882828712463379, 0.01894812774658203, 0.018978784561157227, 0.018922496795654296, 0.018893888473510742, 0.018397119522094725, 0.018312192916870116, 0.01841049575805664, 0.018309120178222657, 0.018355199813842774, 0.018215967178344727, 0.018113504409790038, 0.018373632431030275, 0.017963008880615236, 0.01823641586303711, 0.018147327423095702, 0.01879859161376953, 0.018380800247192384, 0.018332672119140626, 0.018139135360717772, 0.018008064270019532, 0.018236448287963867, 0.01829680061340332, 0.01824051284790039, 0.01799577522277832, 0.01802444839477539, 0.01838591957092285, 0.018339839935302735, 0.01810534477233887, 0.01785651206970215, 0.017928192138671875, 0.018020511627197266, 0.01837552070617676, 0.01826304054260254, 0.018059263229370116, 0.01823744010925293, 0.01820159912109375, 0.018020351409912108, 0.018190336227416993, 0.018371583938598633]",tokens/s,53.61256470417404,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,4009.381888,12218.53184,0.0,11572.084736,11067.830272,s,10,6.230304809570312,0.6230304809570313,0.001216742381839673,0.6227249145507812,0.625022705078125,0.6253476257324219,0.6256075622558593,"[0.6223314208984375, 0.6219492797851562, 0.623067138671875, 0.6228473510742187, 0.6228085327148437, 0.6217056274414062, 0.6226412963867187, 0.6249505004882813, 0.6223311157226562, 0.6256725463867188]",tokens/s,410.8948242897535,kWh,7.348882986439599e-06,4.026913079488782e-06,3.961776045235301e-05,5.0993556518281394e-05,tokens/kWh,5020242.114476227,MB,4009.381888,12218.53184,0.0,11572.084736,11417.44384,s,10,338.039375,33.8039375,0.002081163339272258,33.80410546875,33.80615429687499,33.806391601562495,33.806581445312496,"[33.8018515625, 33.8005546875, 33.805953125, 33.8061015625, 33.80662890625, 33.8039609375, 33.80425, 33.8032109375, 33.80115234375, 33.8057109375]",tokens/s,1.8636882167942712,kWh,0.0003990108090970251,0.00021869259385011794,0.0021626746124912507,0.0027803780154383936,tokens/kWh,22658.789434452687,,s,629,343.1073381958008,0.5454806648581888,0.07408968048294812,0.5365196533203125,0.5368479614257813,0.536985595703125,1.1595506787109375,"[0.5361858520507813, 0.5362073364257812, 0.5362841796875, 0.5364111328125, 0.5360650634765625, 0.5361038818359375, 0.5362667236328125, 0.536321044921875, 0.5364100952148437, 0.5362626342773438, 0.5362349853515626, 0.53634765625, 0.5364408569335938, 0.5362196655273438, 0.5362616577148438, 0.5365176391601563, 0.5366251220703125, 0.53658935546875, 0.5364049682617188, 0.5366445922851563, 0.536431640625, 0.536553466796875, 0.5363967895507813, 0.5365227661132812, 0.5363916625976562, 0.5366077270507813, 0.5364869384765625, 0.5368197021484375, 0.536510498046875, 0.536553466796875, 0.5362565307617188, 0.5363148803710938, 0.53648486328125, 0.5367255249023437, 0.5364561767578125, 0.5369609985351562, 0.53667431640625, 0.5362073364257812, 0.5365841674804688, 0.53644287109375, 0.5365555419921875, 0.5366876220703125, 0.5368094482421875, 0.5368944702148437, 0.5367091064453124, 0.5365360717773437, 0.5365299072265625, 0.5365791015625, 0.536508544921875, 0.5366076049804688, 0.5366968383789062, 0.5368914184570313, 0.5367470092773438, 0.5367080688476562, 0.5367398681640625, 0.5366343383789063, 0.5366242065429687, 0.5366281127929687, 0.5367459716796875, 0.5371760864257813, 0.5366435546875, 0.5366405029296875, 1.1591527099609376, 0.5361038818359375, 0.5361018676757813, 0.536263671875, 0.5362094116210937, 0.5359912719726563, 0.5361182861328125, 0.5362841796875, 0.5365995483398438, 0.536662109375, 0.5365564575195313, 0.5362994995117187, 0.5364090576171875, 0.536268798828125, 0.5362012329101562, 0.536300537109375, 0.5364172973632813, 0.5365196533203125, 0.5369375610351562, 0.5364633178710938, 0.5365770263671875, 0.536489990234375, 0.536511474609375, 0.5364254760742188, 0.5364879150390625, 0.53650634765625, 0.5367684936523438, 0.5366251220703125, 0.5367716064453125, 0.5364961547851562, 0.5364869384765625, 0.5362063598632812, 0.536541259765625, 0.5363352661132812, 0.5366138916015625, 0.5362503662109375, 0.536637451171875, 0.5363988647460938, 0.5364029541015625, 0.536647705078125, 0.53635888671875, 0.5366128540039062, 0.5366681518554688, 0.5366998901367187, 0.5365596313476563, 0.5366507568359375, 0.53677978515625, 0.5365350341796875, 0.5365678100585938, 0.5364971313476562, 0.536750244140625, 0.5368533325195313, 0.5366978759765625, 0.5362718505859375, 0.5368770751953125, 0.5368719482421875, 0.53650634765625, 0.5365380859375, 0.5366179809570313, 0.536711181640625, 0.5368678588867187, 0.5365084228515625, 0.5365411376953125, 1.1604705810546876, 0.536459228515625, 0.5363988647460938, 0.5366630249023437, 0.5363927001953125, 0.5361377563476563, 0.536052734375, 0.53621044921875, 0.5363937377929687, 0.5364285278320312, 0.5365135498046875, 0.5362964477539063, 0.5364961547851562, 0.5361838989257812, 0.5362216186523437, 0.5362186279296876, 0.5362073364257812, 0.5364940795898437, 0.5364930419921875, 0.5364019165039062, 0.5366181030273437, 0.5363280639648438, 0.5363056640625, 0.5379686279296875, 0.5366077270507813, 0.5363896484375, 0.5363722534179688, 0.53621142578125, 0.5367080688476562, 0.5363896484375, 0.5365463256835937, 0.53640087890625, 0.5365811157226562, 0.5365125122070312, 0.536753173828125, 0.536416259765625, 0.5366016235351563, 0.5365422973632813, 0.5365462646484375, 0.5365678100585938, 0.5364786987304687, 0.53650634765625, 0.5365770263671875, 0.5364725952148437, 0.5364940795898437, 0.53674609375, 0.5368380737304688, 0.5385840454101563, 0.5366784057617188, 0.53697021484375, 0.53692724609375, 0.5369292602539063, 0.536827880859375, 0.5362872314453125, 0.5366261596679688, 0.5368565673828125, 0.537196533203125, 0.5369241333007813, 0.5368299560546875, 0.5368002319335937, 0.5368504028320312, 0.5364971313476562, 0.5366364135742188, 1.1599544677734375, 0.5360076904296875, 0.5364992065429688, 0.5362135009765625, 0.5361295166015625, 0.5361039428710938, 0.5363466186523438, 0.5364234008789063, 0.5364490356445313, 0.5364490356445313, 0.5363384399414063, 0.5363200073242187, 0.5377433471679688, 0.5363619995117187, 0.5364172973632813, 0.5363722534179688, 0.5364459228515625, 0.5362647094726563, 0.5363363647460937, 0.5366569213867187, 0.5368955078125, 0.5366435546875, 0.5365043334960937, 0.536322021484375, 0.5366005859375, 0.5361510620117188, 0.5364786987304687, 0.5361930541992187, 0.5367756958007812, 0.5365288696289062, 0.5366661376953125, 0.5365718994140625, 0.5365186767578125, 0.5364132080078124, 0.5367725830078125, 0.5362708740234375, 0.5365698852539063, 0.5366261596679688, 0.5365104370117187, 0.536580078125, 0.5365493774414063, 0.5365054321289062, 0.536596435546875, 0.5365155639648438, 0.5391400756835938, 0.536680419921875, 0.5366333618164062, 0.5368473510742188, 0.5367808227539063, 0.5367900390625, 0.5366754760742187, 0.5367847900390625, 0.5369497680664063, 0.5367030029296875, 0.5367142333984375, 0.5368402099609375, 0.5365791015625, 0.536537109375, 0.5366845703125, 0.5366978759765625, 0.5368914184570313, 0.5366712036132812, 0.5367449340820313, 1.160353759765625, 0.5361346435546875, 0.5365831909179688, 0.5365841674804688, 0.5365616455078125, 0.5362667236328125, 0.5360650024414062, 0.5360445556640625, 0.5363937377929687, 0.53633740234375, 0.536453125, 0.5361203002929688, 0.5364715576171875, 0.5362421264648437, 0.536171630859375, 0.5363198852539063, 0.5363650512695313, 0.5378069458007813, 0.5365338745117187, 0.5363650512695313, 0.5366364135742188, 0.5364306030273438, 0.5364429931640625, 0.536397705078125, 0.53661181640625, 0.53633740234375, 0.5363927001953125, 0.5363558349609375, 0.5372119140625, 0.5364357299804687, 0.5366210327148437, 0.5363937377929687, 0.5364910888671875, 0.5363844604492187, 0.5366077270507813, 0.5363722534179688, 0.5364838256835938, 0.5362329711914062, 0.5363148803710938, 0.5363414916992187, 0.536573974609375, 0.536474609375, 0.5364971923828125, 0.5366926879882813, 0.5367337036132812, 0.5366364135742188, 0.5366098022460938, 0.5364827880859375, 0.5366292724609375, 0.5366312866210937, 0.5369446411132812, 0.5368104858398437, 0.5367337036132812, 0.5364234008789063, 0.5366712036132812, 0.5366630249023437, 0.53658740234375, 0.5400728149414062, 0.5371002807617188, 0.537069580078125, 0.5373511962890625, 0.5366661376953125, 0.5367490844726562, 1.1589600830078124, 0.5360947265625, 0.536141845703125, 0.53627392578125, 0.5362769775390624, 0.5361336059570313, 0.536121337890625, 0.5363435668945312, 0.5363701782226562, 0.5363619995117187, 0.5364551391601563, 0.5362340087890625, 0.5364981689453125, 0.5366824951171875, 0.5364613037109375, 0.53637939453125, 0.5361807250976562, 0.5362708740234375, 0.5365606689453125, 0.5362247924804687, 0.5365718994140625, 0.5362657470703125, 0.5364121704101562, 0.5365299072265625, 0.5366445922851563, 0.5364019165039062, 0.5365278930664062, 0.536384521484375, 0.536816650390625, 0.5365125122070312, 0.5366333618164062, 0.5365258178710938, 0.5368606567382812, 0.53661181640625, 0.5364807739257812, 0.5361397705078125, 0.5365555419921875, 0.5365155639648438, 0.5399480590820313, 0.5366773681640625, 0.53646337890625, 0.5363404541015625, 0.5365657348632813, 0.5365616455078125, 0.5366026000976563, 0.5365616455078125, 0.5365657348632813, 0.5363281860351562, 0.5365985107421875, 0.5366835327148437, 0.5366210327148437, 0.5367255249023437, 0.5375068359375, 0.5364992065429688, 0.5366947631835938, 0.536764404296875, 0.5366251220703125, 0.5365504150390625, 0.5366343383789063, 0.53659033203125, 0.5368985595703125, 0.5366558837890625, 0.5366343383789063, 1.1597823486328125, 0.5358673706054687, 0.536205322265625, 0.5361622924804688, 0.536258544921875, 0.5363712158203126, 0.5364418334960938, 0.5364623413085937, 0.5364715576171875, 0.5363302612304688, 0.5364930419921875, 0.53625341796875, 0.5364490356445313, 0.5365145874023437, 0.53642138671875, 0.5365514526367188, 0.5367142333984375, 0.5362974853515625, 0.536384521484375, 0.53635888671875, 0.5364592895507813, 0.5362432250976562, 0.5363446044921875, 0.5386342163085938, 0.5366722412109375, 0.5365196533203125, 0.53634765625, 0.5361971435546875, 0.5368033447265625, 0.5364520874023437, 0.536658935546875, 0.5361612548828125, 0.5363824462890625, 0.5366333618164062, 0.53676953125, 0.5364285278320312, 0.5366610107421875, 0.5364459228515625, 0.5364572143554688, 0.536479736328125, 0.5364132080078124, 0.5366138916015625, 0.5366507568359375, 0.5371443481445313, 0.536655029296875, 0.53648779296875, 0.5365811157226562, 0.5365380859375, 0.5367193603515625, 0.536656982421875, 0.5365994873046875, 0.536816650390625, 0.536933349609375, 0.536574951171875, 0.5366651000976562, 0.5367971801757813, 0.5366210327148437, 0.53647265625, 0.5367152099609375, 0.5368822021484375, 0.536890380859375, 0.5368883056640625, 0.5370931396484375, 1.1623106689453124, 0.5358325805664063, 0.5361561889648437, 0.5362392578125, 0.536377197265625, 0.5361858520507813, 0.536184814453125, 0.5361663818359375, 0.536474609375, 0.5364551391601563, 0.5365575561523438, 0.536453125, 0.5364387817382813, 0.5364971313476562, 0.5363200073242187, 0.5364295654296874, 0.5377137451171875, 0.5364940185546875, 0.5364981689453125, 0.5364070434570313, 0.5366548461914062, 0.536290283203125, 0.5363271484375, 0.5363353881835937, 0.5365288696289062, 0.5366047973632813, 0.5365256958007812, 0.5362319946289062, 0.5367490234375, 0.5363404541015625, 0.5365657348632813, 0.536394775390625, 0.536468505859375, 0.536543212890625, 0.5367501220703125, 0.5364920043945313, 0.5366876220703125, 0.5363660278320312, 0.5365309448242187, 0.5366896362304687, 0.53644287109375, 0.53656884765625, 0.5368340454101562, 0.5368033447265625, 0.536943603515625, 0.536679443359375, 0.5365780639648438, 0.5364777221679687, 0.53671728515625, 0.53667431640625, 0.5367234497070312, 0.537006103515625, 0.536953857421875, 0.53675830078125, 0.5368831787109375, 0.5366773681640625, 0.5365708618164062, 0.53650537109375, 0.5366497802734375, 0.53669580078125, 0.536648681640625, 0.5364735717773438, 0.5366621704101563, 1.1597054443359376, 0.53639990234375, 0.5362175903320312, 0.5362780151367188, 0.5361448974609375, 0.536247314453125, 0.5360568237304687, 0.5364736938476562, 0.5365626220703125, 0.5363394775390625, 0.53640087890625, 0.5362821044921875, 0.5363189697265625, 0.5361725463867187, 0.536069091796875, 0.5361285400390625, 0.5363609619140625, 0.5378284301757813, 0.5366773071289063, 0.5363701782226562, 0.5366312866210937, 0.5362667236328125, 0.5362298583984375, 0.5362298583984375, 0.5367193603515625, 0.5366599731445313, 0.53675927734375, 0.5367686157226562, 0.5367100219726563, 0.536322021484375, 0.5364961547851562, 0.5364592895507813, 0.5364735717773438, 0.5364807739257812, 0.5365616455078125, 0.5365258178710938, 0.53671728515625, 0.5364623413085937, 0.5364080810546875, 0.53646435546875, 0.5363988647460938, 0.5365821533203124, 0.5366937866210938, 0.5366118774414063, 0.5366373901367187, 0.5368207397460938, 0.5365913696289063, 0.5362380981445313, 0.5364398193359375, 0.53650634765625, 0.536754150390625, 0.5364541625976562, 0.5366640625, 0.5365985107421875, 0.53686376953125, 0.5367234497070312, 0.5365729370117187, 0.5365166015625, 0.5366917114257812, 0.5367388305664063, 0.5368545532226563, 0.5365196533203125, 0.536605712890625, 1.1630111083984376, 0.535952392578125, 0.5362728881835938, 0.5364592895507813, 0.5362789916992188, 0.53621044921875, 0.5360670776367188, 0.5362063598632812, 0.5364869384765625, 0.5363455810546875, 0.5363486938476563, 0.5361018676757813, 0.5364193115234375, 0.5362063598632812, 0.5360137939453125, 0.5360752563476563, 0.5363312377929688, 0.536426513671875, 0.536796142578125, 0.53637939453125, 0.5366405029296875, 0.5363671264648437, 0.5363568725585938, 0.5364091186523438, 0.5364940185546875, 0.5363896484375, 0.5387694091796875, 0.5365647583007812, 0.5367736206054687, 0.5362626342773438, 0.536474609375, 0.5363302612304688, 0.536427490234375, 0.536369140625, 0.5367982177734375, 0.5363814697265625, 0.5368094482421875, 0.536711181640625, 0.5365145874023437, 0.5365894165039062, 0.536303466796875, 0.5364930419921875, 0.5365411987304688, 0.53646337890625, 0.536594482421875, 0.5366395874023437, 0.5366680297851563, 0.5364961547851562, 0.53675830078125, 0.5365463256835937, 0.5366353759765625, 0.53671630859375, 0.5367705688476563, 0.53673681640625, 0.5369097900390625, 0.536932373046875, 0.537027587890625, 0.5365985107421875, 0.5367613525390625, 0.536838134765625, 0.536995849609375, 0.5366098022460938, 0.5366384887695312]",tokens/s,1.8332455473192157,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,4901.904384,5954.338816,0.0,5307.891712,5047.542272,s,10,3.4311535644531252,0.3431153564453125,0.0005038409151925485,0.3432273101806641,0.3436598815917969,0.3437290954589844,0.34378446655273437,"[0.3431885375976563, 0.34266885375976563, 0.34253961181640624, 0.34248388671875, 0.3432660827636719, 0.3435628356933594, 0.34364450073242186, 0.3435517578125, 0.3424491882324219, 0.3437983093261719]",tokens/s,746.1047580387228,kWh,4.048144022623698e-06,2.2182009251991985e-06,2.0107016085600207e-05,2.6373361033423104e-05,tokens/kWh,9706764.324636886,MB,4901.904384,5956.435968,0.0,5309.988864,5098.334208,s,10,188.48311328124998,18.848311328125,0.0009324486308739057,18.8480966796875,18.8496212890625,18.84981552734375,18.84997091796875,"[18.850009765625, 18.849115234375, 18.84820703125, 18.847986328125, 18.848294921875, 18.847962890625, 18.849578125, 18.84715234375, 18.846998046875, 18.84780859375]",tokens/s,3.342474500938071,kWh,0.00022252111416723992,0.00012195848409537575,0.0010845150342780015,0.0014289946325406172,tokens/kWh,44086.93956253143,,s,629,191.27160934448247,0.30408840913272245,0.04077856997503405,0.2991708068847656,0.29932748413085936,0.29945037231445315,0.64191595703125,"[0.2991697998046875, 0.2990540771484375, 0.2991175537109375, 0.2990592041015625, 0.29923019409179685, 0.29914111328125, 0.2991964111328125, 0.29955072021484375, 0.299082763671875, 0.2989486083984375, 0.2992650146484375, 0.29918618774414063, 0.2990018615722656, 0.2991523742675781, 0.29931622314453127, 0.2990110778808594, 0.29905612182617186, 0.29913088989257813, 0.29915545654296877, 0.2989014892578125, 0.29901824951171874, 0.29899365234375, 0.2990950317382812, 0.29925274658203127, 0.2991974487304688, 0.2991718444824219, 0.2994155578613281, 0.29913088989257813, 0.29904281616210937, 0.2991298522949219, 0.29918618774414063, 0.2991912841796875, 0.29923019409179685, 0.29919436645507813, 0.29950054931640624, 0.2992209777832031, 0.2992384033203125, 0.2991401062011719, 0.299114501953125, 0.2991124572753906, 0.2991575012207031, 0.29937255859375, 0.2991073303222656, 0.2991575012207031, 0.29959271240234375, 0.29922918701171874, 0.2991933288574219, 0.2992568359375, 0.299156494140625, 0.29919436645507813, 0.29918310546875, 0.29925784301757813, 0.2993653869628906, 0.2991708068847656, 0.2991984558105469, 0.299315185546875, 0.2992056274414063, 0.2991431579589844, 0.29914520263671873, 0.2992127990722656, 0.2992650146484375, 0.29922305297851565, 0.642862060546875, 0.2992332763671875, 0.2990223388671875, 0.2990592041015625, 0.29894552612304687, 0.2993438720703125, 0.29911041259765625, 0.2991718444824219, 0.2988984375, 0.2996254577636719, 0.298967041015625, 0.2990653381347656, 0.2991994934082031, 0.29911141967773436, 0.29905816650390626, 0.29907763671875, 0.299082763671875, 0.29911346435546876, 0.2990469055175781, 0.2991206359863281, 0.29902334594726565, 0.2991964111328125, 0.29902642822265624, 0.2991022033691406, 0.29904177856445313, 0.29907763671875, 0.29909811401367187, 0.2991964111328125, 0.2992814025878906, 0.29920254516601563, 0.29913088989257813, 0.2991513671875, 0.29914111328125, 0.29920870971679686, 0.29918106079101564, 0.2991278076171875, 0.29918618774414063, 0.29922509765625, 0.29908990478515624, 0.299298828125, 0.299146240234375, 0.29921893310546877, 0.2992906188964844, 0.2992322692871094, 0.299514892578125, 0.29945242309570314, 0.2991401062011719, 0.29927218627929686, 0.2996971435546875, 0.2993438720703125, 0.2991749267578125, 0.2992599182128906, 0.29926400756835936, 0.2993131408691406, 0.29910324096679686, 0.29924658203125, 0.2992353210449219, 0.29931622314453127, 0.29932135009765626, 0.29919540405273437, 0.29922918701171874, 0.2991626281738281, 0.2993039245605469, 0.6417797241210937, 0.2990848083496094, 0.2990192565917969, 0.29909197998046877, 0.2990673828125, 0.2990489501953125, 0.29906842041015624, 0.29916571044921875, 0.2989916076660156, 0.29911654663085935, 0.29913189697265624, 0.2990172119140625, 0.29910015869140627, 0.2991493225097656, 0.29930804443359377, 0.2992568359375, 0.2990049133300781, 0.29910324096679686, 0.2990643310546875, 0.29920050048828123, 0.29905816650390626, 0.29928857421875, 0.29907763671875, 0.29903564453125, 0.2991431579589844, 0.29914727783203127, 0.2990848083496094, 0.29923123168945315, 0.29905816650390626, 0.2991933288574219, 0.299072509765625, 0.2991349792480469, 0.2992476196289062, 0.29907559204101564, 0.2991206359863281, 0.2991155090332031, 0.2993326110839844, 0.2991851501464844, 0.29917901611328124, 0.29917593383789065, 0.2991595458984375, 0.2993602600097656, 0.2992435302734375, 0.2993489990234375, 0.299261962890625, 0.29966848754882813, 0.2992537536621094, 0.29918310546875, 0.2992353210449219, 0.2992762756347656, 0.29911859130859375, 0.29929266357421874, 0.2993029174804688, 0.2993029174804688, 0.2990868530273437, 0.29918206787109375, 0.2992179260253906, 0.2993879089355469, 0.29927423095703126, 0.299251708984375, 0.29920254516601563, 0.2991544189453125, 0.2992957458496094, 0.6418739013671875, 0.29907354736328123, 0.29900799560546876, 0.299040771484375, 0.298998779296875, 0.29912884521484373, 0.29905816650390626, 0.29907559204101564, 0.29908377075195314, 0.29907864379882815, 0.29897625732421873, 0.29901824951171874, 0.29899777221679685, 0.2992814025878906, 0.29921893310546877, 0.2990387268066406, 0.29944525146484374, 0.29907046508789065, 0.29900082397460936, 0.2991728515625, 0.298903564453125, 0.2992015380859375, 0.29904177856445313, 0.2990469055175781, 0.2990458984375, 0.2991728515625, 0.29911346435546876, 0.29913394165039064, 0.29911346435546876, 0.29922406005859375, 0.29911859130859375, 0.2991851501464844, 0.2991513671875, 0.29923019409179685, 0.29922509765625, 0.29916363525390627, 0.299209716796875, 0.29918719482421874, 0.29921383666992185, 0.2992547912597656, 0.2993960876464844, 0.2992209777832031, 0.29927218627929686, 0.2992117614746094, 0.2991728515625, 0.2992557983398437, 0.29905816650390626, 0.29917901611328124, 0.299298828125, 0.2992414855957031, 0.2998128662109375, 0.29929470825195315, 0.2991708068847656, 0.29912472534179685, 0.29926400756835936, 0.29932748413085936, 0.29922509765625, 0.2992752685546875, 0.29927935791015625, 0.29925274658203127, 0.2991441955566406, 0.29928652954101564, 0.29926705932617187, 0.6425958251953126, 0.29906124877929685, 0.299082763671875, 0.2990888977050781, 0.29896499633789064, 0.29904794311523436, 0.2992179260253906, 0.29917388916015625, 0.29889739990234376, 0.29929470825195315, 0.2990868530273437, 0.2990663757324219, 0.29902029418945314, 0.2996674499511719, 0.29909707641601563, 0.29932852172851565, 0.29918206787109375, 0.2994841613769531, 0.29911141967773436, 0.2992414855957031, 0.29906942749023435, 0.29909912109375, 0.29933978271484374, 0.29935205078125, 0.29904794311523436, 0.2990469055175781, 0.29906228637695315, 0.299072509765625, 0.2991544189453125, 0.2992015380859375, 0.29913394165039064, 0.2991595458984375, 0.29914727783203127, 0.29922509765625, 0.29901516723632815, 0.29917388916015625, 0.2992404479980469, 0.2990458984375, 0.2991011962890625, 0.29917593383789065, 0.29919540405273437, 0.2992261047363281, 0.29925274658203127, 0.29912677001953125, 0.2992271423339844, 0.2991493225097656, 0.29914111328125, 0.2990162048339844, 0.2993592224121094, 0.29924658203125, 0.29923635864257814, 0.2993960876464844, 0.29932235717773437, 0.29934796142578124, 0.2991697998046875, 0.29917593383789065, 0.2991964111328125, 0.29927117919921875, 0.2991544189453125, 0.29914727783203127, 0.29923941040039065, 0.29920050048828123, 0.2991964111328125, 0.6428948364257813, 0.2989619140625, 0.2990469055175781, 0.29888204956054687, 0.2990643310546875, 0.29899264526367186, 0.29902847290039064, 0.2990663757324219, 0.29907659912109374, 0.2990715026855469, 0.2991298522949219, 0.299156494140625, 0.29918923950195314, 0.2992066650390625, 0.29902542114257813, 0.2990448608398438, 0.2990888977050781, 0.2992261047363281, 0.29902542114257813, 0.29928652954101564, 0.299156494140625, 0.29922918701171874, 0.2991964111328125, 0.2992046203613281, 0.29915545654296877, 0.2990366821289063, 0.29951589965820313, 0.29911654663085935, 0.29921893310546877, 0.2991401062011719, 0.29907659912109374, 0.299171875, 0.29901513671875, 0.2992384033203125, 0.29911859130859375, 0.29911654663085935, 0.29927423095703126, 0.29912576293945314, 0.29905307006835935, 0.2991441955566406, 0.29913394165039064, 0.2993008728027344, 0.29917388916015625, 0.2992752685546875, 0.29923019409179685, 0.299146240234375, 0.2993377380371094, 0.2991933288574219, 0.2992384033203125, 0.29916775512695315, 0.29917388916015625, 0.299251708984375, 0.29923736572265625, 0.2992568359375, 0.2992281494140625, 0.2996858825683594, 0.29929165649414063, 0.29924249267578124, 0.29922406005859375, 0.29927423095703126, 0.29921893310546877, 0.299298828125, 0.2992353210449219, 0.6428385009765625, 0.2989793395996094, 0.29916671752929686, 0.29902542114257813, 0.2990192565917969, 0.2989906005859375, 0.29903564453125, 0.2990929870605469, 0.2991401062011719, 0.299040771484375, 0.29905307006835935, 0.2991708068847656, 0.2992015380859375, 0.29914727783203127, 0.2991994934082031, 0.299114501953125, 0.2990745544433594, 0.299109375, 0.2990602111816406, 0.2992599182128906, 0.29930804443359377, 0.2992455749511719, 0.2991718444824219, 0.2991984558105469, 0.29924249267578124, 0.299251708984375, 0.2989793395996094, 0.2992486267089844, 0.299177978515625, 0.2993100891113281, 0.2992537536621094, 0.29917593383789065, 0.29913702392578123, 0.2991974487304688, 0.29913189697265624, 0.29927423095703126, 0.29912371826171874, 0.2992557983398437, 0.2992568359375, 0.29929470825195315, 0.2991349792480469, 0.2990858154296875, 0.29928961181640623, 0.299208740234375, 0.29920867919921873, 0.2991718444824219, 0.29924453735351564, 0.2993510437011719, 0.2992557983398437, 0.2990797119140625, 0.2992424621582031, 0.2993326110839844, 0.29956402587890624, 0.2991626281738281, 0.29923431396484373, 0.2992117614746094, 0.29960498046875, 0.2993326110839844, 0.2993459167480469, 0.29935000610351564, 0.29927935791015625, 0.29934182739257814, 0.2992332763671875, 0.6429163818359375, 0.2990960693359375, 0.29910015869140627, 0.2990520324707031, 0.299109375, 0.2990458984375, 0.299030517578125, 0.2991022033691406, 0.29903564453125, 0.29908377075195314, 0.2990663757324219, 0.299156494140625, 0.29912677001953125, 0.29916363525390627, 0.2989916076660156, 0.2990663757324219, 0.29891278076171873, 0.29893939208984377, 0.2990469055175781, 0.29912576293945314, 0.2989956970214844, 0.298998779296875, 0.2990643310546875, 0.29916159057617187, 0.2990715026855469, 0.2994575500488281, 0.29900698852539065, 0.2995394592285156, 0.29913394165039064, 0.29909912109375, 0.29915341186523436, 0.29929165649414063, 0.29929983520507814, 0.2992281494140625, 0.2991523742675781, 0.2993172607421875, 0.29921484375, 0.29922406005859375, 0.29923431396484373, 0.2991708068847656, 0.29913906860351563, 0.29933465576171875, 0.2991575012207031, 0.299325439453125, 0.2992353210449219, 0.2993029174804688, 0.299177978515625, 0.29916876220703126, 0.2991513671875, 0.29921383666992185, 0.29926400756835936, 0.29923431396484373, 0.2992056274414063, 0.29919232177734373, 0.2990888977050781, 0.29924453735351564, 0.2991011962890625, 0.2991431579589844, 0.29923635864257814, 0.29907354736328123, 0.2992599182128906, 0.29916058349609376, 0.2992455749511719, 0.6419323120117187, 0.29938381958007815, 0.2990120849609375, 0.29897012329101563, 0.29900799560546876, 0.2991482849121094, 0.2989434814453125, 0.2991216735839844, 0.29890048217773435, 0.29918310546875, 0.29889022827148437, 0.2989854736328125, 0.29906228637695315, 0.29916876220703126, 0.2990223388671875, 0.29912371826171874, 0.2991022033691406, 0.29912677001953125, 0.29914520263671873, 0.29911962890625, 0.2989465637207031, 0.2991380615234375, 0.29913906860351563, 0.29906228637695315, 0.2993336181640625, 0.2991124572753906, 0.2991544189453125, 0.29918923950195314, 0.2996398010253906, 0.29912472534179685, 0.29916058349609376, 0.2991278076171875, 0.2992281494140625, 0.2991022033691406, 0.29903359985351563, 0.2992046203613281, 0.29924249267578124, 0.2990162048339844, 0.2991595458984375, 0.29906842041015624, 0.2990438537597656, 0.29918618774414063, 0.29923123168945315, 0.2992046203613281, 0.29913394165039064, 0.2991298522949219, 0.29950872802734374, 0.2992281494140625, 0.299219970703125, 0.2991749267578125, 0.2992066650390625, 0.29915341186523436, 0.29912677001953125, 0.29921484375, 0.2991482849121094, 0.29915545654296877, 0.29924966430664063, 0.29932748413085936, 0.2993008728027344, 0.2992179260253906, 0.2992629699707031, 0.2992414855957031, 0.299283447265625, 0.6432184448242187, 0.29901312255859375, 0.2996264953613281, 0.29940225219726563, 0.2994810791015625, 0.29907354736328123, 0.2989823913574219, 0.299082763671875, 0.2989260864257812, 0.29896499633789064, 0.2991544189453125, 0.2990653381347656, 0.29907659912109374, 0.29894247436523436, 0.2990100402832031, 0.2991728515625, 0.2991933288574219, 0.2991994934082031, 0.29903973388671873, 0.29916775512695315, 0.2990888977050781, 0.298977294921875, 0.2990520324707031, 0.29907559204101564, 0.2992732238769531, 0.2992599182128906, 0.29906842041015624, 0.2990243835449219, 0.2990950317382812, 0.29923019409179685, 0.2992824401855469, 0.29916876220703126, 0.2991974487304688, 0.29922305297851565, 0.299146240234375, 0.29922918701171874, 0.2990796813964844, 0.2992015380859375, 0.29915853881835935, 0.29912677001953125, 0.29899365234375, 0.2992353210449219, 0.29910324096679686, 0.2992015380859375, 0.2995189819335938, 0.29924966430664063, 0.2991206359863281, 0.29929779052734373, 0.2990796813964844, 0.29906329345703125, 0.29916159057617187, 0.2991278076171875, 0.29920870971679686, 0.2991708068847656, 0.29911859130859375, 0.29946676635742187, 0.29919232177734373, 0.29923941040039065, 0.2992332763671875, 0.2991329345703125, 0.29944729614257815, 0.2992957458496094, 0.2993489990234375]",tokens/s,3.2885173191969312,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,MB,1841.31584,7019.692032,0.0,6371.147776,6308.565504,s,10,1.3136222534179687,0.1313622253417969,8.207828410841184e-05,0.13137908935546874,0.13145328979492185,0.13146728363037108,0.13147847869873047,"[0.13134970092773438, 0.13143008422851563, 0.13122621154785155, 0.13139634704589845, 0.1314812774658203, 0.13138873291015624, 0.13136944580078125, 0.13145018005371092, 0.13129183959960938, 0.13123843383789063]",tokens/s,1948.8098601702493,kWh,1.5535538484938756e-06,8.512708269886807e-07,6.390939456181827e-06,8.795764131664383e-06,tokens/kWh,29104918.704949204,MB,1876.877312,7021.789184,0.0,6373.244928,6309.223424,s,10,20.129480834960937,2.0129480834960938,0.0053349894763728084,2.010189880371094,2.021866662597656,2.0219719665527345,2.022056209716797,"[2.00925146484375, 2.010943115234375, 2.0094366455078125, 2.015591064453125, 2.0090806884765624, 2.0220772705078125, 2.0161739501953124, 2.02184326171875, 2.008358154296875, 2.0067252197265626]",tokens/s,31.297379458779396,kWh,2.37839480224785e-05,1.3034254062747913e-05,7.324695758741844e-05,0.00011006515967264485,tokens/kWh,572388.2124677257,,s,630,20.12762109184265,0.03194860490768675,0.0005119588656906425,0.03182489585876465,0.032560229873657226,0.032789248085021976,0.03360792469024658,"[0.032835582733154296, 0.032247806549072264, 0.03287142562866211, 0.031735807418823245, 0.032048126220703126, 0.03200307083129883, 0.0317071361541748, 0.031715328216552735, 0.031631359100341795, 0.03168767929077149, 0.031703039169311525, 0.031716352462768556, 0.031676416397094724, 0.03159552001953125, 0.032438270568847655, 0.03202150344848633, 0.031801343917846676, 0.0319866886138916, 0.03189452743530274, 0.03181875228881836, 0.03198361587524414, 0.03158220863342285, 0.031693824768066405, 0.031527936935424807, 0.031541248321533204, 0.03145011138916016, 0.0314839038848877, 0.03172147178649903, 0.03158835220336914, 0.03151872062683105, 0.03161907196044922, 0.03175833511352539, 0.0320819206237793, 0.032612350463867186, 0.032048126220703126, 0.03171737670898438, 0.03256729507446289, 0.03242598342895508, 0.03208294296264649, 0.031955968856811526, 0.03159145545959473, 0.03161084747314453, 0.03201945495605469, 0.03234304046630859, 0.03212492752075195, 0.03265433502197266, 0.032314369201660156, 0.031710208892822264, 0.03170918464660644, 0.03192934417724609, 0.03191910362243652, 0.03162112045288086, 0.0318023681640625, 0.03160678482055664, 0.0315863037109375, 0.031940607070922854, 0.03176959991455078, 0.032105472564697264, 0.03149007987976074, 0.03188732719421387, 0.031904767990112305, 0.03156377601623535, 0.031765504837036135, 0.03244543838500977, 0.032075775146484374, 0.03172147178649903, 0.03167027282714844, 0.03209523010253906, 0.03174297523498535, 0.031899744033813474, 0.031320991516113283, 0.03136716842651367, 0.03262464141845703, 0.03165593528747559, 0.0317071361541748, 0.031888383865356446, 0.03188121604919433, 0.03175628852844238, 0.0317890567779541, 0.03186175918579102, 0.03174604797363281, 0.031591424942016604, 0.03182489585876465, 0.03162112045288086, 0.03198975944519043, 0.031817728042602536, 0.031703039169311525, 0.03167231941223145, 0.031710208892822264, 0.03172454452514648, 0.03140505599975586, 0.03215769577026367, 0.032246784210205076, 0.032140289306640625, 0.031893503189086916, 0.03225804901123047, 0.0315361270904541, 0.03162419128417969, 0.03148902320861816, 0.03154022407531738, 0.031474687576293944, 0.031884288787841795, 0.03211775970458984, 0.031905792236328126, 0.03215871810913086, 0.032710655212402344, 0.032261119842529294, 0.03192934417724609, 0.03269836807250977, 0.03220479965209961, 0.03235123062133789, 0.0317071361541748, 0.03219968032836914, 0.032456703186035156, 0.03214745712280274, 0.03155046463012695, 0.0322426872253418, 0.03194777679443359, 0.032123905181884765, 0.03163545608520508, 0.03211673736572265, 0.03156889533996582, 0.031715328216552735, 0.03256115341186523, 0.03184639930725098, 0.032039936065673826, 0.032023551940917966, 0.031513599395751955, 0.031663103103637694, 0.03190784072875977, 0.032008190155029294, 0.03155353546142578, 0.03155353546142578, 0.0323870735168457, 0.03183718490600586, 0.03174502372741699, 0.031714303970336914, 0.03208396911621094, 0.031971328735351565, 0.031886335372924804, 0.03180646324157715, 0.03194879913330078, 0.0319498233795166, 0.03151155281066895, 0.031459327697753905, 0.031888383865356446, 0.03176345634460449, 0.03135078430175781, 0.031319040298461914, 0.031285247802734374, 0.03205222320556641, 0.031492095947265625, 0.03139891242980957, 0.032366592407226565, 0.034411518096923825, 0.032696319580078126, 0.03205734252929687, 0.03168563270568848, 0.03177267265319824, 0.03201126480102539, 0.03159552001953125, 0.03215769577026367, 0.03200307083129883, 0.03202969741821289, 0.03357593536376953, 0.03236454391479492, 0.031221759796142577, 0.03128422355651855, 0.03242086410522461, 0.032007167816162106, 0.03153305625915527, 0.031938560485839845, 0.03155353546142578, 0.03246182250976563, 0.03175833511352539, 0.031270912170410156, 0.03136409568786621, 0.03181875228881836, 0.031853567123413085, 0.031666175842285156, 0.031661056518554685, 0.0323768310546875, 0.03203276824951172, 0.03223961639404297, 0.03144601631164551, 0.03156582450866699, 0.03191398429870605, 0.032228351593017575, 0.03183103942871094, 0.032024574279785153, 0.03169177627563476, 0.031749120712280275, 0.03199180793762207, 0.031286272048950195, 0.03198873519897461, 0.03142860794067383, 0.03176345634460449, 0.03199283218383789, 0.03147980880737305, 0.031732736587524416, 0.03182489585876465, 0.03156991958618164, 0.03300556945800781, 0.03310182571411133, 0.031784959793090824, 0.03170508766174317, 0.031987712860107424, 0.03322675323486328, 0.03275468826293945, 0.032336894989013674, 0.032107521057128906, 0.031834112167358396, 0.03146854400634766, 0.031704063415527346, 0.03171641540527344, 0.03192518424987793, 0.031732736587524416, 0.03162112045288086, 0.03163750457763672, 0.03158527946472168, 0.03215359878540039, 0.03184332847595215, 0.03181977653503418, 0.03205120086669922, 0.03202969741821289, 0.0317071361541748, 0.031752191543579104, 0.03188019180297851, 0.03147980880737305, 0.031920127868652344, 0.03452108764648437, 0.03348889541625977, 0.03259904098510742, 0.03179417610168457, 0.031882240295410154, 0.031922176361083986, 0.03182592010498047, 0.03199897575378418, 0.03163443183898926, 0.03267891311645508, 0.033620990753173825, 0.031954944610595705, 0.03178598403930664, 0.031866880416870115, 0.03187507247924805, 0.031507455825805664, 0.03163545608520508, 0.031821823120117186, 0.03186483192443847, 0.03138150405883789, 0.0314654712677002, 0.03188019180297851, 0.03211673736572265, 0.03171737670898438, 0.0317890567779541, 0.031645696640014646, 0.031833087921142575, 0.031699968338012696, 0.03196108818054199, 0.032791553497314455, 0.032094207763671875, 0.03184537506103516, 0.031764480590820314, 0.03190169525146484, 0.0317255687713623, 0.03156172752380371, 0.03157708740234375, 0.03193036842346191, 0.031492095947265625, 0.03119923210144043, 0.03146854400634766, 0.03142451286315918, 0.0318023681640625, 0.03174604797363281, 0.031848447799682614, 0.03177676773071289, 0.031937536239624024, 0.03165286445617676, 0.031735807418823245, 0.031939584732055666, 0.03157094383239746, 0.03177984046936035, 0.03196928024291992, 0.03256320190429687, 0.03188531112670898, 0.031527936935424807, 0.03173990440368652, 0.03189043235778809, 0.03172249603271484, 0.031834112167358396, 0.03162214469909668, 0.03196723175048828, 0.03417292785644531, 0.03166720008850098, 0.03167231941223145, 0.03160780715942383, 0.03179520034790039, 0.03141836738586426, 0.031853567123413085, 0.03176652717590332, 0.031631359100341795, 0.032317440032958986, 0.03249663925170899, 0.03165184020996094, 0.03153919982910156, 0.03162623977661133, 0.03171225547790527, 0.03150233650207519, 0.032121856689453124, 0.03213926315307617, 0.032072704315185545, 0.0317573127746582, 0.03253247833251953, 0.03319398498535156, 0.03258367919921875, 0.03144704055786133, 0.0317706241607666, 0.03267583847045898, 0.03246182250976563, 0.03171225547790527, 0.032535552978515625, 0.032361473083496094, 0.03232563018798828, 0.032007167816162106, 0.03220889663696289, 0.03196313667297363, 0.032985088348388675, 0.032677886962890625, 0.03256012725830078, 0.03297689437866211, 0.032645118713378905, 0.031676416397094724, 0.03193548774719238, 0.03184127998352051, 0.03191193580627441, 0.031458303451538085, 0.03157401657104492, 0.031801343917846676, 0.03205734252929687, 0.03373056030273437, 0.03278643035888672, 0.03189043235778809, 0.03233894348144531, 0.0319180793762207, 0.031681535720825195, 0.03245260620117187, 0.0316753921508789, 0.03187302398681641, 0.032780288696289066, 0.031513599395751955, 0.03162112045288086, 0.03304755020141602, 0.03238809585571289, 0.031692800521850584, 0.03216998291015625, 0.03150028800964356, 0.03175935935974121, 0.03243212890625, 0.032718849182128903, 0.03278438568115234, 0.031783935546875, 0.03173785591125488, 0.031492095947265625, 0.03213824081420898, 0.03180851173400879, 0.03191193580627441, 0.031677440643310545, 0.03160985565185547, 0.03246899032592773, 0.0326297607421875, 0.032010238647460935, 0.03197337532043457, 0.03179212760925293, 0.031615999221801756, 0.031647743225097655, 0.03168358421325684, 0.03179110336303711, 0.03180031967163086, 0.032023551940917966, 0.03160678482055664, 0.03207167816162109, 0.03215871810913086, 0.03195699119567871, 0.03171225547790527, 0.03186073684692383, 0.03192831993103027, 0.031936511993408204, 0.03284787368774414, 0.03198054313659668, 0.032486400604248046, 0.03151667213439941, 0.03174195289611816, 0.03189760017395019, 0.03186483192443847, 0.03183206367492676, 0.03162623977661133, 0.03151667213439941, 0.032173057556152344, 0.032089088439941404, 0.03212083053588867, 0.03143168067932129, 0.031833087921142575, 0.03167334365844727, 0.031514623641967776, 0.03158527946472168, 0.03172249603271484, 0.03276902389526367, 0.032059391021728514, 0.03216896057128906, 0.03209011077880859, 0.031888383865356446, 0.03193548774719238, 0.03291545486450195, 0.03238092803955078, 0.03133132743835449, 0.03202252960205078, 0.03280588912963867, 0.03194675254821777, 0.03263385772705078, 0.03207884979248047, 0.03149516868591309, 0.032949249267578126, 0.03271680068969727, 0.03142758369445801, 0.031541248321533204, 0.03197644805908203, 0.03180646324157715, 0.032476158142089845, 0.03224576187133789, 0.03215155029296875, 0.032892929077148435, 0.0326297607421875, 0.031719423294067385, 0.03176959991455078, 0.0317573127746582, 0.031643648147583005, 0.031780864715576174, 0.03181875228881836, 0.0323164176940918, 0.03150028800964356, 0.03164057540893555, 0.0383631362915039, 0.03297792053222656, 0.031339519500732424, 0.03165184020996094, 0.03277926254272461, 0.033808383941650394, 0.03274956893920898, 0.03171123123168945, 0.0315043830871582, 0.031768575668334964, 0.03192831993103027, 0.031849472045898435, 0.031575040817260744, 0.0319815673828125, 0.03212492752075195, 0.03181977653503418, 0.03348582458496094, 0.033413120269775394, 0.03183103942871094, 0.03182489585876465, 0.031545343399047854, 0.03164262390136719, 0.03183616065979004, 0.03193139266967773, 0.03159654426574707, 0.03155763244628906, 0.031665151596069335, 0.03171225547790527, 0.031887359619140625, 0.03197235107421875, 0.031719423294067385, 0.03239014434814453, 0.03244236755371094, 0.03253350448608398, 0.03158937644958496, 0.03266457748413086, 0.031704063415527346, 0.031718399047851564, 0.031921152114868165, 0.032486400604248046, 0.031503360748291014, 0.0315996150970459, 0.03199180793762207, 0.03172863960266113, 0.03175935935974121, 0.031660032272338864, 0.03194470405578613, 0.03156480026245117, 0.03169484710693359, 0.03182284736633301, 0.03192831993103027, 0.03249049758911133, 0.032723968505859374, 0.031821823120117186, 0.03191500854492187, 0.03273625564575195, 0.0316753921508789, 0.03209523010253906, 0.03168563270568848, 0.031529983520507815, 0.03153715133666992, 0.03194777679443359, 0.03129958343505859, 0.032366592407226565, 0.03223859024047852, 0.03224371337890625, 0.03175424003601074, 0.03181670379638672, 0.03177984046936035, 0.032710655212402344, 0.03184435272216797, 0.031498239517211916, 0.031730688095092774, 0.03198361587524414, 0.03153305625915527, 0.031904767990112305, 0.03157606315612793, 0.03177574348449707, 0.031893503189086916, 0.03209625625610352, 0.031903743743896484, 0.03306598281860352, 0.0328642578125, 0.031765504837036135, 0.03172454452514648, 0.032024574279785153, 0.031677440643310545, 0.031768575668334964, 0.03173785591125488, 0.031749120712280275, 0.03199078369140625, 0.03295948791503906, 0.03194777679443359, 0.03204608154296875, 0.031632383346557616, 0.031736831665039066, 0.03165593528747559, 0.031664127349853514, 0.031953920364379884, 0.03168358421325684, 0.031341567993164066, 0.03138047981262207, 0.03158527946472168, 0.031903743743896484, 0.03159040069580078, 0.03164057540893555, 0.0318791675567627, 0.03174195289611816, 0.03151667213439941, 0.03172249603271484, 0.03188121604919433, 0.031923200607299806, 0.031681535720825195, 0.03172659111022949, 0.03191910362243652, 0.03199180793762207, 0.03172659111022949, 0.03155046463012695, 0.0322242546081543, 0.03161702346801758, 0.031731712341308595, 0.03172761535644531, 0.031835136413574217, 0.03167231941223145, 0.03191296005249023, 0.032449535369873043, 0.03158323287963867, 0.03184025573730469, 0.03178700828552246, 0.032363521575927735, 0.03171123123168945, 0.03178291130065918, 0.03200614547729492, 0.031734783172607424, 0.03194572830200195, 0.03156070327758789, 0.03180031967163086, 0.031902719497680664, 0.032396289825439455, 0.03312844848632813, 0.03212799835205078, 0.031942655563354495, 0.031734783172607424, 0.0317388801574707, 0.031953920364379884, 0.03181670379638672, 0.03162521553039551, 0.03159552001953125, 0.03161702346801758, 0.03183923149108887, 0.03153919982910156, 0.031699968338012696, 0.03181875228881836, 0.032263168334960936, 0.03194367980957031, 0.03167231941223145, 0.03165593528747559, 0.03169075202941894, 0.03156480026245117, 0.031486976623535154, 0.03146342468261719, 0.03194470405578613, 0.0316866569519043, 0.03237580871582031, 0.032873470306396486, 0.031850496292114255, 0.03178700828552246, 0.03169587135314941, 0.031514623641967776, 0.031649791717529296, 0.03252326583862305, 0.031768575668334964, 0.03168563270568848, 0.031681535720825195, 0.03153715133666992, 0.03201228713989258, 0.03159040069580078, 0.031545343399047854, 0.03174092864990234, 0.0317071361541748, 0.03141222381591797, 0.031643648147583005, 0.03188019180297851, 0.031703039169311525, 0.03181875228881836, 0.03216896057128906, 0.032118785858154295, 0.032143360137939454, 0.03214438247680664]",tokens/s,31.300271260339215,,,,,,gemma,MB,4748.365824,6422.003712,0.0,5792.333824,5786.358272,s,1,11.0886728515625,11.0886728515625,0.0,11.0886728515625,11.0886728515625,11.0886728515625,11.0886728515625,[11.0886728515625],,kWh,5.021991096527643e-05,2.750872009994756e-05,7.062394538799588e-05,0.00014835257645321988, @@ -9968,48 +9806,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-eager,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,,,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,3918.962688,5287.44448,0.0,4640.997376,4330.799616,s,10,3.2259157409667965,0.32259157409667966,0.0005529953640440514,0.3225971221923828,0.3233306610107422,0.32335841522216796,0.3233806185913086,"[0.32232916259765626, 0.32165042114257814, 0.3231456298828125, 0.32269686889648436, 0.3224973754882812, 0.3220177917480469, 0.32332449340820313, 0.32209555053710937, 0.3233861694335938, 0.3227722778320313]",tokens/s,793.5731139812028,kWh,3.802365012880829e-06,2.0835388033233454e-06,1.894839189480954e-05,2.4834295711013717e-05,tokens/kWh,10308325.348903172,MB,3918.962688,5287.44448,0.0,4640.997376,4515.66848,s,10,175.46737109375,17.546737109375,0.001177635321808986,17.546580078125,17.548339062500002,17.548616796875,17.548838984375,"[17.54512890625, 17.5464765625, 17.54570703125, 17.54668359375, 17.54525, 17.54827734375, 17.54630078125, 17.547208984375, 17.54889453125, 17.547443359375]",tokens/s,3.590411117878998,kWh,0.00020716305749490858,0.000113541591573937,0.0010045058435346013,0.0013252104926034468,tokens/kWh,47539.61755632732,,s,629,178.09222891235356,0.28313549906574487,0.03836657912459716,0.27851162719726563,0.27869696044921877,0.2787557373046875,0.6009740600585938,"[0.2784675903320312, 0.2784378967285156, 0.2781951904296875, 0.2783365173339844, 0.27832830810546877, 0.2783682556152344, 0.2783068237304688, 0.27829656982421874, 0.27835595703125, 0.2783907775878906, 0.27824639892578124, 0.2782801818847656, 0.27823822021484373, 0.2784276428222656, 0.27843072509765626, 0.27827508544921875, 0.27839898681640624, 0.2785382385253906, 0.2784798583984375, 0.2784778137207031, 0.2784051208496094, 0.2784163818359375, 0.2784532470703125, 0.2783907775878906, 0.27848089599609377, 0.27852593994140623, 0.27842868041992186, 0.2785771484375, 0.27849215698242186, 0.27850341796875, 0.27848907470703127, 0.27865191650390625, 0.278687744140625, 0.27855462646484375, 0.27841741943359377, 0.278445068359375, 0.2784675903320312, 0.27849932861328125, 0.2784747619628906, 0.2785249328613281, 0.27833856201171875, 0.27862322998046873, 0.2785730590820312, 0.27850341796875, 0.27848089599609377, 0.2786488342285156, 0.27850137329101565, 0.27865396118164065, 0.2785515441894531, 0.27858328247070313, 0.27858123779296873, 0.2786918334960938, 0.2785577087402344, 0.2786856994628906, 0.27846246337890623, 0.278719482421875, 0.2786211853027344, 0.2786273193359375, 0.2784972839355469, 0.27865191650390625, 0.27861709594726564, 0.27877069091796874, 0.6005718994140625, 0.2785904541015625, 0.27841741943359377, 0.278265869140625, 0.27840103149414064, 0.278761474609375, 0.27836517333984373, 0.278371337890625, 0.27834674072265625, 0.27851263427734374, 0.2784460754394531, 0.27822695922851565, 0.27832321166992186, 0.27824026489257814, 0.27833956909179686, 0.27833856201171875, 0.2783477783203125, 0.2783631286621094, 0.2784829406738281, 0.27833139038085936, 0.278445068359375, 0.27841741943359377, 0.27845632934570314, 0.278518798828125, 0.2784481201171875, 0.27845428466796873, 0.27843994140625, 0.278403076171875, 0.2785658874511719, 0.27850546264648435, 0.27858123779296873, 0.27847064208984373, 0.278645751953125, 0.2785935363769531, 0.27854437255859377, 0.278513671875, 0.2785935363769531, 0.27843890380859376, 0.27848703002929687, 0.27840716552734374, 0.278572021484375, 0.27874200439453123, 0.27856179809570314, 0.27862014770507815, 0.27857510375976563, 0.27847271728515627, 0.2787010498046875, 0.2785290222167969, 0.27869287109375, 0.2786938781738281, 0.27897445678710936, 0.2785535888671875, 0.2785587158203125, 0.2785372009277344, 0.27866827392578125, 0.27850750732421875, 0.27867852783203123, 0.2785105895996094, 0.27856179809570314, 0.27850546264648435, 0.2785904541015625, 0.27853619384765627, 0.27867852783203123, 0.602208251953125, 0.278476806640625, 0.27831808471679687, 0.2782474365234375, 0.27843173217773437, 0.2783057861328125, 0.27829556274414063, 0.2782945251464844, 0.27835186767578124, 0.278297607421875, 0.2783129577636719, 0.2783918151855469, 0.2784460754394531, 0.278329345703125, 0.27832321166992186, 0.27822283935546877, 0.2788935546875, 0.2784245910644531, 0.27843585205078125, 0.27840921020507814, 0.2784132995605469, 0.27833242797851565, 0.278619140625, 0.2783498229980469, 0.27841537475585937, 0.278392822265625, 0.2784245910644531, 0.27871743774414065, 0.27856997680664064, 0.27848193359375, 0.27866009521484375, 0.2784880676269531, 0.2784696350097656, 0.27849932861328125, 0.2784460754394531, 0.27853515625, 0.2785382385253906, 0.2783631286621094, 0.2785986633300781, 0.2784296875, 0.2785321044921875, 0.27840615844726563, 0.2784778137207031, 0.2785730590820312, 0.2785024108886719, 0.2784245910644531, 0.27858740234375, 0.2785238952636719, 0.27861505126953123, 0.27867034912109373, 0.2786283569335937, 0.2786007080078125, 0.2790328369140625, 0.27862527465820314, 0.2787256469726562, 0.2784747619628906, 0.27861196899414065, 0.278619140625, 0.278724609375, 0.2785935363769531, 0.27863348388671877, 0.278540283203125, 0.27875839233398436, 0.6016409301757812, 0.27838360595703127, 0.2785607604980469, 0.2782525329589844, 0.27845733642578124, 0.2782443542480469, 0.2783703002929688, 0.2783139953613281, 0.27840408325195315, 0.2783375244140625, 0.27838772583007815, 0.27835903930664063, 0.278392822265625, 0.27826995849609376, 0.27838055419921875, 0.2785269775390625, 0.2785331115722656, 0.27831808471679687, 0.27850958251953123, 0.2784471130371094, 0.27848089599609377, 0.2784481201171875, 0.2783477783203125, 0.2785208435058594, 0.27845016479492185, 0.27845016479492185, 0.27848602294921876, 0.27848397827148436, 0.27852801513671877, 0.2785433654785156, 0.27850750732421875, 0.27853927612304685, 0.27850650024414064, 0.2785740661621094, 0.2785577087402344, 0.27850546264648435, 0.2784829406738281, 0.27839385986328125, 0.27899493408203124, 0.27845120239257815, 0.2785689697265625, 0.27843585205078125, 0.27867544555664064, 0.2785535888671875, 0.2785884094238281, 0.2786611328125, 0.27860275268554685, 0.27862322998046873, 0.278677490234375, 0.2786437072753906, 0.2785607604980469, 0.2785710144042969, 0.27888128662109374, 0.27861093139648435, 0.2785730590820312, 0.278513671875, 0.2786631774902344, 0.27858740234375, 0.2785740661621094, 0.278498291015625, 0.27868057250976563, 0.2785986633300781, 0.27873382568359373, 0.6014218139648437, 0.27839385986328125, 0.2783211669921875, 0.278245361328125, 0.2784296875, 0.27821466064453126, 0.2783621215820313, 0.2782853088378906, 0.2784296875, 0.2782576599121094, 0.27834368896484374, 0.2782033996582031, 0.2784378967285156, 0.2782167053222656, 0.27833856201171875, 0.2784532470703125, 0.2783006591796875, 0.2784163818359375, 0.27851263427734374, 0.27829656982421874, 0.2784829406738281, 0.2782607421875, 0.27854745483398435, 0.2784122924804687, 0.2784747619628906, 0.27844403076171875, 0.27850546264648435, 0.27843994140625, 0.27860379028320315, 0.27854745483398435, 0.27849932861328125, 0.2784655456542969, 0.27863961791992187, 0.2785003662109375, 0.27850137329101565, 0.27850341796875, 0.27838156127929686, 0.2785167236328125, 0.2785791931152344, 0.2784675903320312, 0.27858944702148436, 0.2783692932128906, 0.2785310668945313, 0.2785157165527344, 0.2785218505859375, 0.27848907470703127, 0.27869696044921877, 0.2786590576171875, 0.27863348388671877, 0.2786990051269531, 0.278697998046875, 0.2785228881835938, 0.27875326538085937, 0.27864474487304686, 0.2785310668945313, 0.27853619384765627, 0.27875225830078126, 0.27861709594726564, 0.278640625, 0.27856280517578125, 0.2788239440917969, 0.2785249328613281, 0.2788229064941406, 0.6009180297851563, 0.27875738525390625, 0.27840408325195315, 0.2787317810058594, 0.27844198608398435, 0.2784952392578125, 0.2783703002929688, 0.2784204711914062, 0.27834878540039065, 0.2785208435058594, 0.27833139038085936, 0.2782863464355469, 0.27851980590820313, 0.27829144287109375, 0.2784184265136719, 0.27834060668945315, 0.27849114990234375, 0.27847885131835937, 0.2785331115722656, 0.27837542724609377, 0.2785218505859375, 0.27835595703125, 0.27865805053710935, 0.27854232788085936, 0.27851776123046873, 0.27856280517578125, 0.2784716796875, 0.2784163818359375, 0.2786263122558594, 0.27854232788085936, 0.27852801513671877, 0.2785372009277344, 0.2785607604980469, 0.2785638427734375, 0.27857818603515627, 0.27848193359375, 0.2784655456542969, 0.2784696350097656, 0.27858944702148436, 0.27878091430664065, 0.2787389526367188, 0.278572021484375, 0.2785587158203125, 0.2785300598144531, 0.278513671875, 0.2785515441894531, 0.27863961791992187, 0.2785525817871094, 0.2786160583496094, 0.27866622924804685, 0.27864678955078126, 0.2784737243652344, 0.27865191650390625, 0.2786570129394531, 0.2786816101074219, 0.2785464172363281, 0.2786365356445312, 0.2785863647460938, 0.2786990051269531, 0.27853619384765627, 0.2786836547851563, 0.27861505126953123, 0.27869696044921877, 0.60225537109375, 0.2784429931640625, 0.2783365173339844, 0.2782105712890625, 0.27842868041992186, 0.27831195068359377, 0.27833242797851565, 0.2782791748046875, 0.27833447265625, 0.27844403076171875, 0.2783211669921875, 0.27834674072265625, 0.278361083984375, 0.27834573364257814, 0.27851162719726563, 0.27837542724609377, 0.27849319458007815, 0.27838360595703127, 0.2784901123046875, 0.2784737243652344, 0.2785658874511719, 0.27838772583007815, 0.27850750732421875, 0.2784481201171875, 0.27856179809570314, 0.27863961791992187, 0.27845632934570314, 0.2784112548828125, 0.2784471130371094, 0.27855563354492185, 0.27854745483398435, 0.27850341796875, 0.2786283569335937, 0.27844403076171875, 0.27850341796875, 0.2784655456542969, 0.27848089599609377, 0.27845428466796873, 0.2786590576171875, 0.2785300598144531, 0.2785771484375, 0.2784051208496094, 0.2785986633300781, 0.2785863647460938, 0.278540283203125, 0.27851162719726563, 0.27865191650390625, 0.2786641845703125, 0.27860888671875, 0.2785986633300781, 0.2785607604980469, 0.27842355346679687, 0.2787317810058594, 0.27872869873046874, 0.2786344909667969, 0.278513671875, 0.2787471313476563, 0.27864984130859377, 0.2786918334960938, 0.27854437255859377, 0.278729736328125, 0.2786068420410156, 0.27870822143554685, 0.600995849609375, 0.2783549499511719, 0.27836007690429687, 0.278371337890625, 0.27842355346679687, 0.2783150024414062, 0.2783477783203125, 0.2783150024414062, 0.278302734375, 0.278265869140625, 0.27850650024414064, 0.27838772583007815, 0.27848397827148436, 0.27830374145507814, 0.2784696350097656, 0.2783365173339844, 0.278434814453125, 0.27846246337890623, 0.2786211853027344, 0.27852801513671877, 0.2786058349609375, 0.2782320556640625, 0.27850958251953123, 0.27844198608398435, 0.27843994140625, 0.27850445556640624, 0.27860479736328125, 0.2783682556152344, 0.2785986633300781, 0.27849215698242186, 0.278518798828125, 0.27844403076171875, 0.2786017150878906, 0.2785689697265625, 0.27850750732421875, 0.27848193359375, 0.278582275390625, 0.2784532470703125, 0.27861810302734374, 0.27844403076171875, 0.27855667114257815, 0.2784184265136719, 0.278572021484375, 0.2786222229003906, 0.2785904541015625, 0.2785658874511719, 0.27858944702148436, 0.27860479736328125, 0.2786621398925781, 0.2785904541015625, 0.2787461242675781, 0.2785249328613281, 0.278729736328125, 0.27870513916015627, 0.2786385803222656, 0.27873690795898437, 0.2787164306640625, 0.278656005859375, 0.2786437072753906, 0.27850546264648435, 0.27900827026367186, 0.2785638427734375, 0.278751220703125, 0.6023884887695312, 0.2784880676269531, 0.2782607421875, 0.27817062377929686, 0.2784378967285156, 0.27832421875, 0.27845733642578124, 0.27837850952148435, 0.27836517333984373, 0.2786263122558594, 0.2783846435546875, 0.2783068237304688, 0.2785535888671875, 0.27822796630859375, 0.278466552734375, 0.27838668823242185, 0.2784409484863281, 0.27841537475585937, 0.2784962463378906, 0.27847885131835937, 0.2784583740234375, 0.2783365173339844, 0.27845220947265625, 0.2784747619628906, 0.2787461242675781, 0.2785228881835938, 0.27894476318359374, 0.27845733642578124, 0.27854132080078126, 0.2785638427734375, 0.27855563354492185, 0.27856997680664064, 0.27877786254882814, 0.2786375732421875, 0.27846450805664064, 0.278371337890625, 0.27854849243164065, 0.27875839233398436, 0.27850650024414064, 0.2785464172363281, 0.27856793212890624, 0.27850341796875, 0.2785372009277344, 0.2785372009277344, 0.2785802307128906, 0.2786058349609375, 0.2786017150878906, 0.2786488342285156, 0.27872357177734375, 0.27849114990234375, 0.27882699584960935, 0.2785935363769531, 0.2786324462890625, 0.27864166259765627, 0.27864678955078126, 0.278508544921875, 0.2786897888183594, 0.27858123779296873, 0.2786375732421875, 0.27848193359375, 0.2791280517578125, 0.2786713562011719, 0.27919769287109375, 0.601754638671875, 0.27856484985351565, 0.27835699462890623, 0.2782443542480469, 0.27838772583007815, 0.27829861450195315, 0.2783529052734375, 0.27829861450195315, 0.27851263427734374, 0.27839898681640624, 0.27828839111328124, 0.27840615844726563, 0.2784962463378906, 0.2782300109863281, 0.27842868041992186, 0.2783846435546875, 0.27839486694335935, 0.27840716552734374, 0.27844198608398435, 0.2783999938964844, 0.2784880676269531, 0.27841537475585937, 0.27849114990234375, 0.278540283203125, 0.27850650024414064, 0.27849420166015626, 0.27862322998046873, 0.27851162719726563, 0.2787327880859375, 0.2785658874511719, 0.2785986633300781, 0.27849215698242186, 0.27883621215820314, 0.27854437255859377, 0.2786068420410156, 0.2784635009765625, 0.2785802307128906, 0.27857510375976563, 0.27845120239257815, 0.278466552734375, 0.27857818603515627, 0.27845938110351565, 0.27869491577148436, 0.27854132080078126, 0.2785986633300781, 0.2784737243652344, 0.278582275390625, 0.2786631774902344, 0.27878091430664065, 0.27861196899414065, 0.278540283203125, 0.27854949951171876, 0.27867239379882813, 0.2786641845703125, 0.27863961791992187, 0.27857614135742187, 0.27870925903320315, 0.27857614135742187, 0.2787891235351562, 0.2785321044921875, 0.278697998046875, 0.2785003662109375, 0.278793212890625]",tokens/s,3.531877858126851,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch @@ -10893,6 +10716,33 @@ ChildProcessError: Traceback (most recent call last): OSError: / does not appear to have a file named config.json. Checkout 'https://huggingface.co///tree/None' for available files. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-9b,google/recurrentgemma-9b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,recurrent_gemma,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,2120.617984,5971.116032,0.0,5324.668928,5080.87552,s,10,3.160778564453125,0.3160778564453125,0.00014221386743603723,0.31608947753906247,0.31621777343749996,0.316263916015625,0.316300830078125,"[0.31631005859375, 0.31595333862304686, 0.31619436645507815, 0.3159564208984375, 0.315891845703125, 0.3161977844238281, 0.31620751953125, 0.31612481689453126, 0.3158882751464844, 0.31605413818359374]",tokens/s,809.9270315201371,kWh,3.7333295784062813e-06,2.0456940575684257e-06,1.835831329775016e-05,2.4137336933724864e-05,tokens/kWh,10605975.327887764,MB,2120.617984,5971.116032,0.0,5324.668928,5290.117632,s,10,173.24443359375002,17.324443359375003,0.0011951264280211484,17.324284179687503,17.325962304687497,17.325986035156248,17.326005019531248,"[17.32349609375, 17.326009765625, 17.3251484375, 17.325763671875, 17.32453515625, 17.323236328125, 17.32595703125, 17.324033203125, 17.32393359375, 17.3223203125]",tokens/s,3.636480474041204,kWh,0.00020452947697291774,0.00011209910411696113,0.00099598438706465,0.0013126129681545287,tokens/kWh,47995.868948769414,,s,629,175.81415676879894,0.27951376274848777,0.03754764840079795,0.27497882080078123,0.27517562255859374,0.27527659912109376,0.5910434716796875,"[0.2751549377441406, 0.27478631591796876, 0.274740234375, 0.2748467102050781, 0.2747781066894531, 0.27483648681640627, 0.27480575561523435, 0.27484774780273435, 0.27487026977539064, 0.2749368286132812, 0.2748989562988281, 0.2748334045410156, 0.27480679321289064, 0.2747279357910156, 0.27483444213867186, 0.2748395385742187, 0.2749573059082031, 0.2748907470703125, 0.2748989562988281, 0.274777099609375, 0.2748989562988281, 0.27485696411132815, 0.2748805236816406, 0.2748784790039063, 0.27539968872070314, 0.2749306945800781, 0.2748108825683594, 0.27496652221679685, 0.2749921264648437, 0.2748805236816406, 0.2750218200683594, 0.2749276123046875, 0.2749573059082031, 0.27505767822265625, 0.2749849548339844, 0.2750382080078125, 0.2751651916503906, 0.27496038818359375, 0.2750003051757812, 0.27507302856445315, 0.2750699462890625, 0.27509246826171874, 0.275346435546875, 0.275018798828125, 0.2749961853027344, 0.2750627746582031, 0.2750822448730469, 0.2749767761230469, 0.2750279541015625, 0.2750679016113281, 0.2748856201171875, 0.27508120727539065, 0.27509246826171874, 0.27501669311523436, 0.27497882080078123, 0.2750330810546875, 0.27511911010742185, 0.27523687744140624, 0.27515802001953127, 0.275162109375, 0.27502590942382815, 0.2751109008789063, 0.5910446166992187, 0.2748846435546875, 0.27493679809570315, 0.27485592651367186, 0.27479757690429685, 0.2749266052246094, 0.2747801513671875, 0.2748385314941406, 0.27494503784179686, 0.2749010009765625, 0.27492044067382815, 0.27491122436523435, 0.27478842163085937, 0.2750320129394531, 0.27486822509765624, 0.2753914794921875, 0.2748610534667969, 0.274977783203125, 0.27492352294921873, 0.27486309814453125, 0.27502899169921874, 0.27495526123046876, 0.2748989562988281, 0.2748948364257813, 0.27489279174804687, 0.27484774780273435, 0.2748805236816406, 0.2749081726074219, 0.27503411865234373, 0.27519793701171874, 0.27506591796875, 0.27506375122070315, 0.27527474975585936, 0.2748713073730469, 0.2750074768066406, 0.274904052734375, 0.2754652099609375, 0.2749624328613281, 0.2749819030761719, 0.2750412902832031, 0.27497369384765624, 0.27500851440429686, 0.27519180297851564, 0.274872314453125, 0.275061767578125, 0.2750320739746094, 0.2752962646484375, 0.27497268676757813, 0.2749819030761719, 0.2750351257324219, 0.2755778503417969, 0.27515597534179687, 0.2751201171875, 0.27508837890625, 0.2750494689941406, 0.27514776611328123, 0.27520306396484373, 0.2751528930664062, 0.2750115966796875, 0.275156005859375, 0.27518563842773436, 0.2749910888671875, 0.27510885620117187, 0.5912586059570313, 0.27490304565429685, 0.27485491943359375, 0.2748385314941406, 0.2748313598632813, 0.27479244995117186, 0.27500851440429686, 0.27482110595703124, 0.2751180725097656, 0.27487640380859374, 0.2750003051757812, 0.27502081298828124, 0.2749696044921875, 0.2748538818359375, 0.2746982421875, 0.2748958740234375, 0.2749091796875, 0.2749286499023437, 0.275240966796875, 0.27487744140625, 0.2748395385742187, 0.2748037109375, 0.2750064697265625, 0.2748907470703125, 0.27497882080078123, 0.27496652221679685, 0.2749358215332031, 0.27493991088867187, 0.274845703125, 0.274914306640625, 0.2749286499023437, 0.27517645263671875, 0.2750771179199219, 0.2749706115722656, 0.27486822509765624, 0.275156982421875, 0.27497268676757813, 0.2749921264648437, 0.2750484619140625, 0.2749644775390625, 0.2750044250488281, 0.2749798278808594, 0.2752276611328125, 0.27546112060546873, 0.275125244140625, 0.2750402526855469, 0.27501773071289065, 0.2749358215332031, 0.2750412902832031, 0.2750894165039062, 0.2750074768066406, 0.2750545959472656, 0.2751150207519531, 0.2751928405761719, 0.27498086547851563, 0.27527783203125, 0.2749931640625, 0.2750330810546875, 0.27503411865234373, 0.2751528930664062, 0.27550515747070314, 0.2750269470214844, 0.2751150207519531, 0.59104052734375, 0.2749900817871094, 0.2748436584472656, 0.2749163513183594, 0.27488357543945313, 0.274850830078125, 0.27485592651367186, 0.2748692626953125, 0.27481497192382814, 0.2749962158203125, 0.2748784790039063, 0.27504229736328123, 0.274956298828125, 0.274934814453125, 0.2749541931152344, 0.2749941711425781, 0.27521844482421876, 0.2750484619140625, 0.27506381225585935, 0.27490509033203125, 0.2747699279785156, 0.2748641357421875, 0.27487850952148435, 0.27475555419921877, 0.27495013427734377, 0.2749266052246094, 0.2748743591308594, 0.2752542724609375, 0.2749706115722656, 0.27507611083984373, 0.2751109008789063, 0.2750330810546875, 0.2754150390625, 0.274977783203125, 0.27513345336914063, 0.27508428955078124, 0.27500238037109376, 0.27498699951171873, 0.275009521484375, 0.27486822509765624, 0.27494708251953126, 0.27500851440429686, 0.27512216186523436, 0.27486822509765624, 0.2749327392578125, 0.27506072998046877, 0.2749296569824219, 0.275009521484375, 0.27506072998046877, 0.2750382080078125, 0.2750146484375, 0.27508837890625, 0.2751744689941406, 0.2754303283691406, 0.2751958923339844, 0.2751508483886719, 0.2751016845703125, 0.27512115478515625, 0.27508633422851564, 0.275019775390625, 0.27512933349609375, 0.275093505859375, 0.275167236328125, 0.5909524536132813, 0.2748692626953125, 0.2747965393066406, 0.274798583984375, 0.274872314453125, 0.274951171875, 0.2747801513671875, 0.2747484130859375, 0.2748190612792969, 0.2747678833007812, 0.27482009887695313, 0.2748661804199219, 0.2749358215332031, 0.27501055908203126, 0.27486822509765624, 0.27493585205078125, 0.27487127685546875, 0.2749347839355469, 0.27503103637695314, 0.2748323974609375, 0.2747781066894531, 0.27493991088867187, 0.2749276123046875, 0.27532186889648436, 0.27497268676757813, 0.2752706604003906, 0.2750003051757812, 0.2747832336425781, 0.2749266052246094, 0.2750044250488281, 0.27501669311523436, 0.2750402526855469, 0.27488973999023436, 0.2749440002441406, 0.2749849548339844, 0.2750136413574219, 0.2751754150390625, 0.27496856689453125, 0.2750382080078125, 0.2750597229003906, 0.2750392456054688, 0.2752174072265625, 0.27517745971679686, 0.27495321655273436, 0.275156982421875, 0.27486004638671874, 0.2751426696777344, 0.2749839477539063, 0.27503717041015624, 0.2749931640625, 0.27507916259765625, 0.2750403442382813, 0.2750545043945313, 0.2751549377441406, 0.2750699462890625, 0.27504434204101563, 0.2751047668457031, 0.2751129455566406, 0.2751170654296875, 0.2751405944824219, 0.2752563171386719, 0.2750802001953125, 0.275093505859375, 0.591404052734375, 0.27478732299804687, 0.2747463684082031, 0.2747463684082031, 0.2748743591308594, 0.27486309814453125, 0.27487640380859374, 0.2747135925292969, 0.27524197387695315, 0.27495321655273436, 0.27483648681640627, 0.27523275756835935, 0.27500238037109376, 0.2752081909179688, 0.27478427124023436, 0.2749214782714844, 0.27479962158203125, 0.2749573059082031, 0.27495834350585935, 0.27488973999023436, 0.2750689392089844, 0.2748938293457031, 0.2749347839355469, 0.274798583984375, 0.2747658386230469, 0.2748211669921875, 0.2749531860351562, 0.2749931640625, 0.275009521484375, 0.27486822509765624, 0.2750822448730469, 0.27492556762695314, 0.27491839599609375, 0.27491738891601564, 0.27533108520507815, 0.27536587524414063, 0.27496652221679685, 0.27481808471679686, 0.27493991088867187, 0.2749531860351562, 0.27497573852539064, 0.274914306640625, 0.2750218200683594, 0.27485491943359375, 0.27489178466796876, 0.2750146484375, 0.2750074768066406, 0.2750136413574219, 0.2751447143554687, 0.2749767761230469, 0.2750484619140625, 0.275135498046875, 0.27490203857421874, 0.2749685363769531, 0.2750413513183594, 0.2751610107421875, 0.27513446044921874, 0.2751436767578125, 0.27492044067382815, 0.27500238037109376, 0.2750679016113281, 0.27492352294921873, 0.2751355285644531, 0.5913036499023437, 0.27488153076171873, 0.27485797119140626, 0.27473101806640626, 0.27497882080078123, 0.27478631591796876, 0.2751488037109375, 0.27473202514648437, 0.27486822509765624, 0.2748518371582031, 0.2749644775390625, 0.2750003051757812, 0.2749214782714844, 0.2749296569824219, 0.2748794860839844, 0.2749767761230469, 0.27479962158203125, 0.2754744262695312, 0.27506585693359376, 0.2748374938964844, 0.27503103637695314, 0.27487539672851563, 0.2749153137207031, 0.27516619873046877, 0.27500338745117187, 0.2749542541503906, 0.27485491943359375, 0.2749767761230469, 0.27487335205078123, 0.27509246826171874, 0.27496139526367186, 0.27499725341796877, 0.27492044067382815, 0.27487335205078123, 0.2748037109375, 0.2749992980957031, 0.274988037109375, 0.27499725341796877, 0.2749624328613281, 0.2749593505859375, 0.275557373046875, 0.2749491271972656, 0.27513650512695315, 0.27502899169921874, 0.2751129455566406, 0.2749962158203125, 0.27495526123046876, 0.2748590087890625, 0.2750382080078125, 0.2751129455566406, 0.27525323486328124, 0.2751539306640625, 0.27510784912109376, 0.2751170654296875, 0.275314697265625, 0.27513345336914063, 0.2750320739746094, 0.2750699462890625, 0.27560345458984375, 0.2751047668457031, 0.27525531005859377, 0.2750750732421875, 0.2749819030761719, 0.5911541748046875, 0.27480172729492186, 0.2748220825195313, 0.2748968811035156, 0.2747218017578125, 0.2748538818359375, 0.27491122436523435, 0.2748108825683594, 0.2747555847167969, 0.2747412414550781, 0.2749276123046875, 0.2749245300292969, 0.274904052734375, 0.2748374938964844, 0.27484979248046876, 0.27491326904296876, 0.2747781066894531, 0.27500338745117187, 0.27496551513671874, 0.274798583984375, 0.274956298828125, 0.27488973999023436, 0.27482827758789063, 0.2748784790039063, 0.2750003051757812, 0.27475250244140625, 0.2749429626464844, 0.27521331787109377, 0.2749491271972656, 0.2751150207519531, 0.2752850036621094, 0.2750074768066406, 0.27493374633789064, 0.2749347839355469, 0.27506072998046877, 0.2751426696777344, 0.2750392456054688, 0.2748118896484375, 0.27491943359375, 0.27496856689453125, 0.2750136413574219, 0.274808837890625, 0.275030029296875, 0.27494708251953126, 0.274977783203125, 0.27509451293945314, 0.27521331787109377, 0.27501568603515625, 0.27507098388671875, 0.27510272216796877, 0.2749358215332031, 0.27550515747070314, 0.2751405944824219, 0.27502490234375, 0.2750003051757812, 0.2750330810546875, 0.27517132568359376, 0.27509451293945314, 0.2750965881347656, 0.2751242370605469, 0.2751856689453125, 0.2751109008789063, 0.2755092468261719, 0.591119384765625, 0.2749716491699219, 0.2747607116699219, 0.27478118896484377, 0.27481497192382814, 0.2748661804199219, 0.27484466552734377, 0.27487539672851563, 0.27488973999023436, 0.2748272705078125, 0.27498291015625, 0.2748958740234375, 0.27494091796875, 0.27488153076171873, 0.2748334045410156, 0.2749521789550781, 0.27494503784179686, 0.27500750732421875, 0.2749541931152344, 0.27481396484375, 0.2748385314941406, 0.2752174072265625, 0.27478936767578127, 0.27481805419921873, 0.2750044250488281, 0.2748856201171875, 0.2749296569824219, 0.2748856201171875, 0.275061767578125, 0.27486209106445314, 0.2749542541503906, 0.274956298828125, 0.27497573852539064, 0.27501873779296876, 0.2750433349609375, 0.27517849731445315, 0.2750965881347656, 0.2750126037597656, 0.27508428955078124, 0.2750269470214844, 0.2750136413574219, 0.27496038818359375, 0.2749716491699219, 0.274956298828125, 0.27498904418945314, 0.2750955505371094, 0.2750392456054688, 0.27496652221679685, 0.2749573059082031, 0.27499520874023436, 0.27500851440429686, 0.275030029296875, 0.2750873718261719, 0.2749859924316406, 0.2750771179199219, 0.275009521484375, 0.2750412902832031, 0.27533721923828125, 0.2750392456054688, 0.27516415405273437, 0.27510067749023437, 0.27538534545898435, 0.2750955505371094, 0.5911121826171875, 0.27488461303710937, 0.2748856201171875, 0.2748938293457031, 0.27470950317382814, 0.27487850952148435, 0.2748907165527344, 0.274914306640625, 0.27480062866210936, 0.27482623291015623, 0.27487026977539064, 0.2751856689453125, 0.2748098449707031, 0.274740234375, 0.27479757690429685, 0.27485592651367186, 0.2749327392578125, 0.27489996337890626, 0.27496038818359375, 0.2748272705078125, 0.27478732299804687, 0.2748323974609375, 0.274845703125, 0.27469515991210935, 0.2749358215332031, 0.2748661804199219, 0.27492556762695314, 0.2748293151855469, 0.27499520874023436, 0.2749327392578125, 0.2750412902832031, 0.2750218200683594, 0.2752102355957031, 0.2749224853515625, 0.2749010009765625, 0.275061767578125, 0.27493991088867187, 0.27485592651367186, 0.2750115966796875, 0.27513037109375, 0.2749706115722656, 0.27488153076171873, 0.27488461303710937, 0.27494708251953126, 0.27503717041015624, 0.27507098388671875, 0.27512832641601564, 0.27502590942382815, 0.2750474243164063, 0.2748876647949219, 0.2750802001953125, 0.27499725341796877, 0.2751907958984375, 0.27516110229492186, 0.27500543212890627, 0.2750382385253906, 0.2752030334472656, 0.2751068115234375, 0.27504434204101563, 0.27506381225585935, 0.27498699951171873, 0.27508428955078124, 0.2751539306640625]",tokens/s,3.577641366088369,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -11113,48 +10963,33 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp0etlqmat/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-65b,huggyllama/llama-65b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 86.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -11255,48 +11090,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2_moe,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mixtral-8x7B-v0.1,mistralai/Mixtral-8x7B-v0.1,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,mixtral,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1323.073536,1313.34144,0.0,666.894336,583.411712,s,10,0.33887715530395507,0.033887715530395504,0.0016204514052615168,0.033726848602294926,0.03456971397399902,0.03648557682037353,0.03801826709747315,"[0.03840143966674805, 0.03366169738769531, 0.03284896087646484, 0.03403376007080078, 0.033791999816894534, 0.03261516952514648, 0.03414396667480469, 0.03399248123168945, 0.03299187088012695, 0.032395809173583985]",tokens/s,7554.359920496306,kWh,3.993419203180093e-07,2.1878426233662458e-07,1.2312050014795668e-06,1.8493311841342008e-06,tokens/kWh,138428423.31123683,MB,1323.401216,1313.34144,0.0,666.894336,644.0192,s,10,20.749383911132814,2.0749383911132813,0.01847720266802181,2.0773916015625,2.0924076904296878,2.1002045776367186,2.1064420874023435,"[2.10800146484375, 2.077233154296875, 2.090675048828125, 2.08299755859375, 2.077550048828125, 2.084087646484375, 2.068962646484375, 2.070968505859375, 2.0433494873046874, 2.045558349609375]",tokens/s,30.362347272488496,kWh,2.4737376349533467e-05,1.355642783399752e-05,6.97736295801321e-05,0.00010806743376366308,tokens/kWh,582969.33503369,,s,629,21.01463651084899,0.03340959699658028,0.0040683413241558415,0.03282534408569336,0.03364966506958008,0.03407605819702149,0.06562567657470704,"[0.03426201629638672, 0.03604585647583008, 0.03490304183959961, 0.040574943542480466, 0.036342784881591796, 0.03460812759399414, 0.03511500930786133, 0.03456819152832031, 0.034374656677246096, 0.03416064071655273, 0.034753536224365236, 0.03387187194824219, 0.03286326217651367, 0.03288982391357422, 0.03429580688476563, 0.034716670989990234, 0.03346739196777344, 0.033320960998535154, 0.03332505416870117, 0.03349401473999023, 0.03330252838134766, 0.033291263580322264, 0.032540672302246096, 0.032382976531982424, 0.03256012725830078, 0.03259392166137695, 0.032036865234375, 0.0323583984375, 0.032481281280517575, 0.03250995254516602, 0.032589824676513675, 0.032389118194580076, 0.03236761474609375, 0.032004096984863284, 0.0320706558227539, 0.03216486358642578, 0.03244134521484375, 0.03250483322143555, 0.03295334243774414, 0.03302195358276367, 0.032996353149414064, 0.03249151992797852, 0.03244441604614258, 0.03224063873291016, 0.032486400604248046, 0.03226214218139648, 0.03251302337646484, 0.03243212890625, 0.03377766418457031, 0.03330681610107422, 0.0334508171081543, 0.03508736038208008, 0.03350742340087891, 0.033242015838623046, 0.03337932968139649, 0.033486846923828126, 0.033380447387695314, 0.03334134292602539, 0.033516544342041016, 0.03349913787841797, 0.033519615173339845, 0.033554431915283206, 0.06713241577148438, 0.03251814270019531, 0.03228672027587891, 0.03229695892333984, 0.03233894348144531, 0.032353279113769534, 0.03302604675292969, 0.03314483261108398, 0.033477695465087894, 0.033221569061279294, 0.0332042236328125, 0.03198566436767578, 0.03241267013549805, 0.03217919921875, 0.03301497650146484, 0.03311494445800781, 0.033344512939453126, 0.03325132751464844, 0.03299532699584961, 0.03322163009643555, 0.03341516876220703, 0.03313356781005859, 0.03316428756713867, 0.033419391632080075, 0.03301574325561524, 0.03311199951171875, 0.0330250244140625, 0.03208396911621094, 0.03245260620117187, 0.03241062545776367, 0.032519168853759765, 0.03237068939208984, 0.03228364944458008, 0.031893503189086916, 0.03355955123901367, 0.03332403182983398, 0.033535999298095705, 0.03316121673583984, 0.033259521484375, 0.032909313201904294, 0.03245260620117187, 0.03257548904418945, 0.03335270309448242, 0.03322777557373047, 0.03356159973144531, 0.0332564468383789, 0.03325030517578125, 0.0332492790222168, 0.03347967910766601, 0.033555454254150394, 0.033516639709472655, 0.03334953689575195, 0.033107967376708985, 0.03333324813842774, 0.03346432113647461, 0.032421886444091795, 0.032361473083496094, 0.033020927429199216, 0.03334656143188477, 0.03339263916015625, 0.0336097297668457, 0.03338649749755859, 0.032083999633789065, 0.06566294097900391, 0.03242399978637695, 0.03236038589477539, 0.032276481628417966, 0.03206246566772461, 0.032350208282470705, 0.032379905700683595, 0.03327078247070313, 0.03344998550415039, 0.03341107177734375, 0.03249049758911133, 0.03262156677246094, 0.032322559356689456, 0.03207993698120117, 0.032043968200683594, 0.032328704833984374, 0.03258265686035156, 0.03280691146850586, 0.03335372924804687, 0.03371724700927734, 0.033928192138671875, 0.034783233642578126, 0.03350425720214844, 0.03309686279296875, 0.03217289733886719, 0.03322982406616211, 0.0337448959350586, 0.03395379257202148, 0.033445888519287106, 0.03337420654296875, 0.03326259231567383, 0.03350732803344727, 0.03337932968139649, 0.03338547134399414, 0.03296255874633789, 0.03361075210571289, 0.03495423889160156, 0.033492992401123044, 0.03338137435913086, 0.033242111206054685, 0.03335270309448242, 0.03321446228027344, 0.0330618896484375, 0.03313868713378906, 0.033051647186279294, 0.03337123107910156, 0.0333853759765625, 0.03339273452758789, 0.03335465621948242, 0.03240857696533203, 0.032691200256347655, 0.033683456420898435, 0.03398144149780274, 0.03329536056518555, 0.033428638458251954, 0.03344060897827148, 0.033611873626708984, 0.03377449417114258, 0.03340902328491211, 0.03355136108398438, 0.03346960067749023, 0.03338940811157227, 0.03366912078857422, 0.0679925765991211, 0.03344486236572266, 0.032589824676513675, 0.03256934356689453, 0.032247806549072264, 0.032331775665283204, 0.03263488006591797, 0.033025184631347654, 0.033526622772216796, 0.03363123321533203, 0.033358848571777344, 0.033054752349853514, 0.033529823303222656, 0.033508350372314456, 0.033341503143310545, 0.03328828811645508, 0.03319795227050781, 0.03331887817382812, 0.03338444900512695, 0.03336703872680664, 0.03355136108398438, 0.033827838897705076, 0.033442817687988284, 0.03353702545166016, 0.03315814590454102, 0.03340595245361328, 0.03322675323486328, 0.033503231048583985, 0.0334837760925293, 0.033598464965820314, 0.03364761734008789, 0.03336806488037109, 0.03325337600708008, 0.032415744781494144, 0.03284182357788086, 0.03333929443359375, 0.033701889038085936, 0.03358617782592774, 0.0335994873046875, 0.03264921569824219, 0.03205836868286133, 0.03266867065429688, 0.032628734588623046, 0.032585758209228516, 0.03253449630737305, 0.03262054443359375, 0.03220598220825195, 0.03283756637573242, 0.03292969512939453, 0.03274137496948242, 0.03234201431274414, 0.03215359878540039, 0.032840705871582034, 0.03271987152099609, 0.03233894348144531, 0.032244735717773435, 0.032573440551757815, 0.03246899032592773, 0.033323009490966796, 0.03384326553344726, 0.03332294464111328, 0.03335987091064453, 0.03327078247070313, 0.06724403381347656, 0.033887233734130856, 0.033454078674316406, 0.03342335891723633, 0.03354521560668945, 0.033027233123779295, 0.03340579223632813, 0.03290419387817383, 0.03249663925170899, 0.032522239685058595, 0.03214950561523437, 0.03219558334350586, 0.032366592407226565, 0.03216691207885742, 0.03221913528442383, 0.032998401641845705, 0.033355777740478515, 0.03334041595458984, 0.03336806488037109, 0.03384729766845703, 0.033620990753173825, 0.03345100784301758, 0.03312844848632813, 0.03322163009643555, 0.03385651016235352, 0.03347967910766601, 0.03376639938354492, 0.03314483261108398, 0.03211264038085938, 0.032473087310791016, 0.03192934417724609, 0.03241267013549805, 0.03240652847290039, 0.032494590759277346, 0.03246387100219727, 0.03290828704833984, 0.033364990234375, 0.03334656143188477, 0.03342131042480469, 0.03293088150024414, 0.03324316787719726, 0.033579936981201174, 0.03332505416870117, 0.03344179153442383, 0.03244851303100586, 0.032626686096191404, 0.03229388809204101, 0.03217407989501953, 0.03268198394775391, 0.032352256774902347, 0.03412275314331055, 0.03473612976074219, 0.03313971328735352, 0.03324937438964844, 0.03319798278808594, 0.033844223022460936, 0.03304755020141602, 0.03207987213134766, 0.03254476928710937, 0.032350208282470705, 0.032440319061279296, 0.03197337532043457, 0.03261542510986328, 0.06552985382080079, 0.032295936584472655, 0.03228057479858398, 0.032282623291015625, 0.03223756790161133, 0.03211161422729492, 0.03241984176635742, 0.032489471435546875, 0.03244134521484375, 0.03242496109008789, 0.032467967987060545, 0.03236048126220703, 0.033349601745605466, 0.03325545501708985, 0.03217404937744141, 0.03344486236572266, 0.03338240051269531, 0.03354022216796875, 0.03331571197509765, 0.03297689437866211, 0.0334919662475586, 0.03356671905517578, 0.0333199348449707, 0.03332006454467774, 0.03330342483520508, 0.03344998550415039, 0.033258495330810545, 0.03327078247070313, 0.03359641647338867, 0.03331891250610351, 0.033276927947998046, 0.033535999298095705, 0.03353497695922852, 0.03357183837890625, 0.03336710357666016, 0.03501561737060547, 0.0347586555480957, 0.033811454772949216, 0.033665088653564455, 0.03315091323852539, 0.03302604675292969, 0.033111038208007815, 0.03321548843383789, 0.03365785598754883, 0.033841152191162106, 0.033584129333496096, 0.033148929595947264, 0.03232665634155273, 0.03253145599365234, 0.03249151992797852, 0.03270560073852539, 0.03262355041503906, 0.032677886962890625, 0.032571392059326174, 0.03247513580322266, 0.03238195037841797, 0.03246080017089844, 0.03247411346435547, 0.03319193649291992, 0.03358003234863281, 0.033552383422851564, 0.033341503143310545, 0.03342943954467773, 0.06773554992675782, 0.03240652847290039, 0.0322529296875, 0.033445888519287106, 0.03304054260253906, 0.03326857757568359, 0.033018878936767575, 0.03354316711425781, 0.034129920959472655, 0.033565696716308595, 0.03357798385620117, 0.03372851181030274, 0.0334919662475586, 0.033334270477294925, 0.03297894287109375, 0.03330867385864258, 0.03314585494995117, 0.033642494201660156, 0.03354828643798828, 0.032156673431396485, 0.03203891372680664, 0.0323583984375, 0.03229388809204101, 0.03268307113647461, 0.0322303352355957, 0.0323583984375, 0.03234918212890625, 0.03239014434814453, 0.03235532760620117, 0.033248287200927734, 0.033182750701904296, 0.03302905654907227, 0.032492542266845705, 0.032092159271240234, 0.03241779327392578, 0.03222118377685547, 0.032508926391601564, 0.032244735717773435, 0.03227033615112305, 0.03237887954711914, 0.031905792236328126, 0.03237068939208984, 0.0323164176940918, 0.032540672302246096, 0.03234815979003906, 0.032648193359375, 0.03258265686035156, 0.03229388809204101, 0.0324771842956543, 0.03261644744873047, 0.03253964614868164, 0.03243008041381836, 0.03250380706787109, 0.032541694641113283, 0.03244339370727539, 0.03379609680175781, 0.03357388687133789, 0.03336703872680664, 0.033339393615722655, 0.033310752868652344, 0.03344790267944336, 0.033589313507080075, 0.0336690559387207, 0.06774169921875, 0.033633281707763675, 0.033358848571777344, 0.03467059326171875, 0.03400601577758789, 0.033685504913330076, 0.0334837760925293, 0.03338652801513672, 0.03372233581542969, 0.0338155517578125, 0.03362303924560547, 0.03344998550415039, 0.033438720703125, 0.03353190231323242, 0.03314483261108398, 0.03337932968139649, 0.032395263671875, 0.03218534469604492, 0.032312416076660154, 0.03252624130249023, 0.03250380706787109, 0.03261644744873047, 0.03348889541625977, 0.03342134475708008, 0.0336209602355957, 0.03357491302490234, 0.03350630569458008, 0.033476608276367184, 0.03317964935302734, 0.03339980697631836, 0.03351555252075195, 0.032237537384033205, 0.03235747146606445, 0.032178142547607425, 0.03349190521240234, 0.0339681282043457, 0.032557056427001956, 0.03225907135009766, 0.03232460784912109, 0.03215871810913086, 0.03219968032836914, 0.032173057556152344, 0.032489471435546875, 0.032054271697998044, 0.03183103942871094, 0.03226419067382812, 0.032350208282470705, 0.03221401596069336, 0.032353279113769534, 0.032366592407226565, 0.03233587265014649, 0.0323768310546875, 0.03253964614868164, 0.03237376022338867, 0.03253145599365234, 0.03237478256225586, 0.03243622589111328, 0.0323061752319336, 0.0324771842956543, 0.032173057556152344, 0.03259392166137695, 0.032672767639160154, 0.03218636703491211, 0.06592205047607422, 0.03243929672241211, 0.032543743133544925, 0.03255807876586914, 0.0324659194946289, 0.03224166488647461, 0.03282534408569336, 0.03225804901123047, 0.03247206497192383, 0.032473087310791016, 0.032638111114501954, 0.03223023986816406, 0.03221196746826172, 0.03254886245727539, 0.03238809585571289, 0.032551937103271485, 0.032389118194580076, 0.03272294235229492, 0.032369663238525394, 0.032726016998291016, 0.032499713897705076, 0.03233484649658203, 0.032502784729003906, 0.03261644744873047, 0.03243929672241211, 0.03239731216430664, 0.03210137557983399, 0.03238604736328125, 0.03237171173095703, 0.03223756790161133, 0.03236454391479492, 0.03194470405578613, 0.03223660659790039, 0.03229075241088867, 0.03240755081176758, 0.032451583862304685, 0.03253760147094727, 0.03243212890625, 0.03215564727783203, 0.033007614135742186, 0.03275475311279297, 0.032383937835693356, 0.032626686096191404, 0.03235753631591797, 0.032218975067138673, 0.032473087310791016, 0.033484798431396484, 0.03295129776000977, 0.03196416091918945, 0.03215359878540039, 0.03249663925170899, 0.03240758514404297, 0.032306144714355466, 0.03200614547729492, 0.03237276840209961, 0.03265737533569336, 0.03217715072631836, 0.03253452682495117, 0.03220479965209961, 0.0321003532409668, 0.032395263671875, 0.032539806365966796, 0.03214422225952149, 0.06471987152099609, 0.03255295944213867, 0.03217919921875, 0.032395263671875, 0.03217414474487305, 0.03244947052001953, 0.03224883270263672, 0.03234918212890625, 0.03223551940917969, 0.032205825805664064, 0.03241984176635742, 0.03253452682495117, 0.032454654693603514, 0.0323583984375, 0.032066654205322266, 0.032269214630126955, 0.03241164779663086, 0.03215359878540039, 0.0322242546081543, 0.03235430526733398, 0.032299007415771484, 0.032295936584472655, 0.03232153701782227, 0.03226828765869141, 0.03220684814453125, 0.03228160095214844, 0.03236761474609375, 0.032284671783447266, 0.03242092895507812, 0.03312223815917969, 0.03338956832885742, 0.03359743881225586, 0.03297792053222656, 0.032307201385498044, 0.03250483322143555, 0.03389952087402344, 0.03437059020996094, 0.03361584091186524, 0.03344384002685547, 0.03237887954711914, 0.032467967987060545, 0.032277503967285154, 0.03224063873291016, 0.03226521682739258, 0.03248332977294922, 0.03225907135009766, 0.03223046493530273, 0.032271358489990236, 0.03228768157958985, 0.03213737487792969, 0.03225276947021485, 0.03198988723754883, 0.03240537643432617, 0.03232153701782227, 0.032315391540527344, 0.03250380706787109, 0.03234099197387695, 0.03219251251220703, 0.03249868774414062, 0.03239014434814453, 0.032508926391601564, 0.0322591667175293, 0.03197942352294922]",tokens/s,29.931519380564733,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,databricks/dbrx-base,databricks/dbrx-base,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch @@ -11821,48 +11641,33 @@ ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please r ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1266.651136,1017.643008,0.0,371.195904,315.446272,s,10,0.18135516929626466,0.018135516929626467,0.000441309971516047,0.018002960205078126,0.01841672248840332,0.018911593055725096,0.01930748950958252,"[0.019406463623046873, 0.018306751251220704, 0.01791129684448242, 0.017998048782348633, 0.018007871627807617, 0.018032480239868164, 0.01785843276977539, 0.01802566337585449, 0.017830368041992187, 0.017977792739868163]",tokens/s,14115.947231798746,kWh,2.0866427887969734e-07,1.1433704629500951e-07,4.900008827740528e-07,8.130022079487598e-07,tokens/kWh,314882293.6728539,MB,1266.651136,1017.643008,0.0,371.195904,340.901888,s,10,11.019724121093748,1.101972412109375,0.0034140012491174305,1.1015155029296875,1.1048577758789062,1.1075598083496092,1.1097214343261719,"[1.1102618408203124, 1.103550537109375, 1.099968994140625, 1.1011197509765625, 1.0969412841796875, 1.1019112548828125, 1.102234375, 1.0996170654296875, 1.10425732421875, 1.0998616943359374]",tokens/s,57.170215250131875,kWh,1.251848763158427e-05,6.859667378969856e-06,2.740274365362118e-05,4.6780898664175304e-05,tokens/kWh,1346703.500765479,,s,629,11.162898439407343,0.017747056342460016,0.0021992864684729457,0.017476608276367187,0.017655813980102537,0.017812684631347657,0.03572195327758789,"[0.019810304641723633, 0.017827840805053712, 0.01761689567565918, 0.01744691276550293, 0.017640447616577147, 0.017473535537719728, 0.017508512496948243, 0.017422176361083983, 0.017655839920043947, 0.017387552261352537, 0.01757279968261719, 0.017315839767456053, 0.01743974494934082, 0.01739888000488281, 0.01757583999633789, 0.0175994873046875, 0.017535999298095704, 0.017356800079345702, 0.017364992141723632, 0.017386495590209963, 0.01721343994140625, 0.017427455902099608, 0.017482751846313475, 0.01743052864074707, 0.01737215995788574, 0.01740595245361328, 0.01744486427307129, 0.017538047790527343, 0.017513504028320314, 0.01767932891845703, 0.018437120437622072, 0.018547712326049806, 0.017557504653930665, 0.017648639678955077, 0.01840025520324707, 0.017863679885864257, 0.01803468894958496, 0.01736396789550781, 0.017521663665771483, 0.017558528900146485, 0.017506303787231444, 0.017031167984008787, 0.017283071517944337, 0.017443840026855468, 0.017765375137329103, 0.017744895935058593, 0.017523712158203125, 0.017521663665771483, 0.017947647094726564, 0.01741823959350586, 0.01761484718322754, 0.017590335845947266, 0.017390527725219727, 0.01704140853881836, 0.017517568588256836, 0.017399808883666993, 0.01785036849975586, 0.01765273666381836, 0.017484800338745117, 0.017566719055175782, 0.017485824584960938, 0.01763532829284668, 0.03625983810424805, 0.01746227264404297, 0.01762099266052246, 0.017466367721557616, 0.01738444709777832, 0.01745510482788086, 0.01743155288696289, 0.017483776092529296, 0.017560575485229494, 0.017612800598144532, 0.017482751846313475, 0.017482751846313475, 0.017504255294799806, 0.017559551239013673, 0.01760870361328125, 0.017689599990844726, 0.01754521560668945, 0.01745715141296387, 0.017494016647338868, 0.01765171241760254, 0.016977920532226562, 0.017522687911987304, 0.017343488693237305, 0.017398784637451172, 0.017474559783935546, 0.01756159973144531, 0.017330175399780275, 0.017423391342163086, 0.017575904846191405, 0.017612800598144532, 0.017354751586914064, 0.017567743301391603, 0.017539072036743163, 0.017555456161499023, 0.01759846305847168, 0.01742336082458496, 0.017443840026855468, 0.017401920318603516, 0.01774995231628418, 0.017513471603393553, 0.017285184860229494, 0.017431488037109377, 0.017699840545654297, 0.017693695068359376, 0.017543167114257813, 0.0174704647064209, 0.01762611198425293, 0.017506303787231444, 0.01756159973144531, 0.01758515167236328, 0.017512447357177736, 0.01775923156738281, 0.01740287971496582, 0.017533952713012696, 0.017612800598144532, 0.017591360092163087, 0.01751545524597168, 0.01741119956970215, 0.01738431930541992, 0.017539072036743163, 0.01761177635192871, 0.01757900810241699, 0.017451007843017577, 0.03582681655883789, 0.017428352355957032, 0.017571840286254883, 0.017694719314575197, 0.017476608276367187, 0.017496063232421876, 0.016882688522338866, 0.016925695419311524, 0.017467391967773437, 0.017550336837768556, 0.017358848571777344, 0.01739571189880371, 0.017535999298095704, 0.017332223892211913, 0.017574911117553712, 0.017481727600097655, 0.017295360565185547, 0.017506303787231444, 0.017294336318969726, 0.01738444709777832, 0.017270784378051757, 0.017327167510986327, 0.017360832214355467, 0.017458175659179686, 0.017501184463500977, 0.01739366340637207, 0.017408000946044923, 0.0174202880859375, 0.017313791275024415, 0.017532928466796875, 0.01761894416809082, 0.017509408950805664, 0.017504224777221678, 0.017555456161499023, 0.01761894416809082, 0.017510400772094727, 0.017542144775390626, 0.017559551239013673, 0.017657855987548828, 0.017580032348632812, 0.017366016387939453, 0.01760358428955078, 0.017604608535766602, 0.017284095764160155, 0.018619392395019533, 0.017853439331054686, 0.01744179153442383, 0.01741107177734375, 0.0173885440826416, 0.017353727340698243, 0.017293312072753905, 0.017308671951293944, 0.017063936233520507, 0.016698368072509767, 0.01736396789550781, 0.017317888259887695, 0.01757900810241699, 0.017391616821289063, 0.017944608688354492, 0.01769980812072754, 0.017427455902099608, 0.017360895156860352, 0.01740185546875, 0.03591584014892578, 0.017574848175048827, 0.017435647964477538, 0.01749814414978027, 0.017326208114624025, 0.017272735595703127, 0.01733318328857422, 0.017504255294799806, 0.017511423110961915, 0.017319936752319336, 0.017280000686645508, 0.01739468765258789, 0.017477632522583008, 0.017334272384643554, 0.01736396789550781, 0.01742848014831543, 0.017480703353881837, 0.017328256607055663, 0.01738534355163574, 0.01764454460144043, 0.017268735885620116, 0.017377279281616212, 0.0175852165222168, 0.017634239196777344, 0.01760358428955078, 0.01757596778869629, 0.017539039611816406, 0.01760153579711914, 0.017690624237060547, 0.017589248657226563, 0.01761996841430664, 0.017487871170043946, 0.017675264358520508, 0.017500160217285156, 0.017415168762207032, 0.017512447357177736, 0.017333248138427734, 0.017523744583129882, 0.017623008728027342, 0.01759334373474121, 0.01741619110107422, 0.01741926383972168, 0.017468416213989257, 0.0174653434753418, 0.017323007583618166, 0.017442815780639647, 0.01745305633544922, 0.017525760650634766, 0.017296384811401368, 0.01741209602355957, 0.01739263916015625, 0.01756159973144531, 0.01746124839782715, 0.01747564888000488, 0.017464256286621092, 0.017265663146972657, 0.017532928466796875, 0.017724416732788087, 0.017622016906738282, 0.017633279800415038, 0.017529855728149413, 0.017296384811401368, 0.017426431655883787, 0.035972095489501955, 0.017534975051879884, 0.017478784561157226, 0.017347455978393555, 0.017539072036743163, 0.017512447357177736, 0.017518592834472657, 0.017466367721557616, 0.017556480407714844, 0.017503231048583985, 0.017353727340698243, 0.01759436798095703, 0.01743052864074707, 0.017715200424194336, 0.02005606460571289, 0.018166784286499024, 0.01780531120300293, 0.01779302406311035, 0.017492992401123047, 0.01722470474243164, 0.01705779266357422, 0.017370111465454103, 0.017687551498413084, 0.017548288345336914, 0.01743052864074707, 0.017562623977661132, 0.017612800598144532, 0.017565696716308594, 0.017481727600097655, 0.01725542449951172, 0.017056768417358398, 0.016697343826293946, 0.01681510353088379, 0.017500223159790038, 0.017441728591918945, 0.017584287643432617, 0.017467231750488282, 0.017415168762207032, 0.0174704647064209, 0.017254463195800783, 0.017265600204467775, 0.01743974494934082, 0.017563648223876953, 0.01744588851928711, 0.017563648223876953, 0.016541759490966798, 0.016787391662597656, 0.016705535888671876, 0.016734207153320312, 0.01662371253967285, 0.016667552947998047, 0.016906240463256835, 0.017496063232421876, 0.01741107177734375, 0.017484800338745117, 0.017358879089355468, 0.017359840393066407, 0.017534975051879884, 0.016974847793579103, 0.017449983596801756, 0.01756159973144531, 0.017385471343994142, 0.017376256942749024, 0.03600896072387695, 0.01756979179382324, 0.017520671844482423, 0.01744483184814453, 0.017377279281616212, 0.017554431915283202, 0.017314815521240236, 0.01766102409362793, 0.017276832580566406, 0.017452096939086913, 0.017494976043701174, 0.01761484718322754, 0.017519615173339845, 0.017712127685546874, 0.01763839912414551, 0.017748992919921876, 0.017447935104370118, 0.0174653434753418, 0.01742540740966797, 0.017513471603393553, 0.017500160217285156, 0.017375232696533204, 0.017443840026855468, 0.01756876754760742, 0.017663999557495116, 0.017484800338745117, 0.01739673614501953, 0.01760358428955078, 0.017515520095825195, 0.017358848571777344, 0.01745715141296387, 0.01743775939941406, 0.017493951797485353, 0.017296384811401368, 0.017554431915283202, 0.017496063232421876, 0.017435647964477538, 0.017325056076049804, 0.017517568588256836, 0.017476608276367187, 0.017566719055175782, 0.017573888778686524, 0.017950719833374023, 0.017526784896850587, 0.017316864013671874, 0.017482751846313475, 0.017447935104370118, 0.017447935104370118, 0.0170199031829834, 0.017111040115356444, 0.01736403274536133, 0.017418176651000976, 0.01747148895263672, 0.017300479888916014, 0.017505279541015627, 0.017385471343994142, 0.017287168502807617, 0.01743257522583008, 0.018036735534667968, 0.01763532829284668, 0.017472511291503907, 0.01738035202026367, 0.017667072296142578, 0.035743743896484374, 0.017307743072509766, 0.017355712890625, 0.017425376892089842, 0.017517568588256836, 0.01759231948852539, 0.017534975051879884, 0.01744691276550293, 0.01760256004333496, 0.01781760025024414, 0.017624063491821287, 0.01741823959350586, 0.017567743301391603, 0.0173885440826416, 0.017472511291503907, 0.017772544860839845, 0.017665023803710937, 0.017560575485229494, 0.0174653434753418, 0.017452032089233398, 0.017339391708374022, 0.017529855728149413, 0.017574911117553712, 0.017503231048583985, 0.017526784896850587, 0.017483776092529296, 0.017374208450317383, 0.017748992919921876, 0.017480831146240234, 0.01751641654968262, 0.01740185546875, 0.01742131233215332, 0.01738956832885742, 0.01749504089355469, 0.017434623718261717, 0.01761484718322754, 0.01754521560668945, 0.01741414451599121, 0.01748684883117676, 0.017338367462158204, 0.017469440460205078, 0.01726361656188965, 0.017349632263183593, 0.01756979179382324, 0.017484800338745117, 0.017512447357177736, 0.017509376525878906, 0.017492992401123047, 0.017506303787231444, 0.017308671951293944, 0.017291263580322267, 0.017283071517944337, 0.01743974494934082, 0.017331232070922853, 0.017647584915161132, 0.01760870361328125, 0.017538047790527343, 0.017546239852905272, 0.017497087478637697, 0.017567743301391603, 0.017555456161499023, 0.01750432014465332, 0.017498048782348632, 0.03591884613037109, 0.017596416473388672, 0.01759334373474121, 0.017447935104370118, 0.017341440200805663, 0.01747865676879883, 0.017452032089233398, 0.017558528900146485, 0.01738137626647949, 0.017544191360473634, 0.017534975051879884, 0.01740595245361328, 0.017680383682250975, 0.017361919403076173, 0.01743769645690918, 0.017383424758911133, 0.017599552154541016, 0.017034175872802735, 0.017449983596801756, 0.017572864532470703, 0.01743155288696289, 0.01745715141296387, 0.017542144775390626, 0.017490943908691405, 0.017354751586914064, 0.01745408058166504, 0.017312768936157227, 0.017484800338745117, 0.017408000946044923, 0.017383424758911133, 0.017385471343994142, 0.017580032348632812, 0.017597440719604493, 0.017523744583129882, 0.017583072662353514, 0.017377279281616212, 0.01737113571166992, 0.0172728328704834, 0.01726361656188965, 0.017337343215942384, 0.017260543823242186, 0.017511423110961915, 0.01746124839782715, 0.01742848014831543, 0.017301504135131835, 0.01722368049621582, 0.017476608276367187, 0.01739776039123535, 0.017560575485229494, 0.017483776092529296, 0.017518592834472657, 0.017617919921875, 0.01765376091003418, 0.01743052864074707, 0.017280000686645508, 0.017291263580322267, 0.01760870361328125, 0.017369087219238282, 0.017460224151611328, 0.017612831115722656, 0.017415136337280274, 0.017472511291503907, 0.017464319229125978, 0.035555328369140625, 0.017348608016967772, 0.017370176315307618, 0.017435583114624023, 0.017503231048583985, 0.017515520095825195, 0.01744486427307129, 0.017552383422851564, 0.017491968154907226, 0.017554431915283202, 0.01740390396118164, 0.017349695205688475, 0.017408960342407225, 0.01743155288696289, 0.01760767936706543, 0.017442815780639647, 0.017313791275024415, 0.017312768936157227, 0.01746227264404297, 0.01741119956970215, 0.017231744766235353, 0.017435647964477538, 0.0174335994720459, 0.017276927947998046, 0.017601600646972658, 0.01750726318359375, 0.017531936645507812, 0.017571807861328125, 0.017458175659179686, 0.01760767936706543, 0.01747865676879883, 0.017755136489868165, 0.018696191787719727, 0.017904640197753906, 0.017622016906738282, 0.017473535537719728, 0.017296384811401368, 0.017526784896850587, 0.0173885440826416, 0.017452032089233398, 0.017474559783935546, 0.0175861759185791, 0.01737932777404785, 0.01739571189880371, 0.017752063751220702, 0.01741926383972168, 0.017489919662475584, 0.01760358428955078, 0.017625087738037108, 0.017310720443725586, 0.017156095504760743, 0.01740595245361328, 0.017560575485229494, 0.017872896194458008, 0.017304576873779298, 0.017477632522583008, 0.01741721534729004, 0.017315839767456053, 0.017519615173339845, 0.017522687911987304, 0.017757183074951173, 0.01919385528564453, 0.017406976699829102, 0.03566592025756836, 0.01740390396118164, 0.01744895935058594, 0.017533952713012696, 0.017326080322265625, 0.017297407150268555, 0.017399808883666993, 0.017475584030151366, 0.01743257522583008, 0.017382400512695313, 0.017358848571777344, 0.017336416244506835, 0.017456031799316405, 0.017369087219238282, 0.017475584030151366, 0.017542144775390626, 0.01741721534729004, 0.017369087219238282, 0.017469440460205078, 0.017556480407714844, 0.017510400772094727, 0.01758720016479492, 0.017229824066162108, 0.017296384811401368, 0.017532928466796875, 0.01742950439453125, 0.018282495498657226, 0.018606176376342775, 0.01762704086303711, 0.017391616821289063, 0.017655807495117186, 0.017558528900146485, 0.017449983596801756, 0.017325056076049804, 0.017382400512695313, 0.017359968185424804, 0.017475488662719727, 0.01741721534729004, 0.01752787208557129, 0.017253311157226562, 0.017290239334106446, 0.017456127166748048, 0.017573888778686524, 0.017540096282958984, 0.017415168762207032, 0.017519615173339845, 0.01763942337036133, 0.01741209602355957, 0.017307647705078123, 0.017330175399780275, 0.016890880584716796, 0.017188863754272463, 0.017229824066162108, 0.01726464080810547, 0.017331199645996095, 0.017452032089233398, 0.017540096282958984, 0.017480703353881837, 0.017528831481933595, 0.01741004753112793, 0.017466367721557616, 0.017496063232421876, 0.017494016647338868]",tokens/s,56.34737280951149,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Meta-Llama-3-70B,meta-llama/Meta-Llama-3-70B,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,,cuda,0,42,,,,,False,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,56387f0ed9acb6808de92d1e9f4c4b67cc3f2798,4.40.2,,0.30.0,,,,1.19.1,,,,0.10.0,,,MB,4000.587776,12218.53184,0.0,11572.084736,11067.6992,s,10,6.274333862304688,0.6274333862304687,0.0014418384227693484,0.6271485595703126,0.6294188720703124,0.6296758972167968,0.6298815173339843,"[0.62793310546875, 0.6299329223632812, 0.6257842407226563, 0.6269880981445313, 0.6266719360351563, 0.62570068359375, 0.6273090209960938, 0.628746337890625, 0.62590576171875, 0.6293617553710937]",tokens/s,408.0114409244492,kWh,7.395804394036532e-06,4.050755433627273e-06,4.069388672174601e-05,5.2140446549409814e-05,tokens/kWh,4909816.024636592,MB,4000.587776,12218.53184,0.0,11572.084736,11417.442816,s,10,346.35421484375,34.63542148437501,0.005481705204538505,34.636068359375,34.641064453125004,34.6420224609375,34.6427888671875,"[34.6408515625, 34.6320390625, 34.62715234375, 34.6347578125, 34.6315546875, 34.63737890625, 34.64076171875, 34.62712890625, 34.639609375, 34.64298046875]",tokens/s,1.8189471154096113,kWh,0.0004088220052400397,0.00022407180809487723,0.00225202189189386,0.002884915705228777,tokens/kWh,21837.726449273854,,s,629,351.4501739501952,0.5587443147061929,0.07457350902882966,0.5497006225585938,0.5501441894531249,0.5504188598632812,1.176382509765625,"[0.5506068725585938, 0.5495930786132812, 0.5498245239257813, 0.5495715942382813, 0.5498572998046874, 0.5496053466796875, 0.5504716796875, 0.5498091430664063, 0.549686279296875, 0.55003955078125, 0.549813232421875, 0.5506253051757812, 0.5500538940429688, 0.54976513671875, 0.5498408813476563, 0.549712890625, 0.5500037231445313, 0.5498408813476563, 0.5497374877929687, 0.5496514282226562, 0.5495634155273438, 0.5499484252929687, 0.5498009643554688, 0.549749755859375, 0.5499996337890625, 0.5498992919921875, 0.549813232421875, 0.549876708984375, 0.5494835205078125, 0.549986328125, 0.5497159423828125, 0.5493197021484375, 0.54942822265625, 0.5495449829101563, 0.5492377319335937, 0.5497302856445313, 0.5496565551757813, 0.5496156005859375, 0.5496463623046876, 0.5495572509765625, 0.5497221069335938, 0.54980712890625, 0.5500057373046875, 0.5500538940429688, 0.5496944580078125, 0.5496985473632813, 0.5496084594726562, 0.54987060546875, 0.5509058837890625, 0.5502412719726563, 0.5501102294921875, 0.54978662109375, 0.5499432983398438, 0.5499647827148437, 0.5500159912109375, 0.5497108764648437, 0.5498173217773438, 0.549696533203125, 0.5498163452148438, 0.5498910522460938, 0.5500999755859375, 0.5499238891601562, 1.180201904296875, 0.5495961303710938, 0.5500723266601563, 0.5503672485351563, 0.549634033203125, 0.550002685546875, 0.5495111694335938, 0.5496616821289062, 0.549981201171875, 0.5498695678710938, 0.5500579833984375, 0.549718017578125, 0.5496463623046876, 0.5496309814453125, 0.5493821411132812, 0.5495296020507813, 0.549802001953125, 0.5496903686523438, 0.5498193969726562, 0.5498982543945312, 0.5498163452148438, 0.5497221069335938, 0.549560302734375, 0.5495582885742187, 0.5493330078125, 0.5493002319335938, 0.5494978637695312, 0.5496187133789062, 0.549644287109375, 0.5494241333007812, 0.549265380859375, 0.5493729248046875, 0.54909130859375, 0.550029296875, 0.5500794677734375, 0.5493473510742187, 0.549444580078125, 0.54938623046875, 0.5494691772460938, 0.5502136840820312, 0.5493544921875, 0.54940771484375, 0.5494712524414063, 0.5495142211914062, 0.54980712890625, 0.5496698608398437, 0.5528350830078125, 0.5495797729492188, 0.5496033325195312, 0.5498132934570312, 0.5498224487304687, 0.5497057495117188, 0.5496361083984375, 0.549570556640625, 0.5493800659179687, 0.5498798217773437, 0.54930224609375, 0.549923828125, 0.5503836059570313, 0.5497139282226563, 0.5494732666015625, 0.549496826171875, 0.5494712524414063, 1.175841796875, 0.5491527709960937, 0.5492736206054688, 0.5495449829101563, 0.5494661254882812, 0.5495418701171875, 0.549697509765625, 0.5495429077148437, 0.5494323120117187, 0.549265380859375, 0.549232666015625, 0.5496770629882812, 0.5496279296875, 0.5497374877929687, 0.5492398071289063, 0.5496821899414063, 0.5494814453125, 0.5498685302734375, 0.5499617309570313, 0.5497620239257812, 0.5497784423828125, 0.549886962890625, 0.5496299438476563, 0.5515704345703125, 0.550213623046875, 0.5498910522460938, 0.5495255126953125, 0.5491824340820313, 0.549538818359375, 0.5496187133789062, 0.5496196899414062, 0.5493555297851562, 0.549349365234375, 0.5494251708984375, 0.5495367431640625, 0.5495818481445313, 0.5498910522460938, 0.549264404296875, 0.549528564453125, 0.5494200439453125, 0.5492828369140625, 0.549528564453125, 0.5496309814453125, 0.5498746948242188, 0.5496688842773437, 0.5493637084960937, 0.5493759765625, 0.551372802734375, 0.549791748046875, 0.549664794921875, 0.5494620361328125, 0.5496924438476563, 0.5494251708984375, 0.5495244750976562, 0.549538818359375, 0.5495643920898438, 0.5492008666992187, 0.5493524780273438, 0.54949169921875, 0.5494599609375, 0.54976513671875, 0.550413330078125, 0.54963916015625, 1.17722216796875, 0.5498256225585938, 0.5503005981445312, 0.5497630615234375, 0.5497098388671875, 0.5496985473632813, 0.5492327270507813, 0.550520751953125, 0.5497098388671875, 0.5497927856445313, 0.5496770629882812, 0.5494978637695312, 0.5504931640625, 0.5496401977539063, 0.5502904052734375, 0.5507471313476563, 0.5496740112304688, 0.5494783935546875, 0.5495623779296875, 0.5499719848632812, 0.5498204345703125, 0.5498378295898437, 0.5494824829101562, 0.5498490600585938, 0.5496207275390625, 0.5495715942382813, 0.5495480346679688, 0.549734375, 0.5497589721679688, 0.549396484375, 0.5495111694335938, 0.5495029907226563, 0.5494138793945312, 0.5499750366210937, 0.549739501953125, 0.549465087890625, 0.5495879516601563, 0.5494466552734375, 0.5499269409179688, 0.5497784423828125, 0.549307373046875, 0.5492203369140625, 0.5496996459960938, 0.5495797119140625, 0.5520455932617188, 0.5495582885742187, 0.5494783935546875, 0.5492991943359375, 0.5496453247070312, 0.5496002807617187, 0.5495111694335938, 0.550424560546875, 0.549718017578125, 0.5494630126953125, 0.5494774169921876, 0.5496810913085938, 0.549549072265625, 0.550012939453125, 0.5497477416992187, 0.5499432983398438, 0.5497426147460938, 0.5495152587890625, 0.5497979125976562, 1.1767869873046874, 0.549739501953125, 0.54965966796875, 0.5495429077148437, 0.5495787353515625, 0.5499330444335937, 0.5492398071289063, 0.5495439453125, 0.5495961303710938, 0.5495060424804687, 0.5497006225585938, 0.549695556640625, 0.5496063232421875, 0.549580810546875, 0.5494302978515625, 0.5494159545898437, 0.5495818481445313, 0.551183349609375, 0.5498726196289062, 0.5500170288085937, 0.5499586791992187, 0.5495982055664063, 0.5493883666992188, 0.5496759643554687, 0.5500333862304687, 0.5495510864257812, 0.5495152587890625, 0.5494763793945312, 0.5500599975585938, 0.5497907104492188, 0.5495081176757812, 0.54969140625, 0.5493330078125, 0.549623779296875, 0.5499668579101562, 0.5498572998046874, 0.5497098388671875, 0.5492971801757812, 0.5493268432617188, 0.5493104858398438, 0.5492162475585938, 0.549734375, 0.5499269409179688, 0.5494998779296875, 0.5495859375, 0.5493616943359375, 0.5498562622070312, 0.5494691772460938, 0.5497988891601563, 0.5496749877929688, 0.5496934814453125, 0.5498726196289062, 0.5495582885742187, 0.549538818359375, 0.5493759765625, 0.5495357666015624, 0.5493759765625, 0.55271728515625, 0.5497804565429687, 0.5499638061523437, 0.5498859252929688, 0.5494845581054687, 0.5494835205078125, 1.176280029296875, 0.5498716430664062, 0.5497538452148437, 0.550128662109375, 0.5498490600585938, 0.5496893310546875, 0.5495398559570313, 0.5497907104492188, 0.5496514282226562, 0.5496740112304688, 0.549696533203125, 0.550066162109375, 0.5496995849609375, 0.5500098266601563, 0.5500528564453125, 0.5500682373046875, 0.549749755859375, 0.5495162963867187, 0.5499627685546875, 0.5494998779296875, 0.5500108642578125, 0.5495879516601563, 0.54917431640625, 0.5495316162109375, 0.5493677978515625, 0.5499985961914062, 0.55001806640625, 0.549771240234375, 0.5498153076171876, 0.5497159423828125, 0.5498163452148438, 0.5502157592773438, 0.5497650756835938, 0.5495982055664063, 0.5495787353515625, 0.54948046875, 0.549781494140625, 0.5504225463867187, 0.55213671875, 0.5497979125976562, 0.549275634765625, 0.5496381225585938, 0.54999755859375, 0.5501982421875, 0.550055908203125, 0.5494722290039062, 0.5493514404296875, 0.5492357177734375, 0.5492142333984374, 0.5495664672851562, 0.5501439819335937, 0.5502484741210938, 0.5494712524414063, 0.5494579467773437, 0.5491691284179687, 0.5500702514648438, 0.5497671508789063, 0.549982177734375, 0.54965966796875, 0.549876708984375, 0.5498112182617187, 0.5495521240234374, 0.5497507934570313, 1.177450439453125, 0.5494251708984375, 0.5496135864257813, 0.54983984375, 0.5499371337890625, 0.5500845947265625, 0.550044677734375, 0.5497057495117188, 0.5494620361328125, 0.5499945068359375, 0.549612548828125, 0.5502515258789062, 0.549391357421875, 0.5496504516601562, 0.5496207275390625, 0.5498306274414062, 0.5498255615234375, 0.5497354125976562, 0.5500078125, 0.549634033203125, 0.549880859375, 0.5498316650390624, 0.5501747436523438, 0.5520926513671875, 0.5496801147460938, 0.5497426147460938, 0.5495900268554688, 0.5498982543945312, 0.5500016479492188, 0.5504071655273437, 0.5496565551757813, 0.5497374877929687, 0.5496524658203125, 0.5498357543945313, 0.5504757690429688, 0.5500744018554687, 0.5496309204101563, 0.5495275268554688, 0.5499218139648437, 0.5500528564453125, 0.5499259033203125, 0.5500927734375, 0.5498378295898437, 0.5497098388671875, 0.549464111328125, 0.549540771484375, 0.5494671630859375, 0.5494425659179687, 0.5492879638671875, 0.5496729736328125, 0.549949462890625, 0.5500682373046875, 0.5503406372070313, 0.54990234375, 0.5496832275390625, 0.549739501953125, 0.5494528198242188, 0.5497006225585938, 0.5501102294921875, 0.5501583251953125, 0.5500590209960937, 0.5497200927734375, 0.5498224487304687, 1.178892333984375, 0.5495838623046875, 0.5494957885742188, 0.5494210815429688, 0.5492039794921875, 0.5495214233398438, 0.54934326171875, 0.5497119140625, 0.5494619750976563, 0.5495746459960937, 0.5495081176757812, 0.5494824829101562, 0.5495889892578125, 0.5497364501953125, 0.5494661254882812, 0.5495736083984375, 0.5508587646484375, 0.5497988891601563, 0.5496689453125, 0.5499104614257813, 0.5496781005859375, 0.549433349609375, 0.5494845581054687, 0.5494364013671875, 0.549982177734375, 0.5496104736328125, 0.5497610473632812, 0.5496873168945312, 0.5494507446289062, 0.5494353637695313, 0.549749755859375, 0.5502740478515625, 0.5493923950195313, 0.5493842163085938, 0.5498470458984375, 0.5492612915039062, 0.5498347778320313, 0.5493955078125, 0.5496390991210938, 0.5495869140625, 0.5492551879882812, 0.5493821411132812, 0.5500845947265625, 0.5495992431640625, 0.5492684936523438, 0.54944970703125, 0.5492981567382812, 0.5493954467773438, 0.5493319702148437, 0.5496228637695313, 0.5499022827148438, 0.54959716796875, 0.549496826171875, 0.5490667724609375, 0.5494989013671875, 0.5497743530273438, 0.549411865234375, 0.551868408203125, 0.54951220703125, 0.5497180786132813, 0.5496380615234375, 0.549728271484375, 0.5498746948242188, 1.17642236328125, 0.5495439453125, 0.5496812133789063, 0.5500067138671875, 0.5500466918945313, 0.5497927856445313, 0.5496145629882813, 0.5498603515625, 0.5499801635742188, 0.5497774047851562, 0.5499218139648437, 0.5500252075195312, 0.5499617309570313, 0.5498746948242188, 0.5501552734375, 0.5501890258789063, 0.5498470458984375, 0.5514967041015625, 0.5500538940429688, 0.549981201171875, 0.5499453735351563, 0.5500723266601563, 0.5503211669921875, 0.5499566040039062, 0.5496749877929688, 0.5495214233398438, 0.55014501953125, 0.5498644409179687, 0.5502791748046875, 0.5495111694335938, 0.5496258544921875, 0.5497467041015625, 0.5499156494140625, 0.5499330444335937, 0.5498101806640625, 0.549422119140625, 0.5496688232421875, 0.549454833984375, 0.5499730224609375, 0.5499525146484375, 0.5496156005859375, 0.5495214233398438, 0.5496279296875, 0.54949169921875, 0.5499443359375, 0.5496893310546875, 0.5496371459960937, 0.5497661743164063, 0.5493309326171875, 0.5493667602539063, 0.5495982055664063, 0.5498716430664062, 0.5500508422851562, 0.5496002807617187, 0.549580810546875, 0.5497518310546875, 0.5494763793945312, 0.5502894287109374, 0.5498040161132812, 0.5496934204101562, 0.5494661254882812, 0.5495572509765625, 0.5497640991210937, 1.1800985107421875, 0.5502362060546875, 0.5499596557617188, 0.5499771118164063, 0.5499320068359375, 0.54993408203125, 0.5496924438476563, 0.5497088012695313, 0.54962890625, 0.5497692260742187, 0.5496719360351563, 0.5496749877929688, 0.550171630859375, 0.5498040161132812, 0.5495132446289063, 0.5494343872070313, 0.5496535034179687, 0.5498522338867188, 0.5498715209960937, 0.55016650390625, 0.5498357543945313, 0.54962890625, 0.5497313232421875, 0.5496842041015625, 0.5497507934570313, 0.54953369140625, 0.55164111328125, 0.5496196899414062, 0.550055908203125, 0.5495173950195312, 0.5501920776367187, 0.5497261962890625, 0.5495695190429688, 0.5494876098632813, 0.5496985473632813, 0.5498613891601563, 0.5497979125976562, 0.5498275756835938, 0.5494743041992187, 0.5495643920898438, 0.549391357421875, 0.5497763671875, 0.5498480834960937, 0.5502494506835938, 0.550213623046875, 0.549992431640625, 0.5500907592773437, 0.5497804565429687, 0.550150146484375, 0.5504419555664063, 0.5499197387695313, 0.5499852905273438, 0.5498756713867188, 0.5500354614257813, 0.5499658203125, 0.5504122924804687, 0.5497241821289063, 0.549796875, 0.5498849487304688, 0.549918701171875, 0.5500098266601563, 0.5499218139648437, 0.5500538330078125]",tokens/s,1.789727382775849,,,main,False,False,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,5050.339328,5956.435968,0.0,5309.988864,5047.4112,s,10,3.5051743164062503,0.350517431640625,0.0007670843399549683,0.3503496398925781,0.3513632080078125,0.35185547790527344,0.35224929382324216,"[0.35234774780273437, 0.35060940551757813, 0.34943450927734376, 0.3503487548828125, 0.35035052490234375, 0.34977896118164065, 0.35125381469726563, 0.35018606567382815, 0.35063809204101565, 0.3502264404296875]",tokens/s,730.3488411454216,kWh,4.129352663211896e-06,2.262697242233493e-06,1.994309066710337e-05,2.6335140572548764e-05,tokens/kWh,9720851.851721248,MB,5050.339328,5956.435968,0.0,5309.988864,5098.333184,s,10,196.294599609375,19.629459960937503,0.003306332269520929,19.629046875,19.633500976562498,19.63416455078125,19.634695410156247,"[19.629720703125, 19.632705078125, 19.6308828125, 19.628373046875, 19.62533203125, 19.634828125, 19.627736328125, 19.633353515625, 19.627451171875, 19.624216796875]",tokens/s,3.2094617032444903,kWh,0.00023168927829041793,0.00012698317002905933,0.001085903571787293,0.0014445760201067701,tokens/kWh,43611.412015093265,,s,629,199.13568667602536,0.31659091681403084,0.04163273062294084,0.3115284423828125,0.3119087646484375,0.3120863220214844,0.6618047436523438,"[0.31184384155273437, 0.3115591735839844, 0.31158578491210936, 0.31151007080078125, 0.31131027221679686, 0.3114598388671875, 0.3118182373046875, 0.3113861083984375, 0.3115765686035156, 0.3114055786132813, 0.31260980224609375, 0.312263671875, 0.3118950500488281, 0.3115806579589844, 0.3114864501953125, 0.31126220703125, 0.3116441650390625, 0.3115018310546875, 0.3115315246582031, 0.31142501831054686, 0.311541748046875, 0.3113861083984375, 0.31136972045898437, 0.3112591247558594, 0.3113461608886719, 0.31131954956054686, 0.3114598388671875, 0.3114537048339844, 0.3115653076171875, 0.311541748046875, 0.3114598388671875, 0.3115110473632812, 0.31137484741210936, 0.3119503479003906, 0.312026123046875, 0.31166055297851564, 0.31153048706054687, 0.3116718139648437, 0.311868408203125, 0.3118182373046875, 0.3115540466308594, 0.311741455078125, 0.3115653076171875, 0.3113912353515625, 0.3117998046875, 0.31146392822265623, 0.3114485778808594, 0.3119083557128906, 0.31169125366210937, 0.31166055297851564, 0.311630859375, 0.31203225708007815, 0.3113512878417969, 0.3113431091308594, 0.31123968505859373, 0.3115110473632812, 0.31138302612304686, 0.31147930908203125, 0.31122125244140625, 0.311330810546875, 0.31138302612304686, 0.31175167846679686, 0.6627184448242187, 0.3113984069824219, 0.31170968627929685, 0.3114127502441406, 0.31126220703125, 0.31154278564453125, 0.3114198913574219, 0.3116042175292969, 0.3116922912597656, 0.31159194946289065, 0.3114813537597656, 0.3120394287109375, 0.3122575378417969, 0.31202304077148435, 0.3119810485839844, 0.3112878112792969, 0.3118858337402344, 0.3115724792480469, 0.3115530090332031, 0.31155609130859374, 0.3114547119140625, 0.3113154602050781, 0.3114690551757813, 0.31146597290039063, 0.31144140625, 0.3114967041015625, 0.3115970458984375, 0.3113584594726562, 0.3114547119140625, 0.3111997375488281, 0.3114035339355469, 0.31180084228515625, 0.3114537048339844, 0.31141683959960936, 0.31198309326171875, 0.3117250671386719, 0.3118499755859375, 0.31170968627929685, 0.31176602172851564, 0.31219302368164065, 0.31145062255859374, 0.3116390380859375, 0.31146087646484377, 0.3120506896972656, 0.31171890258789064, 0.31187353515625, 0.31155813598632814, 0.31151513671875, 0.31166156005859375, 0.31171173095703125, 0.311625732421875, 0.31178546142578123, 0.311546875, 0.3118653564453125, 0.31160525512695314, 0.3117281188964844, 0.3115683898925781, 0.31145779418945313, 0.3114844055175781, 0.31155508422851563, 0.3114342346191406, 0.31170968627929685, 0.31176397705078124, 0.6612152099609375, 0.3113779296875, 0.31236709594726564, 0.3114844055175781, 0.31142501831054686, 0.311731201171875, 0.31146701049804687, 0.3121192932128906, 0.31149774169921873, 0.3114127502441406, 0.3114178466796875, 0.3117547607421875, 0.3113584594726562, 0.3121397705078125, 0.31130624389648437, 0.31131442260742187, 0.3115397033691406, 0.3114137573242188, 0.31158273315429685, 0.311583740234375, 0.31153765869140626, 0.31172607421875, 0.311552001953125, 0.31155712890625, 0.3116072998046875, 0.3116718139648437, 0.311404541015625, 0.3113922424316406, 0.31131137084960936, 0.31155609130859374, 0.3116861572265625, 0.311625732421875, 0.31160525512695314, 0.31175271606445315, 0.3120947265625, 0.3116974182128906, 0.3114618835449219, 0.31171585083007813, 0.311794677734375, 0.31187353515625, 0.31170355224609375, 0.31179571533203126, 0.3116021728515625, 0.3116339111328125, 0.31165338134765624, 0.31194113159179687, 0.31137588500976565, 0.3114649658203125, 0.3115130920410156, 0.3115212707519531, 0.3118940124511719, 0.31150079345703124, 0.3114434509277344, 0.3113594970703125, 0.31137484741210936, 0.3116216430664063, 0.31164825439453125, 0.3118807067871094, 0.3113963623046875, 0.3114967041015625, 0.3115448303222656, 0.31148236083984376, 0.311404541015625, 0.6620692749023438, 0.31139431762695313, 0.311910400390625, 0.3116124267578125, 0.3113605041503906, 0.311372802734375, 0.31147314453125, 0.3112949829101562, 0.3116492919921875, 0.3115335693359375, 0.31172402954101563, 0.31153253173828127, 0.31165643310546876, 0.31194622802734373, 0.31182745361328124, 0.31166055297851564, 0.3115673522949219, 0.31169537353515625, 0.31190118408203127, 0.3117230224609375, 0.31166873168945314, 0.3114342346191406, 0.3116390380859375, 0.31148544311523435, 0.31161038208007813, 0.3112949829101562, 0.31146701049804687, 0.31134002685546874, 0.3115683898925781, 0.3114055786132813, 0.31139736938476564, 0.311583740234375, 0.31133389282226565, 0.3113963623046875, 0.31168511962890627, 0.31193191528320313, 0.3114076232910156, 0.31169638061523436, 0.31148031616210936, 0.31190118408203127, 0.3117271118164063, 0.3115970458984375, 0.3113994140625, 0.31178546142578123, 0.311657470703125, 0.31192166137695315, 0.3116646423339844, 0.311267333984375, 0.3113789367675781, 0.31148544311523435, 0.3116738586425781, 0.3113932800292969, 0.31144549560546875, 0.3118499755859375, 0.31140966796875, 0.311414794921875, 0.31144549560546875, 0.3115284423828125, 0.31134515380859373, 0.3114117126464844, 0.31143118286132815, 0.3113779296875, 0.3115284423828125, 0.662192138671875, 0.31138302612304686, 0.3115816955566406, 0.31156121826171873, 0.31155813598632814, 0.3113809814453125, 0.31169329833984377, 0.31178546142578123, 0.3113799743652344, 0.3115540466308594, 0.3119083557128906, 0.31207220458984375, 0.31151617431640627, 0.31160321044921874, 0.31140045166015623, 0.31141683959960936, 0.3114158020019531, 0.3113851013183594, 0.31130419921875, 0.3115263977050781, 0.3114127502441406, 0.3112949829101562, 0.31136871337890626, 0.3116789855957031, 0.311330810546875, 0.3113861083984375, 0.31132672119140625, 0.3115038757324219, 0.31150796508789064, 0.31133285522460935, 0.3114158020019531, 0.31143936157226565, 0.31139736938476564, 0.31150283813476565, 0.31133285522460935, 0.3117619323730469, 0.31155712890625, 0.31122637939453124, 0.311436279296875, 0.3119288330078125, 0.3116789855957031, 0.3114844055175781, 0.31126629638671877, 0.3119165344238281, 0.31151617431640627, 0.3117782897949219, 0.31169125366210937, 0.31161651611328123, 0.3115397033691406, 0.3115397033691406, 0.3114281005859375, 0.311541748046875, 0.31147213745117186, 0.3113809814453125, 0.3113226318359375, 0.311362548828125, 0.31155508422851563, 0.31144244384765624, 0.31143014526367185, 0.3113543701171875, 0.31150592041015623, 0.31159295654296876, 0.31152947998046876, 0.6616196899414063, 0.3113656311035156, 0.3113799743652344, 0.3122022399902344, 0.31157760620117186, 0.3115438232421875, 0.3114352722167969, 0.3118960571289062, 0.31142706298828127, 0.311468017578125, 0.311552001953125, 0.31180389404296877, 0.3113543701171875, 0.311900146484375, 0.3113994140625, 0.31151513671875, 0.3119503479003906, 0.311330810546875, 0.31155508422851563, 0.3114649658203125, 0.3115130920410156, 0.31137176513671877, 0.3114547119140625, 0.31152435302734377, 0.3127951354980469, 0.3118182373046875, 0.31182131958007814, 0.31145062255859374, 0.31154791259765624, 0.31154074096679685, 0.3115970458984375, 0.31179571533203126, 0.31168307495117187, 0.311841796875, 0.31201278686523437, 0.3121745910644531, 0.31208447265625, 0.31171380615234373, 0.3116697692871094, 0.3120271301269531, 0.3118397521972656, 0.31170150756835935, 0.31159295654296876, 0.312099853515625, 0.3114629211425781, 0.31185101318359376, 0.3115643005371094, 0.3116093444824219, 0.3114649658203125, 0.3114905700683594, 0.3116656494140625, 0.3116021728515625, 0.3117055969238281, 0.3115284423828125, 0.31146392822265623, 0.31161651611328123, 0.31171072387695314, 0.3116011657714844, 0.3118284912109375, 0.3114260559082031, 0.3115816955566406, 0.31153253173828127, 0.3115192260742187, 0.6630952758789063, 0.31141683959960936, 0.31164312744140626, 0.31152743530273436, 0.3114291076660156, 0.3113584594726562, 0.31125811767578127, 0.3115335693359375, 0.3115970458984375, 0.3116779479980469, 0.31240090942382814, 0.31191244506835936, 0.31150283813476565, 0.3117332458496094, 0.31130419921875, 0.31145166015625, 0.31143731689453125, 0.3113594970703125, 0.3113861083984375, 0.3115816955566406, 0.3115192260742187, 0.3116779479980469, 0.31144140625, 0.3115663452148438, 0.3115591735839844, 0.3113912353515625, 0.3116390380859375, 0.3114291076660156, 0.3114649658203125, 0.3113861083984375, 0.31137176513671877, 0.31148748779296875, 0.3113502807617187, 0.31146392822265623, 0.3112017822265625, 0.31204556274414064, 0.31150796508789064, 0.3115745239257813, 0.31130419921875, 0.31160830688476565, 0.31174041748046877, 0.3113707580566406, 0.3113738098144531, 0.3120425109863281, 0.31171890258789064, 0.31209063720703123, 0.31143218994140626, 0.3114178466796875, 0.31159091186523435, 0.31140863037109373, 0.31129190063476564, 0.3126456298828125, 0.3115643005371094, 0.3116707763671875, 0.31138302612304686, 0.31131341552734376, 0.31147930908203125, 0.31155712890625, 0.3114762268066406, 0.3114700927734375, 0.3113543701171875, 0.31164620971679685, 0.3114014587402344, 0.661876708984375, 0.31147213745117186, 0.31161038208007813, 0.31178854370117187, 0.3116369934082031, 0.31138406372070315, 0.3113994140625, 0.3121100769042969, 0.31174554443359376, 0.3116502990722656, 0.31147930908203125, 0.3120486450195312, 0.31169125366210937, 0.31182437133789065, 0.31171685791015624, 0.31158273315429685, 0.31168307495117187, 0.31170968627929685, 0.31170458984375, 0.31167486572265624, 0.31185101318359376, 0.3117701110839844, 0.31155813598632814, 0.3116707763671875, 0.31168203735351563, 0.31149774169921873, 0.31163595581054687, 0.31133389282226565, 0.3113799743652344, 0.311541748046875, 0.31161038208007813, 0.3115704345703125, 0.31139431762695313, 0.3114598388671875, 0.3114281005859375, 0.3118059387207031, 0.31162777709960937, 0.3116441650390625, 0.3114014587402344, 0.3115048828125, 0.31208755493164064, 0.31144244384765624, 0.3115284423828125, 0.31309515380859376, 0.3117025146484375, 0.3120343017578125, 0.3119083557128906, 0.3115530090332031, 0.3115448303222656, 0.31152230834960937, 0.311510009765625, 0.3115069580078125, 0.3117506408691406, 0.3115540466308594, 0.3115069580078125, 0.3114137573242188, 0.3117055969238281, 0.31158477783203126, 0.3115284423828125, 0.311593994140625, 0.3115950012207031, 0.3112540283203125, 0.3114895324707031, 0.662466552734375, 0.3114035339355469, 0.3113226318359375, 0.31210906982421877, 0.31197491455078125, 0.3115673522949219, 0.31160830688476565, 0.3116871643066406, 0.3120425109863281, 0.3115448303222656, 0.3114495849609375, 0.31190631103515626, 0.311468017578125, 0.31199847412109377, 0.31142501831054686, 0.31138714599609374, 0.31127142333984376, 0.31176397705078124, 0.31159808349609375, 0.3120086975097656, 0.3115663452148438, 0.31146597290039063, 0.3112069091796875, 0.3112509460449219, 0.31143218994140626, 0.3113605041503906, 0.31138919067382814, 0.3113861083984375, 0.31139532470703124, 0.3112755126953125, 0.31115264892578126, 0.3114260559082031, 0.3111403503417969, 0.3112017822265625, 0.31128677368164065, 0.31140045166015623, 0.3121039428710938, 0.3113543701171875, 0.31140045166015623, 0.31126425170898436, 0.3120466003417969, 0.3118028869628906, 0.31142196655273435, 0.31180084228515625, 0.31138302612304686, 0.31169842529296876, 0.31304293823242185, 0.3114485778808594, 0.31143832397460935, 0.31143118286132815, 0.3114281005859375, 0.3113574523925781, 0.3113984069824219, 0.31149261474609374, 0.311372802734375, 0.311404541015625, 0.3115397033691406, 0.31151513671875, 0.31149362182617185, 0.3114629211425781, 0.3114629211425781, 0.3115069580078125, 0.31171173095703125, 0.6619484252929687, 0.31144549560546875, 0.3115090026855469, 0.31149978637695314, 0.31185614013671875, 0.3114260559082031, 0.31134002685546874, 0.3112959899902344, 0.3115192260742187, 0.31148748779296875, 0.31161343383789064, 0.31172607421875, 0.31162673950195313, 0.31148031616210936, 0.31136767578125, 0.31123147583007815, 0.31138919067382814, 0.31133184814453124, 0.3117342834472656, 0.3115806579589844, 0.31134002685546874, 0.3112939453125, 0.31161138916015624, 0.3114342346191406, 0.31162469482421873, 0.3115069580078125, 0.3114403991699219, 0.31136358642578127, 0.3114014587402344, 0.3113512878417969, 0.31128167724609374, 0.3113697509765625, 0.31135638427734375, 0.31134515380859373, 0.3115069580078125, 0.3115048828125, 0.31214797973632813, 0.31165338134765624, 0.31144549560546875, 0.31144549560546875, 0.31171072387695314, 0.31162060546875, 0.3115868225097656, 0.3114710998535156, 0.311947265625, 0.31166873168945314, 0.31137484741210936, 0.31127655029296875, 0.31169638061523436, 0.311468017578125, 0.31132672119140625, 0.311436279296875, 0.31144140625, 0.31263333129882814, 0.31138919067382814, 0.3112171630859375, 0.31134515380859373, 0.3113707580566406, 0.31140045166015623, 0.3112458190917969, 0.31124273681640624, 0.3112949829101562, 0.31147930908203125]",tokens/s,3.158650317777157,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-7b,google/gemma-7b,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.0,,,,1.21.4,,,,0.12.0,,,MB,1768.828928,7019.692032,0.0,6371.147776,6308.565504,s,10,1.3267521209716797,0.13267521209716798,0.00021879637069722476,0.13260414886474609,0.13292575378417967,0.13306385192871092,0.13317433044433594,"[0.13272665405273437, 0.13272746276855468, 0.13244419860839843, 0.13244650268554686, 0.1328950653076172, 0.13320195007324218, 0.13256777954101562, 0.13261747741699217, 0.13253421020507813, 0.1325908203125]",tokens/s,1929.5239551794502,kWh,1.5685975723319054e-06,8.595015667302536e-07,6.284924618578928e-06,8.713023757641087e-06,tokens/kWh,29381304.024965487,MB,1807.290368,7021.789184,0.0,6373.244928,6308.568064,s,10,19.6127119140625,1.9612711914062502,0.003016630357559812,1.9599844970703124,1.9660078369140626,1.9668467529296876,1.9675178857421876,"[1.9658214111328125, 1.9590687255859376, 1.9591705322265625, 1.9598502197265626, 1.963131591796875, 1.9676856689453126, 1.95864892578125, 1.9601187744140625, 1.9590291748046875, 1.9601868896484376]",tokens/s,32.12202385679688,kWh,2.3443418473640925e-05,1.284761757032325e-05,7.058219243062118e-05,0.00010687322847458537,tokens/kWh,589483.4552975211,,s,630,19.610716314315805,0.031128121133834594,0.000435886769887594,0.031071231842041015,0.031481651878356935,0.03185622978210449,0.03283826656341553,"[0.03183718490600586, 0.031105024337768555, 0.03143065643310547, 0.03130060768127441, 0.031247360229492187, 0.031123455047607423, 0.030587903976440428, 0.03054489517211914, 0.030940160751342774, 0.03395993423461914, 0.0319109115600586, 0.030901248931884766, 0.031102975845336913, 0.0314726390838623, 0.030920703887939452, 0.030887935638427736, 0.031340543746948245, 0.03163443183898926, 0.03196416091918945, 0.031070207595825194, 0.03138150405883789, 0.03114188766479492, 0.031102975845336913, 0.03073023986816406, 0.031079423904418944, 0.03341926574707031, 0.03203379058837891, 0.03193036842346191, 0.0315156478881836, 0.031137792587280274, 0.030879743576049806, 0.030889984130859374, 0.031156223297119142, 0.030963712692260743, 0.031025152206420898, 0.03210854339599609, 0.03270348739624023, 0.030883840560913086, 0.030612480163574218, 0.030623743057250977, 0.031119359970092773, 0.030616575241088868, 0.03060736083984375, 0.030826496124267577, 0.030669824600219726, 0.03056844711303711, 0.030983167648315428, 0.03129343986511231, 0.03082137680053711, 0.030665727615356447, 0.03157708740234375, 0.03137433624267578, 0.030901248931884766, 0.03114905548095703, 0.030681087493896485, 0.03055718421936035, 0.03055718421936035, 0.03076300811767578, 0.03140403175354004, 0.03155660820007324, 0.03057868766784668, 0.030769151687622072, 0.030976160049438477, 0.030742528915405274, 0.031252479553222655, 0.031110143661499022, 0.031114240646362305, 0.031054847717285155, 0.030934015274047853, 0.031129600524902344, 0.030875648498535156, 0.03102822494506836, 0.031102975845336913, 0.03096780776977539, 0.031069183349609376, 0.030881792068481444, 0.030649343490600587, 0.030793727874755858, 0.03146649551391602, 0.03074764823913574, 0.03076710319519043, 0.03203583908081055, 0.03134976005554199, 0.030741504669189453, 0.030871551513671876, 0.031178752899169923, 0.03159654426574707, 0.031199264526367187, 0.030904287338256835, 0.03138150405883789, 0.031121408462524414, 0.03282227325439453, 0.031100927352905275, 0.030860288619995117, 0.03080499267578125, 0.031121408462524414, 0.03135385513305664, 0.03095244789123535, 0.031152128219604492, 0.031092735290527345, 0.03080089569091797, 0.03134464073181152, 0.031162368774414063, 0.031031295776367186, 0.030876672744750977, 0.030769151687622072, 0.03074662399291992, 0.03083263969421387, 0.031058944702148438, 0.03082035255432129, 0.032008190155029294, 0.03126681518554687, 0.031038463592529295, 0.031068159103393556, 0.030930944442749023, 0.031222784042358398, 0.03119923210144043, 0.03097702407836914, 0.031253503799438476, 0.031062015533447264, 0.031336448669433595, 0.031124479293823244, 0.030910463333129884, 0.030735359191894532, 0.03078144073486328, 0.03120025634765625, 0.0313384952545166, 0.03266867065429688, 0.03190784072875977, 0.03137433624267578, 0.03096780776977539, 0.031102975845336913, 0.030861312866210938, 0.030814207077026368, 0.031235071182250978, 0.03058073616027832, 0.03079270362854004, 0.03120639991760254, 0.03117568016052246, 0.031072256088256835, 0.031169536590576172, 0.031068159103393556, 0.031109119415283205, 0.030893056869506837, 0.030889984130859374, 0.031194112777709962, 0.031093759536743162, 0.031214591979980468, 0.031033344268798828, 0.0311592960357666, 0.030993408203125, 0.031085567474365236, 0.03129241561889649, 0.03075379180908203, 0.031031295776367186, 0.031124479293823244, 0.031062015533447264, 0.031090688705444337, 0.030955520629882813, 0.03081625556945801, 0.031078399658203124, 0.030930944442749023, 0.03122483253479004, 0.030885887145996094, 0.030922752380371094, 0.031148031234741212, 0.03094937515258789, 0.030889984130859374, 0.030881792068481444, 0.03078963279724121, 0.030835712432861328, 0.03116543960571289, 0.030999551773071288, 0.03096063995361328, 0.0312729606628418, 0.031321088790893556, 0.0309749755859375, 0.030993408203125, 0.03165081596374512, 0.03130060768127441, 0.03103539276123047, 0.03121151924133301, 0.031369216918945314, 0.031063039779663085, 0.03116748809814453, 0.03102841567993164, 0.031013696670532227, 0.031056896209716797, 0.03073843193054199, 0.03096883201599121, 0.030915584564208985, 0.031239168167114258, 0.030889984130859374, 0.03075481605529785, 0.030999551773071288, 0.031088640213012695, 0.03075174331665039, 0.031101951599121092, 0.031220735549926756, 0.03117465591430664, 0.030971904754638672, 0.030993408203125, 0.03133030319213867, 0.03119615936279297, 0.03078348731994629, 0.03077734375, 0.03121151924133301, 0.031253503799438476, 0.030702592849731446, 0.03103539276123047, 0.031237119674682616, 0.03115519905090332, 0.030873600006103515, 0.030902271270751954, 0.030958591461181642, 0.030866432189941406, 0.03151872062683105, 0.031180799484252928, 0.031081472396850586, 0.030931968688964844, 0.030882816314697265, 0.03080806350708008, 0.03080601692199707, 0.03082854461669922, 0.030700544357299804, 0.03135897636413574, 0.031230976104736328, 0.03102617645263672, 0.03099443244934082, 0.031048704147338867, 0.03116748809814453, 0.031205375671386718, 0.03082444763183594, 0.031177728652954102, 0.03092889595031738, 0.031047679901123046, 0.031007743835449218, 0.030988288879394532, 0.030728191375732423, 0.031054847717285155, 0.031122432708740235, 0.03119001579284668, 0.030825471878051756, 0.03100569534301758, 0.031036415100097657, 0.031015935897827147, 0.03116851234436035, 0.031038463592529295, 0.031243263244628908, 0.03395174407958984, 0.032299007415771484, 0.031870975494384765, 0.0314726390838623, 0.031089664459228516, 0.031247360229492187, 0.03139276885986328, 0.031225887298583985, 0.031335391998291016, 0.03152179145812988, 0.03079475212097168, 0.03098419189453125, 0.03093606376647949, 0.03133132743835449, 0.031138816833496095, 0.030875648498535156, 0.031160320281982422, 0.03123200035095215, 0.03124838447570801, 0.031162368774414063, 0.031065088272094726, 0.03129446411132813, 0.031834112167358396, 0.03244441604614258, 0.031239168167114258, 0.031422464370727536, 0.031047679901123046, 0.031161344528198243, 0.03126067161560059, 0.03131596755981445, 0.03118694305419922, 0.0321710090637207, 0.03121049690246582, 0.031025152206420898, 0.030801919937133788, 0.030980096817016602, 0.03135078430175781, 0.03077324867248535, 0.03082444763183594, 0.0311592960357666, 0.03096575927734375, 0.031081472396850586, 0.03080806350708008, 0.03060223960876465, 0.031031295776367186, 0.031185920715332032, 0.03079475212097168, 0.03174198341369629, 0.031070175170898436, 0.031197183609008788, 0.03131494331359863, 0.03120128059387207, 0.030908416748046875, 0.03103443145751953, 0.03130668830871582, 0.03118182373046875, 0.030899200439453125, 0.030859264373779297, 0.03126169586181641, 0.031498239517211916, 0.030991359710693358, 0.031213567733764647, 0.030695423126220703, 0.030926847457885744, 0.030643199920654295, 0.03077017593383789, 0.03135487937927246, 0.03146444892883301, 0.03119820785522461, 0.03112550354003906, 0.031302656173706055, 0.03122483253479004, 0.031145984649658204, 0.030915584564208985, 0.030870559692382813, 0.031114208221435548, 0.03248230361938476, 0.032129024505615236, 0.031335424423217774, 0.031546367645263675, 0.03103539276123047, 0.031073280334472656, 0.031129600524902344, 0.031094783782958983, 0.03132928085327148, 0.03100979232788086, 0.03380428695678711, 0.03284479904174804, 0.03158835220336914, 0.03123200035095215, 0.031074304580688477, 0.030926847457885744, 0.031115264892578126, 0.031253503799438476, 0.031214591979980468, 0.030819328308105468, 0.031041536331176758, 0.03082956886291504, 0.030934015274047853, 0.03136819267272949, 0.030934015274047853, 0.030843904495239258, 0.031014911651611327, 0.030899200439453125, 0.030896127700805662, 0.030996480941772462, 0.030818304061889647, 0.030833663940429686, 0.03112550354003906, 0.031139839172363282, 0.03407462310791016, 0.031285247802734374, 0.03142348861694336, 0.030948352813720704, 0.030793727874755858, 0.031355903625488284, 0.030659584045410155, 0.030748672485351562, 0.03102720069885254, 0.031903743743896484, 0.0313118724822998, 0.031243263244628908, 0.03079475212097168, 0.0309616641998291, 0.030648319244384766, 0.030700544357299804, 0.03083776092529297, 0.0307589111328125, 0.03058278465270996, 0.031108095169067384, 0.03103027153015137, 0.03078246307373047, 0.030653440475463867, 0.030825471878051756, 0.030948352813720704, 0.03160883140563965, 0.03098111915588379, 0.031422464370727536, 0.031476736068725586, 0.03138764762878418, 0.03118387222290039, 0.031648767471313476, 0.03129958343505859, 0.03153408050537109, 0.030819328308105468, 0.030551040649414062, 0.03057459259033203, 0.03141222381591797, 0.03105075263977051, 0.03117977523803711, 0.031440895080566404, 0.030825471878051756, 0.03061862373352051, 0.030766080856323243, 0.03115007972717285, 0.03120639991760254, 0.030846975326538087, 0.030869504928588868, 0.031205375671386718, 0.030971904754638672, 0.031121408462524414, 0.0325560302734375, 0.030975040435791017, 0.03069740867614746, 0.031021055221557618, 0.030839807510375978, 0.03077120018005371, 0.030887935638427736, 0.031838207244873046, 0.03128217506408691, 0.030966783523559572, 0.031243263244628908, 0.031122432708740235, 0.030904319763183592, 0.031079423904418944, 0.030841856002807616, 0.03060633659362793, 0.031054847717285155, 0.031101951599121092, 0.031121408462524414, 0.030870527267456056, 0.031337472915649416, 0.0317829761505127, 0.03111827278137207, 0.031144960403442383, 0.0309268798828125, 0.03124015998840332, 0.031114240646362305, 0.030793727874755858, 0.030741504669189453, 0.03120128059387207, 0.03076300811767578, 0.03119923210144043, 0.031129600524902344, 0.031253503799438476, 0.03076812744140625, 0.030939136505126953, 0.031205375671386718, 0.030874624252319335, 0.03126169586181641, 0.031063039779663085, 0.031244287490844725, 0.031383583068847656, 0.03124323272705078, 0.031108095169067384, 0.03101900863647461, 0.031265792846679685, 0.031409151077270506, 0.0311592960357666, 0.031082496643066407, 0.031046655654907225, 0.03129855918884277, 0.031090688705444337, 0.030827520370483398, 0.030801919937133788, 0.030649343490600587, 0.030905344009399413, 0.030737407684326173, 0.030757888793945313, 0.03062272071838379, 0.030703615188598633, 0.031077375411987306, 0.031065088272094726, 0.031065088272094726, 0.031080448150634765, 0.03098521614074707, 0.03082444763183594, 0.030727167129516602, 0.031045631408691408, 0.03159040069580078, 0.031160320281982422, 0.03115110397338867, 0.031237119674682616, 0.031247360229492187, 0.031082496643066407, 0.03061759948730469, 0.030875648498535156, 0.031031295776367186, 0.030809087753295897, 0.031321088790893556, 0.031121408462524414, 0.03119001579284668, 0.030749696731567383, 0.03082342338562012, 0.031088640213012695, 0.03132825660705566, 0.032778240203857424, 0.03150028800964356, 0.031160320281982422, 0.03116339111328125, 0.031281152725219724, 0.03100057601928711, 0.031094783782958983, 0.032456703186035156, 0.03114905548095703, 0.03176959991455078, 0.03134566307067871, 0.0324659194946289, 0.03155353546142578, 0.030680063247680665, 0.030910463333129884, 0.030612480163574218, 0.030962688446044922, 0.030694400787353516, 0.030862335205078126, 0.031129600524902344, 0.030517248153686522, 0.03102822494506836, 0.030835712432861328, 0.031049728393554688, 0.030737407684326173, 0.03079270362854004, 0.030917631149291993, 0.0307906551361084, 0.030616575241088868, 0.030701568603515625, 0.03080806350708008, 0.03096883201599121, 0.030996480941772462, 0.030887935638427736, 0.030628864288330077, 0.03143270492553711, 0.03099545669555664, 0.031052799224853517, 0.031156223297119142, 0.031117311477661135, 0.03081625556945801, 0.030845951080322266, 0.03075071907043457, 0.030742528915405274, 0.031783935546875, 0.030980096817016602, 0.03160780715942383, 0.03143475151062012, 0.03156480026245117, 0.031086591720581053, 0.031106048583984375, 0.031194112777709962, 0.03116646385192871, 0.03150028800964356, 0.031100927352905275, 0.031140863418579103, 0.03136102485656738, 0.03241164779663086, 0.03138559913635254, 0.0310118408203125, 0.031045631408691408, 0.031145984649658204, 0.030851072311401367, 0.03120639991760254, 0.031070207595825194, 0.03138457679748535, 0.031032320022583007, 0.031344831466674807, 0.031109952926635744, 0.030735359191894532, 0.030735359191894532, 0.03117670440673828, 0.03115110397338867, 0.03123302459716797, 0.03144499206542969, 0.03128934478759766, 0.03177881622314453, 0.031714303970336914, 0.031072256088256835, 0.031197183609008788, 0.030996480941772462, 0.031541248321533204, 0.03126272010803223, 0.031077375411987306, 0.032007167816162106, 0.03146956825256347, 0.03102003288269043, 0.030748672485351562, 0.031095808029174804, 0.031049728393554688, 0.031120384216308594, 0.031029247283935548, 0.031085567474365236, 0.03139481544494629, 0.030555135726928712, 0.030926847457885744, 0.03097907257080078, 0.03058278465270996, 0.030924800872802735, 0.030900224685668946, 0.030681087493896485, 0.030740480422973632, 0.030707712173461913, 0.03062272071838379, 0.031062015533447264, 0.031185920715332032, 0.030686208724975586, 0.031120384216308594, 0.031269887924194335, 0.030632959365844727, 0.030744575500488282, 0.030621696472167968, 0.031040512084960937, 0.03122585678100586, 0.030741504669189453, 0.03058687973022461, 0.030847999572753908, 0.030652416229248046, 0.030864383697509764, 0.030893056869506837, 0.03101388740539551, 0.033242111206054685, 0.03198566436767578, 0.03147980880737305, 0.031202304840087892, 0.030665727615356447, 0.030660608291625976, 0.03126681518554687, 0.03132825660705566, 0.03118694305419922, 0.031244287490844725, 0.031078399658203124, 0.031037439346313478, 0.03139891242980957, 0.03157606315612793]",tokens/s,32.1252926156553,,,,,,gemma,MB,4745.342976,6422.003712,0.0,5792.333824,5786.358272,s,1,11.3428896484375,11.3428896484375,0.0,11.3428896484375,11.3428896484375,11.3428896484375,11.3428896484375,[11.3428896484375],,kWh,5.1149725233335e-05,2.8018309353047524e-05,7.386922576200355e-05,0.00015303726034838607, @@ -11897,48 +11702,33 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,qwen2,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.4.1+cu124,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-70b-hf,meta-llama/Llama-2-70b-hf,cuda,0,42,,,True,,,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.261056,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,0.12.0,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 153, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 82, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 258, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 172, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3904, in from_pretrained - dispatch_model(model, **device_map_kwargs) - File ""/usr/local/lib/python3.10/dist-packages/accelerate/big_modeling.py"", line 489, in dispatch_model - model.to(device) - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 2796, in to - return super().to(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1173, in to - return self._apply(convert) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 779, in _apply - module._apply(fn) - [Previous line repeated 2 more times] - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 853, in _apply - self._buffers[key] = fn(buf) - File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1159, in convert - return t.to( -torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. GPU + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 173, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,llama,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,4119.26528,5287.44448,0.0,4640.997376,4330.668544,s,10,3.1964943542480473,0.31964943542480473,0.0011752270876245364,0.31938330078125,0.3210399475097656,0.3212041931152344,0.3213355895996094,"[0.31863671875, 0.32136843872070314, 0.3186257019042969, 0.3181160888671875, 0.3190706481933594, 0.3183872985839844, 0.31969595336914064, 0.3206474304199219, 0.320942626953125, 0.3210034484863281]",tokens/s,800.8773726122291,kWh,3.759320359677076e-06,2.059510090694167e-06,1.8734077487252143e-05,2.4552907937623384e-05,tokens/kWh,10426463.56392357,MB,4119.26528,5287.44448,0.0,4640.997376,4515.667456,s,10,175.22104296875003,17.522104296875003,0.0010294162792460536,17.5222197265625,17.523270312500003,17.5235814453125,17.523830351562502,"[17.52235546875, 17.520908203125, 17.52276953125, 17.520298828125, 17.52115625, 17.522083984375, 17.523892578125, 17.52198046875, 17.523201171875, 17.522396484375]",tokens/s,3.5954585666537664,kWh,0.0002069021587094499,0.00011339795251749486,0.0010054651863159417,0.0013257652975428866,tokens/kWh,47519.72322458684,,s,629,177.82021145629867,0.28270303888123816,0.03804919348962998,0.27811224365234377,0.278291455078125,0.2783965270996094,0.5975640185546875,"[0.27857000732421877, 0.2780395202636719, 0.27797503662109374, 0.2779525146484375, 0.27803750610351563, 0.27811431884765625, 0.2782115783691406, 0.2780088195800781, 0.2779576416015625, 0.2779299926757812, 0.278034423828125, 0.2779525146484375, 0.27793820190429686, 0.2782955322265625, 0.2780692443847656, 0.27795455932617186, 0.27805389404296876, 0.2780364685058594, 0.2780498046875, 0.27802215576171874, 0.2781245422363281, 0.2780498046875, 0.27813479614257813, 0.27802215576171874, 0.2781532287597656, 0.2781327209472656, 0.27817880249023436, 0.27801803588867186, 0.2782863464355469, 0.278108154296875, 0.27824844360351564, 0.2780682373046875, 0.27830169677734373, 0.2781562805175781, 0.27808154296875, 0.278002685546875, 0.27811328125, 0.27820440673828123, 0.2781337585449219, 0.2781665344238281, 0.27812457275390623, 0.27807025146484377, 0.27822796630859375, 0.2779862976074219, 0.2780989379882812, 0.278108154296875, 0.27817166137695315, 0.27808154296875, 0.27818496704101564, 0.27802520751953125, 0.2784215087890625, 0.27810406494140627, 0.2782023620605469, 0.27815115356445314, 0.2781747131347656, 0.2784276428222656, 0.278086669921875, 0.2781051025390625, 0.27823614501953126, 0.2781521911621094, 0.27816244506835935, 0.2781388854980469, 0.5999144897460937, 0.27822695922851565, 0.2780282897949219, 0.27790643310546875, 0.27799859619140627, 0.27829043579101564, 0.27803237915039064, 0.27796685791015624, 0.2778839111328125, 0.27787982177734377, 0.2779361267089844, 0.27789108276367186, 0.27787161254882814, 0.2779381713867187, 0.27786648559570315, 0.2779586486816406, 0.2779320373535156, 0.2780569458007813, 0.27793508911132814, 0.27819827270507813, 0.27794125366210937, 0.2780354614257812, 0.2785587158203125, 0.27818292236328124, 0.2780968933105469, 0.27815115356445314, 0.2779381713867187, 0.27799859619140627, 0.2783057861328125, 0.2781767578125, 0.27804568481445313, 0.2780692443847656, 0.27811224365234377, 0.2780917663574219, 0.2780559387207031, 0.27807333374023435, 0.27801803588867186, 0.27811532592773436, 0.27813067626953125, 0.27810406494140627, 0.2780610656738281, 0.27840408325195315, 0.27811328125, 0.278408203125, 0.27807232666015624, 0.2780897216796875, 0.27822796630859375, 0.27817062377929686, 0.27822900390625, 0.2781808776855469, 0.27825048828125, 0.27820645141601563, 0.2780579833984375, 0.2781112365722656, 0.2781214599609375, 0.27808767700195314, 0.2780426330566406, 0.27833139038085936, 0.27853927612304685, 0.27829556274414063, 0.2780999755859375, 0.27811224365234377, 0.2782023620605469, 0.5972490234375, 0.2780139465332031, 0.27791461181640625, 0.2779453430175781, 0.27807232666015624, 0.27789926147460936, 0.2779279479980469, 0.27835903930664063, 0.27785830688476565, 0.277928955078125, 0.2780498046875, 0.2779422607421875, 0.2780159912109375, 0.27806515502929685, 0.27808767700195314, 0.27808563232421873, 0.2780149841308594, 0.27797503662109374, 0.2780211181640625, 0.2781542358398438, 0.2780426330566406, 0.2781890563964844, 0.2781552734375, 0.2781767578125, 0.2779658203125, 0.27812249755859375, 0.2780405883789063, 0.2780845947265625, 0.27801702880859375, 0.2782105712890625, 0.27835595703125, 0.27822079467773436, 0.2781368408203125, 0.27814093017578123, 0.27808255004882815, 0.27803955078125, 0.27819931030273437, 0.2780928039550781, 0.2780692443847656, 0.2780948486328125, 0.27811019897460937, 0.27807537841796875, 0.2781521911621094, 0.2783621215820313, 0.27800473022460936, 0.2781562805175781, 0.27817984008789065, 0.27834368896484374, 0.2781603698730469, 0.2783662109375, 0.2781337585449219, 0.27818496704101564, 0.2783641662597656, 0.278192138671875, 0.2781552734375, 0.2782423095703125, 0.27822695922851565, 0.2782248840332031, 0.27817166137695315, 0.2782586975097656, 0.27807745361328123, 0.27811328125, 0.2781695861816406, 0.5976371459960937, 0.27793508911132814, 0.27821261596679686, 0.277923828125, 0.27800985717773435, 0.27803237915039064, 0.27797607421875, 0.27789312744140626, 0.27804672241210937, 0.27797708129882814, 0.27794842529296876, 0.2779309997558594, 0.27805389404296876, 0.27793508911132814, 0.27791156005859374, 0.27805184936523436, 0.2780426330566406, 0.2781562805175781, 0.2780354614257812, 0.27821261596679686, 0.2779893798828125, 0.27812045288085935, 0.2779381713867187, 0.27817166137695315, 0.2780364685058594, 0.2781839294433594, 0.2779955139160156, 0.2781552734375, 0.2780487670898438, 0.27814501953125, 0.27815731811523436, 0.27817062377929686, 0.2780508117675781, 0.2780928039550781, 0.2782791748046875, 0.2779914245605469, 0.277855224609375, 0.2780334167480469, 0.27817166137695315, 0.278255615234375, 0.2780928039550781, 0.2780917663574219, 0.2781839294433594, 0.27820748901367187, 0.27814093017578123, 0.2783354797363281, 0.2781552734375, 0.2781767578125, 0.27807537841796875, 0.2782115783691406, 0.27816448974609376, 0.27819314575195314, 0.278001708984375, 0.27818389892578127, 0.27822796630859375, 0.27817880249023436, 0.27804672241210937, 0.2781532287597656, 0.2782310485839844, 0.27818597412109375, 0.27811431884765625, 0.27812353515625, 0.27820440673828123, 0.5976770629882813, 0.27793508911132814, 0.2780139465332031, 0.27788491821289063, 0.2779473876953125, 0.2779207763671875, 0.27793508911132814, 0.27799755859375, 0.27789208984375, 0.277928955078125, 0.27800167846679685, 0.2780415954589844, 0.2780139465332031, 0.2780149841308594, 0.2780282897949219, 0.27804266357421875, 0.2780118713378906, 0.27807537841796875, 0.27801907348632815, 0.2781399230957031, 0.2780303039550781, 0.2780282897949219, 0.27832525634765626, 0.2781941833496094, 0.27800473022460936, 0.27825665283203127, 0.27798220825195313, 0.27805184936523436, 0.2780968933105469, 0.27815936279296877, 0.27810919189453126, 0.27815936279296877, 0.27806515502929685, 0.27813580322265624, 0.27810714721679686, 0.2780652160644531, 0.27816851806640625, 0.2780989379882812, 0.27828121948242185, 0.2780968933105469, 0.2781665344238281, 0.2781890563964844, 0.27813784790039064, 0.2782095336914063, 0.2779473876953125, 0.2781481018066406, 0.27816448974609376, 0.27832321166992186, 0.2781747131347656, 0.2782843017578125, 0.2781665344238281, 0.27816860961914064, 0.27811734008789063, 0.27814297485351563, 0.2780897216796875, 0.27829656982421874, 0.2781839294433594, 0.27825665283203127, 0.2781562805175781, 0.27817880249023436, 0.2780989379882812, 0.2784952392578125, 0.27819827270507813, 0.5973759765625, 0.2780149841308594, 0.27803955078125, 0.27812353515625, 0.2779873352050781, 0.2780631103515625, 0.2779586486816406, 0.2779525146484375, 0.2778961791992188, 0.2780231628417969, 0.2780201110839844, 0.2779781188964844, 0.2780682373046875, 0.2778746948242187, 0.2778736572265625, 0.2780569458007813, 0.2779473876953125, 0.2781951904296875, 0.2780692443847656, 0.27814093017578123, 0.2780139465332031, 0.2780569458007813, 0.2779873352050781, 0.27845733642578124, 0.2780528564453125, 0.27811532592773436, 0.2780006408691406, 0.2780979309082031, 0.27799859619140627, 0.2780334167480469, 0.2781214599609375, 0.27804672241210937, 0.2782586975097656, 0.27856692504882813, 0.27808563232421873, 0.27827813720703126, 0.2783825988769531, 0.27811328125, 0.27818292236328124, 0.278108154296875, 0.27812045288085935, 0.27821261596679686, 0.2781440124511719, 0.27821978759765625, 0.27803851318359374, 0.2780477294921875, 0.2781481018066406, 0.27811532592773436, 0.27803955078125, 0.27829656982421874, 0.2781245422363281, 0.27817779541015625, 0.2781583251953125, 0.2784716796875, 0.278086669921875, 0.27816448974609376, 0.27813580322265624, 0.2781460571289063, 0.278029296875, 0.27843685913085936, 0.27803033447265624, 0.27825048828125, 0.27829556274414063, 0.5983477783203125, 0.2779914245605469, 0.2780149841308594, 0.278034423828125, 0.2779525146484375, 0.27849215698242186, 0.27799038696289063, 0.2781112365722656, 0.277970947265625, 0.2779361267089844, 0.2780006408691406, 0.2781327209472656, 0.27799755859375, 0.2780006408691406, 0.2780211181640625, 0.27807437133789065, 0.27802725219726565, 0.2781521911621094, 0.27789208984375, 0.2782105712890625, 0.27799755859375, 0.27802215576171874, 0.2780477294921875, 0.27806719970703125, 0.27805902099609375, 0.27865087890625, 0.27810714721679686, 0.27815115356445314, 0.27817266845703126, 0.278302734375, 0.27812762451171874, 0.27827301025390627, 0.27820645141601563, 0.27821774291992185, 0.2780845947265625, 0.27819314575195314, 0.2780631103515625, 0.27807232666015624, 0.2781757507324219, 0.278181884765625, 0.2781419372558594, 0.27867239379882813, 0.278108154296875, 0.2783160400390625, 0.27805389404296876, 0.2783057861328125, 0.278150146484375, 0.27816244506835935, 0.2780405883789063, 0.2783088684082031, 0.2781603698730469, 0.2782290344238281, 0.2781531982421875, 0.2781839294433594, 0.2782597045898437, 0.2781552734375, 0.278181884765625, 0.27816448974609376, 0.27823614501953126, 0.27816448974609376, 0.2781317138671875, 0.2785331115722656, 0.278302734375, 0.5992274169921875, 0.2780313720703125, 0.278044677734375, 0.2780078125, 0.27800677490234377, 0.2780559387207031, 0.27793304443359373, 0.27800677490234377, 0.27792178344726565, 0.2780579833984375, 0.27804364013671873, 0.2780334167480469, 0.277970947265625, 0.2784481201171875, 0.278150146484375, 0.2780641174316406, 0.27803851318359374, 0.2780887145996094, 0.27807537841796875, 0.2781368408203125, 0.27809381103515624, 0.2780897216796875, 0.27805184936523436, 0.2781614074707031, 0.27800677490234377, 0.2781163635253906, 0.27799859619140627, 0.27805184936523436, 0.27798220825195313, 0.2782003173828125, 0.27810302734375, 0.27817166137695315, 0.27818496704101564, 0.27818496704101564, 0.2780569458007813, 0.2780477294921875, 0.2779955139160156, 0.27807232666015624, 0.2781552734375, 0.2780528564453125, 0.27810202026367187, 0.27819009399414063, 0.2780139465332031, 0.2782863464355469, 0.2779986267089844, 0.27810198974609374, 0.27818292236328124, 0.27821261596679686, 0.2781337585449219, 0.27874508666992187, 0.27831808471679687, 0.27817779541015625, 0.278245361328125, 0.2782320556640625, 0.27816244506835935, 0.2781317138671875, 0.27815115356445314, 0.2781737060546875, 0.27821875, 0.2782105712890625, 0.27813479614257813, 0.27818801879882815, 0.2782443542480469, 0.5995693969726562, 0.27800167846679685, 0.27810098266601563, 0.2780897216796875, 0.277897216796875, 0.27807949829101564, 0.27800167846679685, 0.2780528564453125, 0.2780139465332031, 0.2779791259765625, 0.27805490112304687, 0.278002685546875, 0.2781245422363281, 0.27810714721679686, 0.27802215576171874, 0.27804672241210937, 0.27800985717773435, 0.27851263427734374, 0.2779924621582031, 0.2781757507324219, 0.27816549682617187, 0.2783897705078125, 0.2783006591796875, 0.27812966918945314, 0.27797503662109374, 0.2782586975097656, 0.2779873352050781, 0.27809588623046877, 0.2780631103515625, 0.2781767578125, 0.27805184936523436, 0.27824331665039065, 0.2781265869140625, 0.27821261596679686, 0.27800985717773435, 0.2781194152832031, 0.27810205078125, 0.2781490783691406, 0.27807437133789065, 0.2781624755859375, 0.2780200500488281, 0.27825048828125, 0.27810513305664064, 0.2783231811523437, 0.2780917663574219, 0.2781695861816406, 0.2781695861816406, 0.2783047790527344, 0.2782218322753906, 0.2782494812011719, 0.2781051025390625, 0.2781695861816406, 0.27809381103515624, 0.278192138671875, 0.27815731811523436, 0.27819009399414063, 0.2781890563964844, 0.2785372009277344, 0.2781839294433594, 0.2781603698730469, 0.27817880249023436, 0.27817166137695315, 0.2783221740722656, 0.5995499267578125, 0.2779862976074219, 0.27796685791015624, 0.2779842529296875, 0.277970947265625, 0.27801803588867186, 0.2779586486816406, 0.27808563232421873, 0.27782144165039063, 0.27793508911132814, 0.27791668701171873, 0.2779381713867187, 0.2780405883789063, 0.27805389404296876, 0.2779924621582031, 0.27810202026367187, 0.27799859619140627, 0.27811328125, 0.27840103149414064, 0.2781890563964844, 0.2779473876953125, 0.27806515502929685, 0.2780078125, 0.2781603698730469, 0.2779299926757812, 0.27811328125, 0.27796685791015624, 0.2782474365234375, 0.27803851318359374, 0.27858328247070313, 0.2781255798339844, 0.2784962463378906, 0.2780979309082031, 0.27818292236328124, 0.2782156677246094, 0.2781317138671875, 0.2780702819824219, 0.2781757507324219, 0.2781265869140625, 0.27816448974609376, 0.2780897216796875, 0.27811532592773436, 0.2780405883789063, 0.27824331665039065, 0.278076416015625, 0.27814093017578123, 0.2782423095703125, 0.278181884765625, 0.27828839111328124, 0.2782494812011719, 0.2781562805175781, 0.27818496704101564, 0.27810714721679686, 0.2782597045898437, 0.2783846435546875, 0.27833447265625, 0.2781685791015625, 0.27818701171875, 0.2782156677246094, 0.2782771301269531, 0.27810714721679686, 0.2782115783691406, 0.2782771301269531]",tokens/s,3.537280688447406,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,r,r,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.293824,Linux,x86_64,Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.3.1,,4.42.4,,0.32.1,,,,1.21.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch