diff --git "a/perf-df-unquantized-32vCPU-C7i.csv" "b/perf-df-unquantized-32vCPU-C7i.csv" --- "a/perf-df-unquantized-32vCPU-C7i.csv" +++ "b/perf-df-unquantized-32vCPU-C7i.csv" @@ -3823,6 +3823,423 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,3374.178304,,,,,s,1,48.558202806001646,48.558202806001646,0.0,48.558202806001646,48.558202806001646,48.558202806001646,48.558202806001646,[48.558202806001646],,kWh,0.0005439857744826251,0.00029649203179826094,0,0.0008404778062808861,,MB,3384.48384,,,,,s,10,4.331008567000026,0.43310085670000265,0.00990644255129294,0.4294304135000857,0.4465573489986127,0.4475254699995276,0.4482999668002594,"[0.42853586899946094, 0.4463422109984094, 0.4484935910004424, 0.44167812400155526, 0.43006467600025644, 0.418377504000091, 0.42879615099991497, 0.44064103699929547, 0.4261737489996449, 0.42190565500095545]",tokens/s,591.0863394512386,kWh,5.013990916695569e-06,2.7327431492861644e-06,0.0,7.746734065981733e-06,tokens/kWh,33046184.084745325,MB,3414.638592,,,,,s,10,27.287819859999217,2.728781985999922,0.0051947481996214954,2.730266948999997,2.733845954201206,2.7349333081013354,2.7358031912214393,"[2.7294153410002764, 2.7336043200011773, 2.7235425189992384, 2.725377329999901, 2.7243706439985544, 2.718889895999382, 2.7333088649993442, 2.7321717260001606, 2.736020662001465, 2.731118556999718]",tokens/s,23.08722364894775,kWh,3.229385477260885e-05,1.7601305703527962e-05,0.0,4.989516047613681e-05,tokens/kWh,1262647.5072693834,,s,630,27.281638503000067,0.043304188100000106,0.0007377116499941429,0.04321132699988084,0.04413209750018723,0.04461441564944835,0.045947955209794604,"[0.04416177300117852, 0.04293827099900227, 0.04191493899998022, 0.04233325300083379, 0.042304630998842185, 0.042840791000344325, 0.04295588800050609, 0.04325220899954729, 0.04388025499974901, 0.043713719000152196, 0.04331308600012562, 0.0435892010009411, 0.04253194400007487, 0.04206018100012443, 0.044620644999668, 0.04324712099878525, 0.04241506600010325, 0.04273395800009894, 0.04239724500075681, 0.04471484299938311, 0.0431137800005672, 0.04227339699900767, 0.04187192800054618, 0.042351213000074495, 0.04231981500015536, 0.04262780200042471, 0.04265072300040629, 0.042445454999324284, 0.04305314100020041, 0.04302354299943545, 0.043944557999566314, 0.04333757099993818, 0.04356555800040951, 0.043478827999933856, 0.043676837000020896, 0.043717540000216104, 0.042886113000349724, 0.043460848999529844, 0.04303748399979668, 0.04408602399962547, 0.04435381500115909, 0.04478721099985705, 0.04662445499889145, 0.044273764000536175, 0.04352596600074321, 0.04516055299973232, 0.04292490999978327, 0.042349662000560784, 0.0426248410003609, 0.04254339899853221, 0.042824778000067454, 0.042949882001266815, 0.0428164659988397, 0.043328712999937125, 0.04340248399967095, 0.04436386700035655, 0.04370287200072198, 0.044127832999947714, 0.04361866199906217, 0.04302339700006996, 0.04353484400053276, 0.043615534999844385, 0.045449377001204994, 0.04514607700002671, 0.04573703800087969, 0.04391718800070521, 0.04361291399982292, 0.04465159899882565, 0.04462472900013381, 0.0432996199997433, 0.04344694900100876, 0.043675734999851556, 0.04581287700057146, 0.0439372419987194, 0.04384454200044274, 0.04301449400009005, 0.04255349000050046, 0.04251212199960719, 0.04221132699967711, 0.04246826000053261, 0.04232697400038887, 0.04263593399991805, 0.042743938000057824, 0.042565420999380876, 0.04284481899958337, 0.042321364000599715, 0.04307257700020273, 0.04410800199912046, 0.042611279999619, 0.0428870880004979, 0.04206361599972297, 0.042540720000033616, 0.043583022999882814, 0.04360237699984282, 0.04267888800131914, 0.04360968599939952, 0.04478914799983613, 0.04317173600065871, 0.042830794998735655, 0.04347746600069513, 0.043840258000273025, 0.04303428700040968, 0.04338576899863256, 0.04370026200012944, 0.04227550500036159, 0.04381419099991035, 0.04277913500118302, 0.04306541599908087, 0.04333260299972608, 0.043214526000156184, 0.04221078599948669, 0.04265837200000533, 0.04251930900136358, 0.042932971999107394, 0.04430741800024407, 0.043689623000318534, 0.04376369999954477, 0.04481940299956477, 0.0434911580014159, 0.04360456699942006, 0.04354567200061865, 0.04381258200010052, 0.04347298499851604, 0.0435189090003405, 0.0439353250003478, 0.04331006599932152, 0.04437678800059075, 0.04278334199989331, 0.04320003700013331, 0.0426211929989222, 0.043955064000329, 0.04327726100018481, 0.04287761299929116, 0.04341099600060261, 0.04231475800042972, 0.04276901299999736, 0.042658250999011216, 0.04319885000040813, 0.04278768100084562, 0.04263231299955805, 0.04228644599970721, 0.042675442000472685, 0.04246632299873454, 0.04316929700144101, 0.042734700999062625, 0.04333794800004398, 0.043211001000599936, 0.04245703699962178, 0.04275375399993209, 0.043583037000644254, 0.042552281000098446, 0.042963971998688066, 0.043750352000643034, 0.043508519000170054, 0.042526318999080104, 0.04315482499987411, 0.042546100999970804, 0.04314206700109935, 0.04460680199917988, 0.04484711000077368, 0.04360839999935706, 0.04244576500059338, 0.04307962500024587, 0.04342168800030777, 0.04331362799894123, 0.044039264001185074, 0.04324254699895391, 0.04374991599979694, 0.04353621500013105, 0.04329383699950995, 0.04337770200072555, 0.04359197300072992, 0.04360329500013904, 0.04370031500002369, 0.04435166199982632, 0.04403144199932285, 0.04357414200057974, 0.04318944499937061, 0.04350843700012774, 0.04321554499983904, 0.042986210999515606, 0.04262775999995938, 0.04269995499998913, 0.043804740000268794, 0.044274369000049774, 0.04272781399959058, 0.04269193800064386, 0.043049358999269316, 0.043072013000710285, 0.044355549000101746, 0.04304859199874045, 0.04305219600064447, 0.04299785799958045, 0.043720879999455065, 0.042585556000631186, 0.042329418000008445, 0.0421985869998025, 0.04252694299975701, 0.042279127999790944, 0.04276245900109643, 0.04252247299882583, 0.04250553000019863, 0.04339756999979727, 0.04234036900015781, 0.04270910600098432, 0.04335115399953793, 0.043827376000990625, 0.0442679349998798, 0.04321165299916174, 0.04300311299994064, 0.04286128599960648, 0.04285193500072637, 0.042481464999582386, 0.04243193000002066, 0.04379824100033147, 0.042727058000309626, 0.04342568799984292, 0.04325469399918802, 0.0435193250013981, 0.043403908999607665, 0.0455099569990125, 0.04282350300127291, 0.04301620099977299, 0.042995734000214725, 0.043300260000251, 0.04308009199849039, 0.04333306100124901, 0.04241301299953193, 0.04257179200067185, 0.042515225999522954, 0.042483764000280644, 0.04258985100022983, 0.04262540799936687, 0.042172585999651346, 0.043236493000222254, 0.04307595600039349, 0.04353296899898851, 0.043702138000298874, 0.04391195500102185, 0.04408113099998445, 0.04268040699935227, 0.04318509400036419, 0.046834441000100924, 0.04316935499991814, 0.043491786000231514, 0.0463939040000696, 0.04415158599840652, 0.04320459000155097, 0.0440030489990022, 0.04392800999994506, 0.04289295200032939, 0.044099706999986665, 0.043731953999667894, 0.042274911000276916, 0.04265276200021617, 0.04269153899986122, 0.0422130949991697, 0.042442293000931386, 0.04272397699969588, 0.04373688699888589, 0.042888929001492215, 0.044274431998928776, 0.043252039000435616, 0.043341037000573124, 0.043027119998441776, 0.043801530000564526, 0.04258706899963727, 0.042920849000438466, 0.04260046300078102, 0.04488172799938184, 0.042859794999458245, 0.04247765000036452, 0.04369579100057308, 0.04352309699970647, 0.04391079599918157, 0.044092727001043386, 0.043920091000472894, 0.042962019999322365, 0.04391162799947779, 0.04436802900090697, 0.04369795699858514, 0.04335462000017287, 0.042875713001194526, 0.042731313998956466, 0.042298705000575865, 0.04351408399998036, 0.04352323099919886, 0.042842266000661766, 0.044206536000274355, 0.045972793999681016, 0.04514331700011098, 0.04303440200055775, 0.04354392599998391, 0.04292114300005778, 0.04234207700028492, 0.042334665999078425, 0.04363179800020589, 0.04386393999993743, 0.04294956500052649, 0.042863941000177874, 0.042496575999393826, 0.04317336100029934, 0.04302784199899179, 0.04310576100033359, 0.04237273500075389, 0.04238337300012063, 0.04229673299960268, 0.043075933999716653, 0.04357893099950161, 0.042351723999672686, 0.04283068900076614, 0.043517756999790436, 0.04373983900040912, 0.043043961999501335, 0.04339724500096054, 0.04399361600007978, 0.042237206998834154, 0.04255446800016216, 0.042415148000145564, 0.042106898999918485, 0.04197175999979663, 0.042146477000642335, 0.04248163800002658, 0.04260643199995684, 0.04319720499916002, 0.04300566900019476, 0.04336452300049132, 0.04276946599929943, 0.04361313200024597, 0.04305056399971363, 0.04319634600142308, 0.04290784799923131, 0.042946280000251136, 0.0430994020007347, 0.043301044999680016, 0.04218045599918696, 0.04363940800067212, 0.04386694599998009, 0.04413394699986384, 0.04317296000044735, 0.04352030299924081, 0.042982442000720766, 0.043368896998799755, 0.04322570700060169, 0.043478937999680056, 0.043115517999467556, 0.04333596200012835, 0.043653166001604404, 0.043111012999361265, 0.04271110099944053, 0.04349252000065462, 0.042963484998836066, 0.04335096800059546, 0.04406833000030019, 0.04379347000030975, 0.043659560000378406, 0.043409572999735246, 0.043094028000268736, 0.043112462999488343, 0.04393090399935318, 0.043337044000509195, 0.04369113699976879, 0.04346224400069332, 0.04299781299960159, 0.043026309000197216, 0.043230008999671554, 0.04349848300080339, 0.04273180199925264, 0.04349585499949171, 0.042959472000802634, 0.04260440399957588, 0.04302624400042987, 0.04290373100047873, 0.04329064399826166, 0.043036543000198435, 0.04251799000121537, 0.043316994999258895, 0.04383480999968015, 0.04406350899989775, 0.04236494799988577, 0.04228003800017177, 0.04336432299896842, 0.04211694900004659, 0.04231207600059861, 0.042533891999482876, 0.04617845699976897, 0.04293390000020736, 0.04273900400039565, 0.04388533000019379, 0.04340722400047525, 0.04304757999852882, 0.04260280100061209, 0.04309693600043829, 0.042232898998918245, 0.04309219200149528, 0.042787079999470734, 0.042483526000069105, 0.04250833900005091, 0.042413134999151225, 0.042859166000198456, 0.042722931000753306, 0.04374690999975428, 0.04361099599918816, 0.04294481300166808, 0.043115745998875354, 0.042831392000152846, 0.04259319599987066, 0.04311804800090613, 0.04341352099982032, 0.043606297998849186, 0.04353519500000402, 0.04399639500115882, 0.043765619999248884, 0.04335770000034245, 0.043548321000344004, 0.044201052998687373, 0.044326491000902024, 0.04320651899979566, 0.0427416649999941, 0.04302681599983771, 0.043464819000291754, 0.04343589299969608, 0.04524716299965803, 0.04570622500068566, 0.04406071300036274, 0.04365910799970152, 0.04365505699934147, 0.04475116400135448, 0.044919219999428606, 0.04325778699967486, 0.04378889600047842, 0.04395545799889078, 0.04334306300006574, 0.04373606100125471, 0.04300947199953953, 0.043079417999251746, 0.042905070000415435, 0.04359068300072977, 0.04400122599872702, 0.043330907999916235, 0.0431155670012231, 0.046500398000716814, 0.04366905700044299, 0.04332961900036025, 0.043027873000028194, 0.043106274999445304, 0.04274828699999489, 0.04333147699981055, 0.043770497000878095, 0.04353981699932774, 0.04269354200005182, 0.044526075000248966, 0.043714290999560035, 0.04363060899959237, 0.04305801600094128, 0.04329592099929869, 0.04314736800006358, 0.043470406000778894, 0.043082909998702235, 0.04295325900056923, 0.04289629199956835, 0.043122377999679884, 0.04263955500027805, 0.04302733700023964, 0.042717108999568154, 0.04273921599997266, 0.043140485000549234, 0.04332080799940741, 0.04431644200121809, 0.043746411000029184, 0.0437685569995665, 0.043807689000459504, 0.04258871499951056, 0.043018874999688705, 0.043039537000368, 0.0441216519993759, 0.04412003600009484, 0.042975499000021955, 0.04346974700092687, 0.04299408999941079, 0.042971109000063734, 0.04287311500047508, 0.04374714000005042, 0.043270745998597704, 0.04383566700016672, 0.04435522600033437, 0.04406314700099756, 0.04361106999931508, 0.04329777299972193, 0.043092113999591675, 0.04351245700127038, 0.04247025099903112, 0.043232137000813964, 0.04298779199962155, 0.04319059100089362, 0.04277136099881318, 0.042736233001051005, 0.04317128099864931, 0.04346384200107423, 0.04321935400002985, 0.04344767399925331, 0.04261952799970459, 0.044472789000792545, 0.04300949599928572, 0.044282075999944936, 0.043132026999956, 0.04272036100155674, 0.04277269199883449, 0.0430096260006394, 0.042815265000172076, 0.04285401099878072, 0.0439680400013458, 0.0437354059995414, 0.04388058599943179, 0.04325377399982244, 0.043652351001583156, 0.04282334799972887, 0.042640853998818784, 0.042787639000380295, 0.04330945900073857, 0.04315972699987469, 0.0434885739996389, 0.042990886000552564, 0.04356805299903499, 0.04296078799961833, 0.0430744670011336, 0.0426882880001358, 0.04343066299952625, 0.04337780700006988, 0.0436222070002259, 0.04311276800035557, 0.04354228400006832, 0.0443134299985104, 0.04419936700105609, 0.04562688199985132, 0.04365052100001776, 0.0432658209992951, 0.04367689300124766, 0.04310673999862047, 0.04305444200144848, 0.04286776599838049, 0.04372747500019614, 0.043904561000090325, 0.0431907490001322, 0.043366444000639603, 0.04309715899944422, 0.04304480300015712, 0.04251268500047445, 0.04288002300017979, 0.043917883000176516, 0.04386997099936707, 0.04369294800017087, 0.04417336000005889, 0.045352242999797454, 0.043494309999005054, 0.04262136100078351, 0.0435603900004935, 0.04291399899921089, 0.04321259800053667, 0.04419922899978701, 0.04413712499990652, 0.04340205500011507, 0.04320135099987965, 0.04314947100101563, 0.04336852699998417, 0.04332018099921697, 0.043510389999937615, 0.04613532199982728, 0.04393558599986136, 0.04235577100007504, 0.04372113899989927, 0.043092003999845474, 0.043487883000125294, 0.043555267999181524, 0.04397194900047907, 0.0432306040002004, 0.042510073999437736, 0.042940569001075346, 0.04271422199963126, 0.042146437999690534, 0.04252099099903717, 0.04311171200060926, 0.04313538899987179, 0.043254592001176206, 0.04292167999847152, 0.0437774130004982, 0.045318811000470305, 0.043311244999131304, 0.0423541700001806, 0.04296251900086645, 0.04257087199948728, 0.04309912099961366, 0.04334200500125007, 0.043387246998463525, 0.04417981799997506, 0.04314010200141638, 0.04298168899913435, 0.042805377999684424, 0.04247605300042778, 0.042361324000012246, 0.04353370799981349, 0.04263829900082783, 0.04299487299977045, 0.04246256799888215, 0.04424470300000394, 0.04456761000074039, 0.04324733199973707, 0.04286875400066492, 0.04278965000048629, 0.0428453360000276, 0.043231593999735196, 0.04326948000016273, 0.04259272499984945, 0.043396365999797126, 0.04284746899975289, 0.04320732300038799, 0.04312784899957478, 0.04339730299943767, 0.043753291000030003, 0.043993961000523996, 0.04326055199999246, 0.043329333999281516, 0.04413189200022316, 0.04588714300007268, 0.04537468700073077, 0.043429497000033734, 0.042881346998910885, 0.043544285001189564, 0.043452906000311486, 0.04344042999946396]",tokens/s,23.09245465336406,, +float16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,12674.56,,,,,s,1,193.0935409169997,193.0935409169997,0.0,193.0935409169997,193.0935409169997,193.0935409169997,193.0935409169997,[193.0935409169997],,kWh,0.0022501933176826473,0.0012264415858629387,0,0.0034766349035455857,,MB,12922.310656,,,,,s,10,14.798830034002094,1.4798830034002095,0.006145138659349014,1.47856284449972,1.4873673155992946,1.4883166158003405,1.489076055961177,"[1.4892659160013864, 1.482803025000976, 1.473320123999656, 1.474013464001473, 1.4703050719999737, 1.4784753349995299, 1.4870313060000626, 1.4871563599990623, 1.4786503539999103, 1.477809078000064]",tokens/s,172.98664787135817,kWh,1.7443648766249885e-05,9.507285540727933e-06,0.0,2.6950934306977818e-05,tokens/kWh,9498743.052248081,MB,13045.399552,,,,,s,10,104.70637704099863,10.470637704099863,0.0548586507269253,10.475131417500052,10.536161597298996,10.540443704149402,10.543869389629727,"[10.535210017998907, 10.544725810999807, 10.387487480000345, 10.401851393999095, 10.530395326000871, 10.494245481999315, 10.483293274999596, 10.466969560000507, 10.405002552000951, 10.457196142999237]",tokens/s,6.016825505798166,kWh,0.00012338343510020838,6.724878534864414e-05,0.0,0.00019063222044885252,tokens/kWh,330479.2854621508,,s,630,104.70143843699589,0.166192759423803,0.0030683421486360873,0.16594851250010834,0.17030235550082579,0.17173697530024584,0.17406530013031443,"[0.16921652299970447, 0.1681069039987051, 0.161090346000492, 0.16292985499967472, 0.16223504500158015, 0.16406782399826625, 0.1659714850011369, 0.16362970499903895, 0.16271438600051624, 0.1634668460010289, 0.16669810799976403, 0.16251985399867408, 0.1620581340011995, 0.1631260909998673, 0.16653610399953322, 0.1652375400008168, 0.16577656799927354, 0.16799629900015134, 0.16466771900013555, 0.16677618799985794, 0.16926093400070386, 0.16609617499852902, 0.16811162900012278, 0.16362483500051894, 0.16551430099934805, 0.16472293900005752, 0.16518066900061967, 0.16879412299931573, 0.16711313800078642, 0.16506578900043678, 0.1679061439990619, 0.1688221149997844, 0.171088533999864, 0.1709098540013656, 0.17313837599976978, 0.16932194699984393, 0.1688535609991959, 0.16805462299998908, 0.1715062900002522, 0.1694026400000439, 0.17149577000054705, 0.16859459099941887, 0.165842395999789, 0.1655927560004784, 0.16976845899989712, 0.17142013299962855, 0.16911453099965001, 0.17117251900162955, 0.16937506399881386, 0.17085919199962518, 0.1719317540009797, 0.16907379899930675, 0.165298462001374, 0.16618753299917444, 0.16949662799925136, 0.16788357000041287, 0.1670408709996991, 0.1704691909999383, 0.16682871900047758, 0.16861109700039378, 0.16651321099925553, 0.1677742030005902, 0.16700645399942005, 0.16642322999905446, 0.16272493399992527, 0.16514880800059473, 0.16332467999927758, 0.16328686000088055, 0.16375932399932935, 0.16333517800012487, 0.16372309300095367, 0.16238804899876413, 0.16453092099982314, 0.1645060540013219, 0.16403476799860073, 0.16409487800046918, 0.16618428800029506, 0.16447490400059905, 0.16516134799894644, 0.172059304000868, 0.16845406500033278, 0.16614650799965602, 0.16903172999991511, 0.1709333170001628, 0.17503574100010155, 0.17348618299911323, 0.17292151500078035, 0.1733807769996929, 0.17040633400029037, 0.1652215639987844, 0.16975962400101707, 0.16765467699951841, 0.1710855340006674, 0.171756400999584, 0.16602993199921912, 0.16566537800099468, 0.16915702400001464, 0.16968071299925214, 0.1663744970010157, 0.1664943209998455, 0.17193254199992225, 0.1703414410003461, 0.1704373669999768, 0.16776797999955306, 0.1684271260000969, 0.1662336649987992, 0.16863843900137, 0.16732205900007102, 0.16757665799923416, 0.16891635399952065, 0.1646386820011685, 0.16448454700002912, 0.16287811899928784, 0.16785254000023997, 0.17007230700073706, 0.16781274599998142, 0.16767221000009158, 0.1683768559996679, 0.16743674899953476, 0.16643690099954256, 0.1702601640008652, 0.165714480999668, 0.16490348099978291, 0.16692365199924097, 0.16729144500095572, 0.16609450800024206, 0.16562570800124377, 0.16110149499945692, 0.1625664980001602, 0.16227285700006178, 0.16105255100046634, 0.16282140799921763, 0.16146719500102336, 0.16234990799966909, 0.15996335099953285, 0.16238161599903833, 0.16543039200041676, 0.16400641499967605, 0.160420617001364, 0.16159763599898724, 0.16239950700037298, 0.16142084800048906, 0.1644602449996455, 0.1608336009994673, 0.16217184100059967, 0.16418656499990902, 0.16251962000023923, 0.16226504699989164, 0.16176115499911248, 0.16220109599998978, 0.16223446100048022, 0.16619870599970454, 0.1661999990010372, 0.16359598799863306, 0.16669007600103214, 0.16592978200060315, 0.16511874599927978, 0.16709136399913405, 0.16425331600112258, 0.16519128700019792, 0.16332635700018727, 0.16469476299971575, 0.1684696209995309, 0.16606620700076746, 0.1665905499994551, 0.16425141999934567, 0.16536354500021844, 0.16525090600043768, 0.16955788099949132, 0.1677855500001897, 0.16962039499958337, 0.1753271739999036, 0.16754453100111277, 0.1676344200004678, 0.16846250599883206, 0.1649640060004458, 0.16700455799946212, 0.16677621999951953, 0.1654585300002509, 0.16578821100119967, 0.16616338199855818, 0.16512507400148024, 0.16721162700014247, 0.1673664350000763, 0.16671960899839178, 0.16741325900147785, 0.16457784900012484, 0.16591029899973364, 0.1668405669988715, 0.1640367879990663, 0.16223330600041663, 0.16154723900035606, 0.1634618119987863, 0.15921041700130445, 0.1612153349997243, 0.16247137800019118, 0.16150072799973714, 0.1628006649989402, 0.16234768700087443, 0.1615224630004377, 0.1625756849989557, 0.1607944999996107, 0.16104719600116368, 0.1641811030003737, 0.16328047400020296, 0.16516873899854545, 0.16146693700102333, 0.1627159309991839, 0.1631725910010573, 0.16297048999877006, 0.16410001999975066, 0.16300138700171374, 0.16325480199884623, 0.1637449000008928, 0.1632913590001408, 0.1637295799991989, 0.1642108649994043, 0.16407297900150297, 0.16360714099937468, 0.165980600999319, 0.16774479299965606, 0.1700228400004562, 0.16606735700042918, 0.16795451799953298, 0.16707242600023164, 0.16656945099930454, 0.16453734700007772, 0.1682437760009634, 0.16706675100067514, 0.16814504699868849, 0.16522660900045594, 0.16676643500068167, 0.1682850809993397, 0.16775253800005885, 0.16803800299931027, 0.16494510600023204, 0.1656470680009079, 0.16449738499977684, 0.16288567099945794, 0.1663093650004157, 0.16281265799989342, 0.16539989699958824, 0.17306971200014232, 0.16878499199992802, 0.17009900599987304, 0.17220186899976397, 0.16898197600130516, 0.16737587099851226, 0.17134191900004225, 0.16850355600035982, 0.16491270200094732, 0.16543135999927472, 0.16556481900079234, 0.16413832399848616, 0.1623981080010708, 0.16383165399929567, 0.16328501300085918, 0.16232870899875707, 0.16575492000083614, 0.16489841399925353, 0.16734486700079287, 0.16399451100005535, 0.16365430599944375, 0.1669896269995661, 0.16263499500018952, 0.16440420300023106, 0.16752890899988415, 0.1633393949996389, 0.16473762600071495, 0.16200541900070675, 0.16247336199921847, 0.16276091399959114, 0.1619003440009692, 0.16437444099938148, 0.1666961499995523, 0.16817651300152647, 0.16704153999853588, 0.16714649700043083, 0.16323076400112768, 0.16439775599974382, 0.1661189939986798, 0.16722599100103253, 0.16516185100044822, 0.1691257359998417, 0.17084230299951741, 0.16516107200004626, 0.1651082309999765, 0.16571712999939336, 0.16488043000026664, 0.16375852800047141, 0.16875584099943808, 0.16508113400050206, 0.16596724299961352, 0.16673340200031816, 0.16705072299919266, 0.16763569900103903, 0.16973977300040133, 0.17275200099902577, 0.17400769900086743, 0.16797214999860444, 0.17523370800154225, 0.17309730599845352, 0.17739599500055192, 0.17305594599929464, 0.17281285100034438, 0.17497310799990373, 0.17008371600059036, 0.16753156800041324, 0.16704828000001726, 0.17305838799984485, 0.1714073909988656, 0.17024577900156146, 0.16817619899848069, 0.16926246600087325, 0.17074065599990718, 0.16603822599972773, 0.16581862900056876, 0.16554341599839972, 0.16415698200034967, 0.16474729900073726, 0.17448703199988813, 0.17341365799984487, 0.17357214000003296, 0.16862557499916875, 0.16548755200165033, 0.16372074499849987, 0.16413176200148882, 0.17000890499912202, 0.16753204000087862, 0.1647477139995317, 0.16292871099904005, 0.16352482000002055, 0.16483141200114915, 0.17043705099968065, 0.16633818300033454, 0.16653455999949074, 0.1663160150001204, 0.16560813800060714, 0.16502726799990342, 0.166296612998849, 0.16351870099970256, 0.1632569860012154, 0.1666788649999944, 0.16523479100033, 0.1658239919997868, 0.16381460999946285, 0.16326873599973624, 0.16741812500004016, 0.1663796850007202, 0.1660332510000444, 0.17174536600032297, 0.1677056939988688, 0.16688306600008218, 0.1723303760009003, 0.167599190999681, 0.16611368499980017, 0.16838833700057876, 0.16517084399856685, 0.16590615400127717, 0.1689145179989282, 0.16662120600085473, 0.1645022800003062, 0.16243121099978453, 0.16346555499876558, 0.1641115490001539, 0.16387674399993557, 0.16960152899991954, 0.1675730980005028, 0.16362867099996947, 0.16553184900112683, 0.1655333549988427, 0.16589053099960438, 0.1635926710005151, 0.16548626199983119, 0.16723606300001848, 0.17408177300058014, 0.1703228129990748, 0.16813301399997727, 0.1658323259998724, 0.16592911599946092, 0.16308700600166048, 0.16499285799909558, 0.1609824099996331, 0.1640213650007354, 0.16268971099998453, 0.16470105199914542, 0.1622771490001469, 0.16624671400131774, 0.16652168599830475, 0.17068000700055563, 0.16377140499935194, 0.16374500100027944, 0.1630334210003639, 0.16694399600055476, 0.169577427999684, 0.1684129210007086, 0.1635911710000073, 0.16458307599896216, 0.1638772730002529, 0.16312026800005697, 0.1680114959999628, 0.16572636099954252, 0.1685212890006369, 0.1649511910000001, 0.16644118700060062, 0.1696490239992272, 0.1731536490005965, 0.16738589399938064, 0.16463619000023755, 0.17128677099935885, 0.16517451199979405, 0.16322588300135976, 0.16375719299867342, 0.16539753700089932, 0.1628460060001089, 0.16412079899964738, 0.16387645900067582, 0.16749921500013443, 0.16690530499909073, 0.16540576899933512, 0.1663803960000223, 0.16542085299988685, 0.16380808900066768, 0.16794941800071683, 0.169773883999369, 0.16696691700053634, 0.163814289999209, 0.16572249099954206, 0.16851203100122802, 0.16400125399923127, 0.16426947100080724, 0.16647397300039302, 0.16881673599891656, 0.16876815400064515, 0.16945484399911948, 0.1701234789998125, 0.16830444600054761, 0.17402496999966388, 0.17030196400082787, 0.17219192599986854, 0.1711744730000646, 0.16662605900091876, 0.16759617999923648, 0.16529658600120456, 0.16884281099919463, 0.1682356469991646, 0.1690920040000492, 0.17105477299992344, 0.1643225830011943, 0.16609024799981853, 0.1688187529998686, 0.1624038179998024, 0.16334828000071866, 0.16202614899884793, 0.16288785400138295, 0.166158871999869, 0.1645616269997845, 0.16578117300014128, 0.16349269399870536, 0.1649601610006357, 0.16424238800027524, 0.16230697999890253, 0.16568751800150494, 0.16173318099936296, 0.1639334110004711, 0.16173963499932142, 0.1619085170004837, 0.16240157699940028, 0.16543751700010034, 0.16647863300022436, 0.16438920099972165, 0.16422079800031497, 0.16610668000066653, 0.1668593089998467, 0.16588804099956178, 0.1627113790000294, 0.16316922400073963, 0.16495149299953482, 0.16893718800019997, 0.16645911399973556, 0.17085499099994195, 0.1710035539999808, 0.1683123990005697, 0.16704601499986893, 0.1643876279995311, 0.16968250699937926, 0.17087259000072663, 0.16625304599983792, 0.16753579699980037, 0.17191319300036412, 0.16965786999935517, 0.16963815900089685, 0.16643619000024046, 0.16588226799831318, 0.1676211530011642, 0.16782485200019437, 0.16956397099966125, 0.1668513320000784, 0.16504905900001177, 0.16301002299951506, 0.16465500599952065, 0.16977582200161123, 0.16502344099899346, 0.1664748929997586, 0.17014404400106287, 0.16235214499829453, 0.16441588700035936, 0.1591064650001499, 0.1595547289998649, 0.1597452969999722, 0.16204214800018235, 0.1629155320006248, 0.16285559899915825, 0.1682286360010039, 0.16249851399879844, 0.1621084480011632, 0.16152716199940187, 0.16173190900008194, 0.16213342400078545, 0.16355409599964332, 0.16667287500058592, 0.16333043899976474, 0.1623949069999071, 0.16519450899977528, 0.16459342899906915, 0.16377785700024106, 0.16423244699944917, 0.1642179300015414, 0.1644091149992164, 0.16515929600063828, 0.1675288839996938, 0.1647068889997172, 0.16449433699926885, 0.16429292200155032, 0.16801606599983643, 0.1656649969991122, 0.16450967600030708, 0.16406307999932324, 0.16729185900112498, 0.1639223559996026, 0.16554147399983776, 0.16618804999961867, 0.165283766000357, 0.16696955500083277, 0.16606621599930804, 0.16388570200069807, 0.16634540499944706, 0.1679861089996848, 0.1686057360002451, 0.16546144600033585, 0.1645174440000119, 0.16613535000033153, 0.16775323000001663, 0.17189978499845893, 0.16978665200076648, 0.16755841899976076, 0.17046546500023396, 0.16704335900067235, 0.16576490299848956, 0.1666901759999746, 0.1640216820014757, 0.16294464400016295, 0.1630339509993064, 0.16626032900057908, 0.16970481999851472, 0.16715631500119343, 0.16809075699893583, 0.16376147699884314, 0.16077366700119455, 0.15948879299867258, 0.1606858650011418, 0.16178609499911545, 0.16114301800007524, 0.16046002000075532, 0.16088847399987571, 0.1611453959994833, 0.16187208399969677, 0.16090464500121016, 0.16486054199958744, 0.1631717580003169, 0.16221002099882753, 0.16358890800074732, 0.16607321700030298, 0.16331344900027034, 0.16325363099895185, 0.1649766480004473, 0.16763891899972805, 0.1648959100002685, 0.16323896699941542, 0.166110612999546, 0.16663977700045507, 0.17111098799978208, 0.17030587900080718, 0.16730255599941302, 0.166607221999584, 0.16572895500030427, 0.16938239299997804, 0.16903089900006307, 0.16636507600014738, 0.16825173600045673, 0.16918992799946864, 0.16994892799993977, 0.168595772000117, 0.16920470200057025, 0.16956398799993622, 0.16560168800060637, 0.16567934099839476, 0.16914552700109198, 0.16635486700033653, 0.16822581799897307, 0.16460021500097355, 0.16792058799910592, 0.1656846099995164, 0.16680219500085514, 0.16864537999936147, 0.16891390800083173, 0.1702857780001068, 0.16548012400016887, 0.16255513499891094, 0.16873675900023954, 0.16582970600029512, 0.1642686010000034, 0.16757167499963543, 0.16734060400085582, 0.16682475699963106, 0.17050966799979506, 0.1681547629996203, 0.16765423200013174, 0.17172672000015154, 0.16855322199990042]",tokens/s,6.017109310098951,, +float16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cpu,,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +bfloat16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cpu,,42,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,6131.986432,,,,,s,1,75.42604925199703,75.42604925199703,0.0,75.42604925199703,75.42604925199703,75.42604925199703,75.42604925199703,[75.42604925199703],,kWh,0.000861831033254877,0.0004697298935277112,0,0.0013315609267825883,,MB,6146.265088,,,,,s,10,10.038142496003275,1.0038142496003275,0.007483750476863813,1.0018462705011189,1.017723323602695,1.0177617248009483,1.0177924457595509,"[0.9997958220010332, 1.001708078001684, 1.0178001259992016, 1.0019844630005537, 0.9947841709981731, 1.0014055100000405, 0.9961156009994738, 1.0177147900030832, 1.0029842959993402, 1.0038496390006912]",tokens/s,255.02726236644617,kWh,1.186182167743204e-05,6.464976446052114e-06,0.0,1.8326798123484153e-05,tokens/kWh,13968615.700085597,MB,6177.23904,,,,,s,10,43.420734133000224,4.342073413300023,0.015974445791138894,4.337487816499561,4.362603413999386,4.369139515499955,4.374368396700411,"[4.3345059099992795, 4.375675617000525, 4.346620583000913, 4.333127848000004, 4.346158248001302, 4.334432168998319, 4.3140871780015, 4.36115094699926, 4.3370174179981404, 4.337958215000981]",tokens/s,14.509197335776808,kWh,5.143513356305322e-05,2.803414958886018e-05,0.0,7.946928315191341e-05,tokens/kWh,792759.1328534983,,s,630,43.41502509499696,0.06891273824602692,0.0010514641258632742,0.06872878699869034,0.07015264370056684,0.07098606734962232,0.07266658849894159,"[0.07180817699918407, 0.06879203900098219, 0.06859868899846333, 0.06893599500108394, 0.06836618799934513, 0.06927468499998213, 0.0686300620000111, 0.06807482899966999, 0.06838783499915735, 0.06802685300135636, 0.06818715999906999, 0.06791560200144886, 0.06703729899891187, 0.06888941299985163, 0.06831255500219413, 0.06777027499992982, 0.06838558500021463, 0.06769325799905346, 0.06852265799898305, 0.06905933799862396, 0.06839501600188669, 0.06920545599859906, 0.06756053500066628, 0.06722488400191651, 0.06793513699813047, 0.06802595400222344, 0.06873387799714692, 0.06841408200125443, 0.06812988700039568, 0.0679873189983482, 0.06844848000037018, 0.068997472000774, 0.06824447899998631, 0.06744469000113895, 0.06839796799977194, 0.06869091599946842, 0.06917675499789766, 0.06850085300175124, 0.06852024600084405, 0.068158426998707, 0.06749038099951576, 0.06862280800123699, 0.0689717479981482, 0.06996959000025527, 0.06890388600004371, 0.0684228610007267, 0.06829644900062704, 0.06956223000088357, 0.06983057899924461, 0.06959625300078187, 0.0701131469977554, 0.06984643699979642, 0.06911684500300908, 0.06944010199731565, 0.06956540900137043, 0.07169789399995352, 0.06993072000113898, 0.06873257099869079, 0.06891786399864941, 0.0688371320029546, 0.07182414699855144, 0.07013282600019011, 0.06927706799979205, 0.07143569199979538, 0.06967536699812626, 0.06849739300014335, 0.06798228200204903, 0.06792344199857325, 0.0685955240005569, 0.06918272300026729, 0.07071926600110601, 0.06946930199774215, 0.0698834610011545, 0.06891304699820466, 0.0686757570001646, 0.06801360200188356, 0.06960151699968264, 0.06868809799925657, 0.06956202400033362, 0.06871937500181957, 0.06884851299764705, 0.07042932200056384, 0.07415868300086004, 0.06967043599797762, 0.0730906259996118, 0.06835198000044329, 0.07009848300003796, 0.06939578600213281, 0.06943804500042461, 0.0690682829990692, 0.0709657959996548, 0.06944805899911444, 0.06786061599996174, 0.06835062599930097, 0.06806960399990203, 0.06814678200316848, 0.06847542799732764, 0.06996901200182037, 0.06779044499853626, 0.0693439960014075, 0.06965918700007023, 0.07190019599875086, 0.06876909800121211, 0.06826798299880465, 0.06849809699997422, 0.06947379299890599, 0.06894013300188817, 0.06895889299994451, 0.0697160529998655, 0.0739290669989714, 0.07510310900033801, 0.06995547699989402, 0.07006883600115543, 0.06844055699912133, 0.06887437900149962, 0.06846600299832062, 0.06806285000129719, 0.07022644599783234, 0.0684547710006882, 0.06836327800192521, 0.06962163599746418, 0.06852073200207087, 0.06904657899940503, 0.06929348500125343, 0.06947851899894886, 0.06857886400030111, 0.0723603299993556, 0.06908430800103815, 0.06879372599723865, 0.0688595620013075, 0.06772928099962883, 0.06864115799908177, 0.0693515090024448, 0.06930053899850463, 0.07025251000231947, 0.06934450999688124, 0.06929860099990037, 0.06831846500062966, 0.06812862600054359, 0.06862181299948134, 0.0706843209991348, 0.06945967400315567, 0.06848524700035341, 0.06878959299865528, 0.06969262100028573, 0.06843943399871932, 0.06774673800100572, 0.06770348799909698, 0.0684282379988872, 0.0674551820011402, 0.06883071700212895, 0.07008111699906294, 0.07188564099851646, 0.06932725500155357, 0.06880522599749384, 0.06785172800300643, 0.06782473599741934, 0.06808479300161707, 0.06815351299883332, 0.06784432300264598, 0.067818087998603, 0.06845222300034948, 0.06875547699746676, 0.06843373000083375, 0.06962154500070028, 0.06990794800003641, 0.06882652800049982, 0.06993443299870705, 0.06979917300122906, 0.0689491849989281, 0.0694473730000027, 0.06878774000142585, 0.06893773399860947, 0.06797405900215381, 0.06801352199909161, 0.07104142499883892, 0.07043690100181266, 0.07041367800047738, 0.06952864199774922, 0.06861022100201808, 0.06885404999775346, 0.07002485400153091, 0.06846055400092155, 0.06799924499864574, 0.06835993899949244, 0.06756043299901648, 0.0688104740002018, 0.06929582300290349, 0.06921469799999613, 0.07300772199960193, 0.06970840499707265, 0.06875301900072373, 0.06869539500257815, 0.06775803999698837, 0.06953223500022432, 0.06979862799926195, 0.06777460100056487, 0.07003503099986119, 0.06809027400231571, 0.07040292499732459, 0.06876122600078816, 0.06803755299915792, 0.06910106200302835, 0.0688760889970581, 0.0697526010008005, 0.0688411859991902, 0.07110576800187118, 0.06944491199828917, 0.06951626100271824, 0.06828473600035068, 0.06842866399892955, 0.06777259099908406, 0.06895878900104435, 0.06797358699986944, 0.0679750259987486, 0.06777221000083955, 0.06820673900074326, 0.06819596900095348, 0.06834961499771453, 0.06893773800038616, 0.06840723999994225, 0.06992296899989014, 0.06791973399958806, 0.06910382800197112, 0.0680545590003021, 0.06850872699942556, 0.07106207700053346, 0.06818686399856233, 0.0696900340008142, 0.06755293599780998, 0.06820279000021401, 0.0686517680005636, 0.06928145799975027, 0.06983338700229069, 0.06849278200024855, 0.06741859199973987, 0.06829902199751814, 0.06754432500019902, 0.06826532000195584, 0.06863872499889112, 0.06808233699848643, 0.06792719300210592, 0.06826346999878297, 0.06768378999913693, 0.0685203110006114, 0.06820354899900849, 0.07023488600316341, 0.06856510499710566, 0.0686278780012799, 0.06906062399866641, 0.06840209599977243, 0.06798870400234591, 0.07105229100125143, 0.06824344699998619, 0.0678640109981643, 0.06916303600155516, 0.06863445799899637, 0.06792906500049867, 0.06797605500105419, 0.0677739089987881, 0.06818368500171346, 0.06772823599749245, 0.06856359200173756, 0.06844417699903715, 0.06862281400026404, 0.06967717899897252, 0.0713304769997194, 0.06885583600160317, 0.06857393400059664, 0.06895803999941563, 0.06835765000141691, 0.06799845399655169, 0.06860428500294802, 0.06834277699817903, 0.06799448500169092, 0.06843851699886727, 0.06796729599955142, 0.06862275900130044, 0.06721802899846807, 0.06851106400063145, 0.06910703500034288, 0.06817858800059184, 0.0694150859999354, 0.06833364799967967, 0.06802324000091176, 0.06890907999695628, 0.06877796500339173, 0.06918644299730659, 0.06921024400071474, 0.06846556500022416, 0.07026242099891533, 0.07214136500260793, 0.07231172799947672, 0.07019858000057866, 0.06876260999706574, 0.06950572300047497, 0.07037973200203851, 0.06862610099778976, 0.06942560800234787, 0.06893650499841897, 0.07095125599880703, 0.06940631900215521, 0.06910623599833343, 0.06858806200034451, 0.07043090200022561, 0.06977383699995698, 0.06947601800129632, 0.06983519499772228, 0.06874956000319798, 0.06904673799726879, 0.06924336000156472, 0.06826098099918454, 0.0680641930011916, 0.06846806799876504, 0.06839114000104018, 0.07136686200101394, 0.06983542199668591, 0.0677184619999025, 0.06882759200016153, 0.0672688400009065, 0.06756842100003269, 0.06735470499916119, 0.06758426500164205, 0.07026973900065059, 0.06888617199729197, 0.06975610900190077, 0.06779767699845252, 0.0677526340004988, 0.06871178499932284, 0.0674958970012085, 0.06816960600190214, 0.06744163399707759, 0.06868491800196352, 0.06911096700059716, 0.06896863099973416, 0.06895377399996505, 0.06901221799853374, 0.06888388700099313, 0.06880948799880571, 0.07018747400070424, 0.0695478039997397, 0.07305708100102493, 0.07008003800001461, 0.07279167999877245, 0.06933370000115247, 0.06869796799946926, 0.06885379899904365, 0.06888121900192345, 0.06813412499832339, 0.0673199980010395, 0.06872750199909206, 0.06783879200156662, 0.06893653999941307, 0.06897303299774649, 0.06904742200276814, 0.06896624499859172, 0.06892837000123109, 0.06851561599978595, 0.06890353600101662, 0.06789630399725866, 0.06800142400243203, 0.06765878399892244, 0.06964496499858797, 0.06885974100077874, 0.06881482400058303, 0.06859651700142422, 0.06923817699862411, 0.06797706300130812, 0.068555429999833, 0.06901574599760352, 0.06889749800029676, 0.06767101299919887, 0.06840623000243795, 0.06913689700013492, 0.06844463199740858, 0.06808575200193445, 0.06802406800125027, 0.06906332699873019, 0.07090084200171987, 0.06905386699872906, 0.06781745700209285, 0.0682541059977666, 0.06868545500037726, 0.06796396800200455, 0.06803025900080684, 0.06761759599976358, 0.06793305199971655, 0.06801805799841532, 0.0680532949991175, 0.06736809500216623, 0.06797112499771174, 0.06824037700062036, 0.06786676800038549, 0.06806969000172103, 0.06817228499858174, 0.06795947100181365, 0.06830926599650411, 0.06834968900147942, 0.06801073599854135, 0.06790428400199744, 0.06878966599833802, 0.06895404200258781, 0.06816655299917329, 0.06814428299912834, 0.06752122300167684, 0.06921782099743723, 0.07070012199983466, 0.06789225500324392, 0.07139182599712512, 0.06797985800221795, 0.0670162349997554, 0.06835601999773644, 0.06756635500278207, 0.06781356699866592, 0.06871041300109937, 0.0696255859984376, 0.06828809400030877, 0.06863028299994767, 0.06910485100161168, 0.06800047599972459, 0.06837226800053031, 0.06805679199896986, 0.06800134500008426, 0.07104237299790839, 0.06886506900264067, 0.06877838899890776, 0.06880612899840344, 0.06881143700229586, 0.06873440599883907, 0.068706107998878, 0.06829691100210766, 0.06873007199828862, 0.06787834100032342, 0.06901394799933769, 0.06967934600106673, 0.06805691100089462, 0.06815172400092706, 0.06808071999694221, 0.0679686550029146, 0.06885297599728801, 0.06818688400016981, 0.07174596699769609, 0.06904518400187953, 0.06865422799819498, 0.06948366399956285, 0.0718056030018488, 0.06812223700035247, 0.06790460699994583, 0.06873914999960107, 0.06842906900055823, 0.06831951399726677, 0.06793497600301635, 0.06871437999870977, 0.06871688300088863, 0.06906843899923842, 0.06813411699840799, 0.06914359200163744, 0.0687777860002825, 0.06888808999792673, 0.07126078200235497, 0.06964832399899024, 0.0692676559992833, 0.06846070699975826, 0.06787399700260721, 0.06870575999710127, 0.069629387002351, 0.07111314599751495, 0.06989394999982323, 0.06884614000227884, 0.06799302299987176, 0.07081683800061, 0.07030458199733403, 0.06853104100082419, 0.06879149399901507, 0.06774486100039212, 0.06898578199979966, 0.06857276099981391, 0.06870664300004137, 0.0684365200031607, 0.07017648199689575, 0.06941476300198701, 0.06908020199989551, 0.06909295799778192, 0.06881412500297301, 0.07012948799820151, 0.06918663699980243, 0.06980593399930513, 0.07051891400260502, 0.06803065799977048, 0.06842366199998651, 0.06836523300080444, 0.06930667199776508, 0.06832906099953107, 0.06821610300175962, 0.06820716899892432, 0.07071491000169772, 0.06923622300018906, 0.06979227699775947, 0.06891248700048891, 0.07067801800076268, 0.0695999790004862, 0.06996831300057238, 0.07036059299935005, 0.07105542199860793, 0.07100265299959574, 0.06835261699961848, 0.0698453500008327, 0.06896385999789345, 0.06824796400178457, 0.06801126999926055, 0.06849223899916979, 0.06868777899944689, 0.06918948700331384, 0.06774440999652143, 0.06858318000013242, 0.06832623300215346, 0.06840134199956083, 0.06862473900037003, 0.06811308299802477, 0.06852089200037881, 0.06890448800186277, 0.0691743530005624, 0.06837944099970628, 0.07023305799884838, 0.06936702700113528, 0.07025972900009947, 0.06877464599892846, 0.06854860100065707, 0.06839345099797356, 0.06961967300230754, 0.0694382340006996, 0.0696434949968534, 0.06929700600085198, 0.06821437899998273, 0.06842842100013513, 0.06900814099935815, 0.06855595700108097, 0.06820182299998123, 0.06889673800105811, 0.06789316299909842, 0.06945287800044753, 0.06850370800020755, 0.06774874599796021, 0.06780728500234545, 0.0685700949979946, 0.06800247200226295, 0.06756391399903805, 0.0688055130012799, 0.06810579899683944, 0.06889781600330025, 0.06866976899982546, 0.0681112940001185, 0.0690300619971822, 0.06864214999950491, 0.0689316370007873, 0.06854225200004294, 0.0688963730026444, 0.06910991999757243, 0.06867673300075694, 0.06937388900041697, 0.07081841599938343, 0.0690725870008464, 0.07014999500097474, 0.06948754799668677, 0.06924370400156477, 0.06864819299880764, 0.0693208720003895, 0.07116927400056738, 0.06889779799894313, 0.06729137699949206, 0.06786528999873553, 0.06824960900121368, 0.06821805499930633, 0.068150595001498, 0.06873157799782348, 0.06785844200203428, 0.0697151690001192, 0.0672825149995333, 0.06786094999915804, 0.06916432200159761, 0.0680361989980156, 0.06810881600176799, 0.06841199399787001, 0.06946380600129487, 0.06923883399940678, 0.06810581200261367, 0.06945775799977127, 0.0679719709987694, 0.06841992699992261, 0.06832726900029229, 0.06785696400038432, 0.06893402800051263, 0.06807649699840113, 0.0680754450004315, 0.06748542200148222, 0.06785208699875511, 0.06823653700121213, 0.0692265559991938, 0.06859360099770129, 0.06879522400049609, 0.06858724600169808, 0.06799878499805345, 0.06854619300065679, 0.06899833900024532, 0.06833157599976403, 0.06831361800141167, 0.06942537699796958, 0.06794491700202343, 0.06928211999911582, 0.06902075100151706, 0.07063152099726722, 0.07064884300052654, 0.07007358000191743, 0.06991609299802803, 0.06864828600009787, 0.06976530800238834, 0.06892309499744442, 0.07001925300210132, 0.07012071099961759, 0.06919236799876671, 0.06955724300132715, 0.06976507799845422, 0.07200298600218957, 0.06959300199741847, 0.06842358699941542, 0.06887628300319193, 0.0697726660000626, 0.06885971499650623, 0.06816399400122464, 0.06882414700157824]",tokens/s,14.511105282594887,, +float16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cpu,,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cpu,,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTJForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +bfloat16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cpu,,42,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,3374.096384,,,,,s,1,40.40298847700069,40.40298847700069,0.0,40.40298847700069,40.40298847700069,40.40298847700069,40.40298847700069,[40.40298847700069],,kWh,0.000448394099476396,0.00024439086347910477,0,0.0006927849629555008,,MB,3386.372096,,,,,s,10,5.081101806001243,0.5081101806001243,0.0032179546311577414,0.507622211499438,0.5092548716005695,0.5132234783006424,0.5163983636607008,"[0.5079443670001638, 0.5074658419998741, 0.507770926999001, 0.507473495999875, 0.5171920850007155, 0.5060488750004879, 0.5048994710014085, 0.5057695409996086, 0.5083729590005532, 0.5081642429995554]",tokens/s,503.82773220099773,kWh,6.0505089412151845e-06,3.297670236284909e-06,0.0,9.348179177500095e-06,tokens/kWh,27385012.111894492,MB,3404.587008,,,,,s,10,23.12246347800101,2.3122463478001007,0.014848238735775832,2.307289389000289,2.33329433179897,2.3358067483999547,2.3378166816807426,"[2.30246206800075, 2.3067284820008354, 2.3383191650009394, 2.307804195999779, 2.3288760220002587, 2.307183516000805, 2.3327360169987514, 2.301445855999191, 2.3073952619997726, 2.2895128939999267]",tokens/s,27.246231812600318,kWh,2.6998047201978276e-05,1.4714908403085284e-05,0.0,4.171295560506356e-05,tokens/kWh,1510322.1310060415,,s,630,23.11729763599942,0.03669412323174511,0.0008016607430829591,0.036572998499650566,0.037623678699492305,0.0383038886497161,0.03921609268065367,"[0.03790680100064492, 0.036813442999118706, 0.037447701000928646, 0.03673492999951122, 0.03649803400003293, 0.035982187999252346, 0.03536354700008815, 0.036021641000843374, 0.037721619999501854, 0.03851450299953285, 0.03778651100037678, 0.036231922000297345, 0.03621369900065474, 0.035697913999683806, 0.036058708999917144, 0.03535627800010843, 0.03572928699941258, 0.035403735000727465, 0.03562995299944305, 0.035848076000547735, 0.03526092799984326, 0.03611745499983954, 0.03590674999941257, 0.035892257999876165, 0.036615491000702605, 0.036096010999244754, 0.03681286200117029, 0.03721185299946228, 0.03661302299951785, 0.037021471000116435, 0.037095745999977225, 0.03644385400002648, 0.03598998500092421, 0.03679102299975057, 0.03617690700048115, 0.03661958199882065, 0.03646255800049403, 0.03644876199905411, 0.03578236800058221, 0.036029066000992316, 0.03571526599989738, 0.0362737819996255, 0.03742712799976289, 0.03655288799927803, 0.03725210000084189, 0.037070391999805, 0.03691500700006145, 0.036864237999907346, 0.03654631999961566, 0.036005166000904865, 0.036343293999379966, 0.038588674000493484, 0.03902937800012296, 0.03634888200031128, 0.03681794099975377, 0.036391116000231705, 0.036482213999988744, 0.036948741999367485, 0.03629457200077013, 0.036229421999451006, 0.03672414599896001, 0.03631436300020141, 0.03649546800079406, 0.0389668839998194, 0.03596958799971617, 0.03578155099967262, 0.035859275001712376, 0.035150828000041656, 0.03593736799848557, 0.035626382999907946, 0.03604373000052874, 0.03595908600073017, 0.0358109560002049, 0.03612995999901614, 0.03626054900087183, 0.03540191999854869, 0.03690726700006053, 0.03696608500104048, 0.03696723299981386, 0.0362281220004661, 0.03634674199929577, 0.03818889400099579, 0.0363015329985501, 0.036293667000791174, 0.03596807000030822, 0.0365960979997908, 0.036608243999580736, 0.035794137000266346, 0.03628768300040974, 0.03573013599998376, 0.03633529000035196, 0.03737342199929117, 0.03603681700042216, 0.03634032399895659, 0.0362554800012731, 0.03656467999826418, 0.035777136001343024, 0.03554753400021582, 0.036346790999232326, 0.03619281400096952, 0.0367679879982461, 0.03583234700090543, 0.03620689799936372, 0.03634822800086113, 0.03673148199959542, 0.03663812900049379, 0.03651528499904089, 0.040629072000228916, 0.03677582000091206, 0.03710446100012632, 0.036893771999530145, 0.039157985000201734, 0.036926335998941795, 0.036822069001573254, 0.03644847399846185, 0.03622472400093102, 0.03650895399914589, 0.036893919999783975, 0.03675953400124854, 0.03669945900037419, 0.037287283999830834, 0.03728904699892155, 0.03706063699974038, 0.03712427600112278, 0.03726873999949021, 0.0382990800007974, 0.039765706000252976, 0.0377993410002091, 0.03788219700072659, 0.035961727999165305, 0.035822076999465935, 0.03688620300090406, 0.036482568000792526, 0.036702519999380456, 0.03617849100010062, 0.037071933000333956, 0.038150940999912564, 0.037371735999840894, 0.03742635999878985, 0.03683287400053814, 0.03693917199962016, 0.039041423000526265, 0.036286386000938364, 0.03597241399984341, 0.03674580799997784, 0.03649367399884795, 0.036110140001255786, 0.03652080799838586, 0.036735985000632354, 0.03734995299964794, 0.03743354300058854, 0.036554065000018454, 0.03729134800050815, 0.036769679998542415, 0.03748015300152474, 0.03708944899881317, 0.036790564001421444, 0.03705645299851312, 0.03686995300085982, 0.037245072999212425, 0.03757425600088027, 0.03690587500022957, 0.03654532600012317, 0.03715445399939199, 0.03693220800050767, 0.03719516899946029, 0.03678558700084977, 0.03680952399918169, 0.03733535399987886, 0.03701017000093998, 0.03856342299877724, 0.03852016500059108, 0.038251239000601345, 0.03763149699989299, 0.03709655499915243, 0.03734185299981618, 0.03731107200110273, 0.03711342899987358, 0.036238253000192344, 0.03722969599948556, 0.03671250799925474, 0.03733983900019666, 0.03690709500006051, 0.03694392999932461, 0.03700446800030477, 0.0365735430004861, 0.03745399000035832, 0.0370426999998017, 0.03719930599982035, 0.0385343460002332, 0.036987178998970194, 0.03640571500000078, 0.036464732000240474, 0.036469857001065975, 0.03599840799870435, 0.036301306001405464, 0.03601056999832508, 0.03559088600013638, 0.0368408089998411, 0.036484667001786875, 0.03582615999948757, 0.03548408899951028, 0.035974058000647346, 0.03660716699960176, 0.03714833699996234, 0.03660538199983421, 0.035698969000804937, 0.03598461799992947, 0.0358603229997243, 0.03671576600027038, 0.03651226699912513, 0.03694169099981082, 0.036570817001120304, 0.036857008999504615, 0.03750944300008996, 0.036562563000188675, 0.03622693899887963, 0.03637144800086389, 0.03597650099982275, 0.03865681599927484, 0.03680148900093627, 0.0367619629996625, 0.03673626099953253, 0.03627496100125427, 0.037296234999303124, 0.03627927299930889, 0.036306449001131114, 0.03658990999974776, 0.03640781399917614, 0.03600431900122203, 0.0358750979994511, 0.036534036000375636, 0.03638348599997698, 0.036382985999807715, 0.03688554399923305, 0.03749001800133556, 0.03760825599965756, 0.03630560199962929, 0.03671872599989001, 0.036421147000510246, 0.0365430589990865, 0.037124579001101665, 0.03687788499883027, 0.036962264000976575, 0.03721718700035126, 0.03687774899844953, 0.03698197599987907, 0.037151980000999174, 0.036636131999330246, 0.03717385300114984, 0.03686308999931498, 0.03667377599958854, 0.03874508399894694, 0.037048009000500315, 0.03623883299951558, 0.036099638000450796, 0.03588256500006537, 0.036713499001052696, 0.03641671699915605, 0.03700464500070666, 0.03628517799916153, 0.036403940999662154, 0.03663292300007015, 0.036196314000335406, 0.03612036599952262, 0.03631679099999019, 0.03671410600145464, 0.03761577099976421, 0.039328000999375945, 0.03671846900033415, 0.03636276199904387, 0.03620182600025146, 0.036463047999859555, 0.03708477000145649, 0.0371359989985649, 0.036900770999636734, 0.037518950000958284, 0.037304817000404, 0.03647677699882479, 0.03671667300113768, 0.0360267340001883, 0.036910504999468685, 0.03648875900034909, 0.03791160299988405, 0.03909626600034244, 0.040226649998658104, 0.03906548600025417, 0.03717940500064287, 0.03681554799914011, 0.03649303800011694, 0.036447784999836585, 0.037108939000972896, 0.03707523299999593, 0.03687372499916819, 0.03698360000089451, 0.037535949999437435, 0.03656974700061255, 0.03668634099994961, 0.036259839000194916, 0.03662298999915947, 0.037505624000914395, 0.03748639299919887, 0.036537037000016426, 0.037033141999927466, 0.03700083300100232, 0.03713871399850177, 0.03719430100136378, 0.03706458299893711, 0.03716835200066271, 0.03715874800036545, 0.03600632899906486, 0.03610198299975309, 0.037008000999776414, 0.03670497800158046, 0.036238228998627164, 0.03923593100080325, 0.03733675100011169, 0.036616440000216244, 0.03693394999936572, 0.03612157499992463, 0.035973980999187916, 0.035648631001095055, 0.03791297200041299, 0.040171193999412935, 0.035974428999907104, 0.035706331000255886, 0.038211197999771684, 0.03669909100062796, 0.03553469399957976, 0.03580450099980226, 0.03532830800031661, 0.035751514000367024, 0.03615323599842668, 0.0361438210002234, 0.03565372100092645, 0.03596280899910198, 0.0364880320012162, 0.0362254479987314, 0.03596990400001232, 0.03634206000060658, 0.03725001299972064, 0.03744937699957518, 0.036460819999774685, 0.036440456000491395, 0.03676620100122818, 0.03635710899834521, 0.03611733700017794, 0.03707196400137036, 0.035829271999318735, 0.036219064000761136, 0.03564180199828115, 0.03610255500097992, 0.036634829999457, 0.036485501999777625, 0.03627823799979524, 0.03653998400113778, 0.03633625899965409, 0.03613634899920726, 0.03638232200137281, 0.03682842299895128, 0.037025994000941864, 0.03650700499929371, 0.03707638699961535, 0.036602900001525995, 0.03605148799942981, 0.036446348000026774, 0.03588400099943101, 0.03709285700097098, 0.03576218099988182, 0.03599582700007886, 0.036174116999973194, 0.03598441699978139, 0.03710393599976669, 0.037288032999640564, 0.0378762310010643, 0.0383078229988314, 0.03853452200019092, 0.03762280999944778, 0.03864258799876552, 0.037305960000594496, 0.03737313300007372, 0.036885479999909876, 0.037817958000232466, 0.03804544699960388, 0.03865906700048072, 0.03910531200017431, 0.03789507699912065, 0.03733990500040818, 0.036952889000531286, 0.036088495999138104, 0.03688814800079854, 0.03699418199903448, 0.03679574200032221, 0.037117079000381636, 0.03661180999915814, 0.0360662120001507, 0.03671205400132749, 0.03739446999861684, 0.03815442599989183, 0.037138982999749715, 0.03856326000095578, 0.0383682060000865, 0.03729638999902818, 0.03619401100149844, 0.03674162799870828, 0.03635433599993121, 0.036588061000657035, 0.03697295399979339, 0.03677390899974853, 0.03572643100051209, 0.03649499799939804, 0.036727922000864055, 0.0368890649988316, 0.03573207000044931, 0.03607171700059553, 0.035897081999792135, 0.0366773280002235, 0.036476150000453345, 0.03623985400008678, 0.03625020299841708, 0.036669220000476344, 0.0359103970004071, 0.036375309000504785, 0.036228334000043105, 0.036572453998815035, 0.03655646200058982, 0.03704748299969651, 0.03710551400035911, 0.03680754599918146, 0.036975299000914674, 0.03670494099969801, 0.03696641900023678, 0.036575292999259545, 0.03684796700144943, 0.0375083969993284, 0.037367953000284615, 0.037731802000052994, 0.03857848900042882, 0.0374754969998321, 0.038034630999391084, 0.03720287900068797, 0.03803880399937043, 0.03644171400082996, 0.036395987999640056, 0.035596026998973684, 0.03561477000039304, 0.03595121600119455, 0.03711882599964156, 0.03657194599873037, 0.036008382001455175, 0.035568548999435734, 0.0354615840005863, 0.03580259699992894, 0.035513393999281107, 0.035390717999689514, 0.035142767999786884, 0.035725130001083016, 0.03556308400038688, 0.03604086599989387, 0.035770593998677214, 0.03596325699982117, 0.03587554700061446, 0.035467298001094605, 0.0357138949984801, 0.0354714260010951, 0.03656224699989252, 0.03578297899912286, 0.03551618200071971, 0.036045892999027274, 0.03564682500109484, 0.03601319500012323, 0.03622196499964048, 0.03619511900069483, 0.036537727999530034, 0.03676917699885962, 0.036242442000002484, 0.036717425000460935, 0.03617694000058691, 0.03711102600027516, 0.03705813600026886, 0.03720920999876398, 0.036760724000487244, 0.036303127999417484, 0.036976033999962965, 0.03721761100132426, 0.03835607499968319, 0.03768139499879908, 0.0368931020002492, 0.037006780001320294, 0.03633502199954819, 0.037901444999079104, 0.03712211099991691, 0.037211426000794745, 0.036477711000770796, 0.0366851869985112, 0.03771916100049566, 0.03696927000055439, 0.03771277799933159, 0.039167523000287474, 0.03832707999936247, 0.036475939999945695, 0.036700921000374365, 0.036987302999477834, 0.037956400999973994, 0.03801575800025603, 0.03679886399913812, 0.036132870000074035, 0.03577595400020073, 0.03590179000093485, 0.03678503699848079, 0.03569562099983159, 0.03566550499999721, 0.035673304000738426, 0.03588180299993837, 0.035691162000148324, 0.036251608000384294, 0.03664430099888705, 0.036585131001629634, 0.03700651199869753, 0.036172090000036405, 0.036884049000946106, 0.03728740999940783, 0.03714959799981443, 0.036697675999675994, 0.03607763900072314, 0.03665171099964937, 0.036793045001104474, 0.03652607000003627, 0.03626112499841838, 0.035881871001038235, 0.03706558299927565, 0.03589311900032044, 0.036374851000800845, 0.03623183399940899, 0.036481675000686664, 0.03589465199911501, 0.03597670400085917, 0.03715134799858788, 0.03622158000143827, 0.03646919999846432, 0.038909453000087524, 0.03626020000046992, 0.03660974900049041, 0.03666655499910121, 0.03615602600075363, 0.036542137000651564, 0.03647969899975578, 0.036897216999932425, 0.03746882099949289, 0.03593070799979614, 0.03696254900023632, 0.0370061909998185, 0.03695417500057374, 0.03752991999863298, 0.03978967500006547, 0.0363244920008583, 0.03594888199950219, 0.037884016001044074, 0.037898212998698, 0.03672458900109632, 0.036747823000041535, 0.03602183099974354, 0.03609441200023866, 0.036499919999187114, 0.03670611500092491, 0.03690377899874875, 0.03635983000094711, 0.037998980000338634, 0.036352564000480925, 0.03635980899889546, 0.03612642300140578, 0.0357992720000766, 0.03593323699897155, 0.035935576999690966, 0.03602392500033602, 0.0362286050003604, 0.0358056619988929, 0.035465027000100235, 0.03579208300106984, 0.03599251500054379, 0.035910297998270835, 0.03562394300024607, 0.035886638001102256, 0.036191360999509925, 0.035876374000508804, 0.035808421998808626, 0.036285714000769076, 0.03636367500075721, 0.03567370299970207, 0.03729610399932426, 0.035862699000063, 0.036659740000686725, 0.03658610099955695, 0.036613959999158396, 0.037045136999950046, 0.03627109599983669, 0.03711175700118474, 0.0362037529994268, 0.03560088599988376, 0.03627619600047183, 0.03630607100058114, 0.03681873099958466, 0.03658314799940854, 0.03585583200037945, 0.03685825100001239, 0.037596164000206045, 0.0377880260002712, 0.03636570500020753, 0.036465862998738885, 0.03594504700049583, 0.036493404999419, 0.03585429400118301, 0.03572473599888326, 0.035988472000099136, 0.03604691800137516, 0.036504433999652974, 0.03651804400033143, 0.03635389299961389, 0.036199041998770554, 0.03617827600101009, 0.0369907839994994, 0.036485115000687074, 0.03707744999883289, 0.03689945600126521, 0.03632140899935621, 0.03650690299946291, 0.03635003900126321, 0.03639143500004138, 0.03630615699876216, 0.03627472200059856]",tokens/s,27.252320315283402,, +bfloat16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cpu,,42,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTJForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float32-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cpu,,42,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +bfloat16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cpu,,42,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float32-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cpu,,42,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 57, in launch + raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") +RuntimeError: Isolated process exited with non-zero code -9 +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float32-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cpu,,42,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,11693.142016,,,,,s,1,29.41541665099794,29.41541665099794,0.0,29.41541665099794,29.41541665099794,29.41541665099794,29.41541665099794,[29.41541665099794],,kWh,0.0003189265374680771,0.0001738259765790067,0,0.0004927525140470838,,MB,11812.90496,,,,,s,10,7.102716182995209,0.7102716182995209,0.005096656777520243,0.7078282490001584,0.7164747309991071,0.7188930624985005,0.7208277276980152,"[0.7062851859991497, 0.7065217999988818, 0.7062391380022746, 0.707326440999168, 0.7049440939990745, 0.7144621869992989, 0.7083300570011488, 0.7113585619990772, 0.7159373239992419, 0.7213113939978939]",tokens/s,360.42549554900705,kWh,8.446054256696951e-06,4.6032773856984855e-06,0.0,1.3049331642395436e-05,tokens/kWh,19617862.969187796,MB,11981.279232,,,,,s,10,51.29337705800572,5.129337705800571,0.03471973986825579,5.127027131000432,5.181126338601098,5.18282017730071,5.184175248260399,"[5.100540788000217, 5.158111803000793, 5.124483988998691, 5.137218851999933, 5.112562635000359, 5.180749930001184, 5.184514016000321, 5.080222856002365, 5.085401915999682, 5.129570273002173]",tokens/s,12.282287424506231,kWh,6.084029665746813e-05,3.3160265560703406e-05,0.0,9.400056221817155e-05,tokens/kWh,670208.7574091261,,s,630,51.28851896800916,0.08141034756826851,0.0020656284087260104,0.08106037899960938,0.08408412230055547,0.08524025579863519,0.08841822099006097,"[0.08434359899911215, 0.08135598900116747, 0.08063318700078526, 0.08336111500102561, 0.08225446999858832, 0.08032969600026263, 0.08338929499950609, 0.07945668500178726, 0.07951266799864243, 0.07965856000009808, 0.08176748299956671, 0.08130000600067433, 0.08008907999828807, 0.07938705600099638, 0.08031909200144582, 0.08020780199876754, 0.08029974799865158, 0.08128857900010189, 0.08027474499976961, 0.0797725590018672, 0.0816415089975635, 0.08175332400060142, 0.08516732800126192, 0.0821527789994434, 0.07954492399949231, 0.07908027700250386, 0.08499813699745573, 0.08153586200205609, 0.0861508159978257, 0.08125240300068981, 0.08095723299993551, 0.08083177500157035, 0.08209521800017683, 0.08317917199747171, 0.08127142500234186, 0.0780251389987825, 0.07873162100077025, 0.0791358839996974, 0.07874153900047531, 0.07969426799900248, 0.0808876400005829, 0.08150139999997918, 0.08204924399979063, 0.07965084399984335, 0.08067613599996548, 0.08081984700038447, 0.0830861009999353, 0.08283192399903783, 0.08175904400195577, 0.08245104899833677, 0.07941551900148625, 0.08013948999723652, 0.08018356899992796, 0.08085417400070583, 0.0799123580000014, 0.08079054600239033, 0.0791244079991884, 0.07903808099945309, 0.07976090000011027, 0.08023501299976488, 0.07969694100029301, 0.07903039799930411, 0.08103991000098176, 0.08274569400236942, 0.07949822299997322, 0.07961581699782982, 0.07937278900135425, 0.08095311999932164, 0.07852035399992019, 0.07900115400116192, 0.07922065900129382, 0.08156216499992297, 0.08019676699768752, 0.07835902000078931, 0.08178172100087977, 0.07981243099857238, 0.0787114400009159, 0.08037147899813135, 0.08037584900012007, 0.08000081600039266, 0.07860558800166473, 0.08246394199886709, 0.08136427299905336, 0.08106338400102686, 0.08353343099952326, 0.08244291899973177, 0.08021308300158125, 0.08155053700102144, 0.07992578099947423, 0.0818153779982822, 0.08147107600234449, 0.08233137499701115, 0.08081164699979126, 0.08886218300176552, 0.08486347600046429, 0.0839919350000855, 0.08201321700107655, 0.08418783199886093, 0.08265866600049776, 0.08155049400011194, 0.08228993900047499, 0.08296992299801786, 0.08138459499969031, 0.08219717099927948, 0.08646272000260069, 0.08089513400045689, 0.07962249099728069, 0.08016036499975598, 0.08167531000071904, 0.08476605499890866, 0.08359672300139209, 0.08412686100200517, 0.08370719199956511, 0.08339975099806907, 0.0836675099999411, 0.08398238700101501, 0.08446352199825924, 0.08449810200181673, 0.08505310599866789, 0.08204921400101739, 0.08337555300022359, 0.08323290000043926, 0.08042243299860274, 0.08037437600069097, 0.0823852169996826, 0.08100089599975036, 0.08860123100021156, 0.08523937799691339, 0.08202214500124683, 0.08175280000068597, 0.07933035499809193, 0.08069278800030588, 0.08211686600043322, 0.07925638899905607, 0.08175117100108764, 0.08518658400134882, 0.08148218399946927, 0.08210832600161666, 0.07963267799641471, 0.08067462700273609, 0.08091380399855552, 0.08013172500068322, 0.07856844300113153, 0.07907441999850562, 0.08283059200039133, 0.08102530700125499, 0.08041498899910948, 0.08138976600093883, 0.08217116299783811, 0.07975683200129424, 0.08054265800092253, 0.07850665399746504, 0.07932500400056597, 0.0796500520009431, 0.08207459899858804, 0.0840832430003502, 0.0818797319989244, 0.08299603800332989, 0.0817877729969041, 0.08329713200146216, 0.08595244100069976, 0.0817027839984803, 0.08148746699953335, 0.08274880500175641, 0.08118620299865142, 0.07902454700160888, 0.08010951599862892, 0.08097824400101672, 0.08129679299963755, 0.08161428500170587, 0.08014024599833647, 0.08174685300036799, 0.07974845799981267, 0.08185910199972568, 0.08613746300034109, 0.08159371700094198, 0.08231327099929331, 0.08272859099815832, 0.07962941500227316, 0.07859475199802546, 0.07861024299927521, 0.0794696120028675, 0.07999930799996946, 0.0806045589997666, 0.08132538099744124, 0.08194140700288699, 0.08067518700045184, 0.0815497219991812, 0.07905281700004707, 0.08487912100099493, 0.08094798100137268, 0.08328502899894374, 0.08108496400018339, 0.08118293399820686, 0.08052631800092058, 0.07851454200135777, 0.08129302399902372, 0.08050136299789301, 0.07952315600050497, 0.07996220300265122, 0.07849465099934605, 0.08021085300060804, 0.07918570199763053, 0.08114151500194566, 0.07912000099895522, 0.08081168300122954, 0.08137794499998563, 0.07923411799856694, 0.0809669459995348, 0.08620067900119466, 0.07889296500070486, 0.07944675399994594, 0.07963304399891058, 0.0797391529995366, 0.07821145900015836, 0.08063204800055246, 0.08010868199926335, 0.08289022000099067, 0.08111690900113899, 0.08000894499855349, 0.08165140699929907, 0.07972888600124861, 0.08437184099966544, 0.0822269279997272, 0.0830511550011579, 0.08351944199966965, 0.08163527699798578, 0.07923918700180366, 0.08024536800076021, 0.08060818699959782, 0.08118010399994091, 0.08184815200002049, 0.08288527900003828, 0.0856724769982975, 0.08729704000143101, 0.08460385699800099, 0.08216496800014284, 0.08410247400024673, 0.08090097300009802, 0.07995015400229022, 0.08797016199969221, 0.08156900299945846, 0.08917290399767808, 0.08140798500244273, 0.07960598099816707, 0.08241963500040583, 0.08405656300237752, 0.08029892399645178, 0.08004470800005947, 0.08086528600324527, 0.08196780099751777, 0.0810935920017073, 0.0826436810020823, 0.08131533799678436, 0.08430248600052437, 0.08098368699938874, 0.08207296700129518, 0.0802896579989465, 0.07955265100099496, 0.07785696100108908, 0.07881661699866527, 0.0829497420018015, 0.08049417299844208, 0.0790910329997132, 0.08430498300003819, 0.0841290870012017, 0.08446829600143246, 0.08365530099763419, 0.07998880200102576, 0.07978443400133983, 0.07939829899987672, 0.07863306999934139, 0.07996505799746956, 0.07965972800229792, 0.08004886599883321, 0.07935111500046332, 0.08073482400141074, 0.08047217299827025, 0.08098284800144029, 0.0798860329996387, 0.08052093999867793, 0.0800345150018984, 0.08279126999696018, 0.07938733100309037, 0.08351175200004946, 0.08160453299933579, 0.07991035799932433, 0.08215252100126236, 0.07990289299777942, 0.08086076000108733, 0.08286864600086119, 0.08141126799819176, 0.08398168799976702, 0.08202080499904696, 0.07925029300167807, 0.08197641700098757, 0.08100900099816499, 0.07896742999946582, 0.08109124100155896, 0.0857474859985814, 0.08212519700100529, 0.08298575500157312, 0.08089185599965276, 0.08431780699902447, 0.08284929000001284, 0.0808067990001291, 0.08006968300105655, 0.07971065499805263, 0.08112939599959645, 0.08123765300115338, 0.08123486900149146, 0.08140675599861424, 0.0783246879982471, 0.08058030100073665, 0.07967274900147459, 0.08795396800269373, 0.08368934999816702, 0.08440743599931011, 0.0823501760023646, 0.08274558699849877, 0.08326191599917365, 0.08235948600122356, 0.08282533699821215, 0.08194154200100456, 0.08172396200097864, 0.08931736299928161, 0.08569052500024554, 0.08232156899975962, 0.08328660899860552, 0.08182652400137158, 0.08370448500136263, 0.08151950699902955, 0.08174719100134098, 0.07892168199759908, 0.07890058200064232, 0.07962296199912089, 0.08040437900126562, 0.08229179699992528, 0.08237514099891996, 0.08139199600191205, 0.08166610599801061, 0.08314393399996334, 0.08273020900014671, 0.08317647000149009, 0.08255473499957589, 0.08790556199892308, 0.08357194700147375, 0.08962981299919193, 0.08213735499884933, 0.0827406480020727, 0.07999009099876275, 0.08289354600128718, 0.08973240799969062, 0.08687384299992118, 0.08345314400139614, 0.08228497199888807, 0.08215581900003599, 0.08238574400093057, 0.07977187799770036, 0.07994144700205652, 0.08166412199716433, 0.08058090900158277, 0.07942098599960445, 0.07967161800115719, 0.07972760899792775, 0.0785655610015965, 0.07975297099983436, 0.07897668499936117, 0.08008350700038136, 0.07975764400180196, 0.07957217099828995, 0.07866714099873207, 0.08030931600296753, 0.07997856899964972, 0.08096509799725027, 0.08526485600305023, 0.08119519499814487, 0.08073254600094515, 0.08311555799809867, 0.08183520900274743, 0.08318634199895314, 0.08620045900170226, 0.0818903159997717, 0.08010149199981242, 0.0797592419985449, 0.08060252900031628, 0.08204390999890165, 0.08199063700158149, 0.08200196200050414, 0.07995581399882212, 0.07898386200031382, 0.07995786500032409, 0.07851540499905241, 0.08106613299969467, 0.08262706700043054, 0.08050906599964947, 0.08211463699990418, 0.08045055999900796, 0.08318115800284431, 0.08109598399823881, 0.08307782300107647, 0.08200577499883366, 0.08667664799941122, 0.08240413400199031, 0.08122864200049662, 0.08454059499854338, 0.08353968899973552, 0.08333583800049382, 0.08321495500058518, 0.08131421600046451, 0.08144169499792042, 0.08254977500109817, 0.08642986400082009, 0.08568639499935671, 0.0835585640015779, 0.0802012169988302, 0.07777994299976854, 0.07988799099985044, 0.07912080599999172, 0.08099477699943236, 0.07904104499903042, 0.08049008800298907, 0.08280263299820945, 0.08368788000007044, 0.08312808199843857, 0.08200405400202726, 0.0822181069997896, 0.08667351600161055, 0.08559814799809828, 0.08213981399967452, 0.08594830299989553, 0.08567040500201983, 0.08173302199793397, 0.08201383499908843, 0.08266668100259267, 0.08447297000020626, 0.08223071399697801, 0.08207893300277647, 0.08168266699794913, 0.0826199950024602, 0.08504136199917411, 0.08409203600240289, 0.08066216899896972, 0.07980582199888886, 0.0799330240006384, 0.08381266500146012, 0.0814168199976848, 0.07854976700036786, 0.07797402700089151, 0.0801948900007119, 0.08146317399950931, 0.0831730019999668, 0.08262908699907712, 0.08030377300019609, 0.0823686080002517, 0.0791155320002872, 0.07892923599865753, 0.08284387300227536, 0.07979641099882429, 0.08107852100147284, 0.08100389399987762, 0.07945478900001035, 0.08090714099671459, 0.08041380100257811, 0.07985747400016407, 0.08146701599980588, 0.08004815399908694, 0.08017080699937651, 0.07977034500072477, 0.0819138059996476, 0.08236872899942682, 0.08321580000119866, 0.08333978500013473, 0.08147299600022961, 0.07955451099769562, 0.08100626800296595, 0.08203618599873153, 0.08113002399841207, 0.07868750600027852, 0.07913872200151673, 0.08027412999945227, 0.0809580349996395, 0.07993556600195006, 0.07946721699772752, 0.08080752900059451, 0.07951335400139214, 0.08004740300020785, 0.07942486799947801, 0.08042291200035834, 0.07898907199705718, 0.08058861500103376, 0.08104825900227297, 0.08112771499872906, 0.07895553799971822, 0.08008937499835156, 0.08315759100150899, 0.07984010400105035, 0.07991118599966285, 0.07983847400100785, 0.0800772729999153, 0.08065382199856685, 0.08057218300018576, 0.079761224998947, 0.0792718379998405, 0.08431816599841113, 0.0826130400018883, 0.0838836029979575, 0.0803783890005434, 0.08190861800176208, 0.08208630700028152, 0.08066570300070453, 0.07962025899905711, 0.07980437999867718, 0.08057743199969991, 0.0795388600017759, 0.08224823499767808, 0.08222593100072118, 0.08378998300031526, 0.08088736900026561, 0.08051190900005167, 0.08010219900097582, 0.08178430199768627, 0.08224057600091328, 0.0826979349985777, 0.08133502400232828, 0.08431205799934105, 0.08022969499870669, 0.0806536230011261, 0.08016037000197684, 0.07910826799707138, 0.0803749010010506, 0.07865734000006341, 0.08042184399891994, 0.08397538000281202, 0.0805309559982561, 0.08011141799943289, 0.08127847600189853, 0.08071292900058324, 0.08082925499911653, 0.08027007399869035, 0.08140800500041223, 0.08124413499899674, 0.07982272300068871, 0.07830316499894252, 0.0788955000025453, 0.07743503099845839, 0.07854925599895068, 0.07740353300323477, 0.078027954998106, 0.07830144300169195, 0.07839412499743048, 0.07889538400195306, 0.07957209699816303, 0.07906173300216324, 0.07934991599904606, 0.07915972599948873, 0.08076609799900325, 0.08243389900235343, 0.07958585899905302, 0.08215492800081847, 0.0786143359982816, 0.07938213699890184, 0.08144204900236218, 0.08025770200038096, 0.0810129040000902, 0.08738204999826849, 0.08328194900241215, 0.0899103500014462, 0.08737445299993851, 0.08424060700053815, 0.08524097400004393, 0.08280292299969005, 0.08454473899837467, 0.0795365199992375, 0.08041841900194413, 0.08327739800006384, 0.08183576100054779, 0.08193525900060195, 0.08050330899641267, 0.08397898600014742, 0.08148171400171123, 0.08046749699860811, 0.08131787800084567, 0.0825631740008248, 0.0806279110001924, 0.08182992699948954, 0.07974956999896676, 0.0798060439992696, 0.08060641100018984, 0.08056308200320927, 0.08034347599823377, 0.08202810700095142, 0.08123883299776935, 0.08029217700095614, 0.08048049599892693, 0.0849331310018897, 0.08217446799972095, 0.0795350119988143, 0.07891505100269569, 0.08004081899707671, 0.08157529100208194, 0.080306046998885, 0.08008373800112167, 0.07916193299752194, 0.07970300400120323, 0.08139491800102405, 0.08114439500059234, 0.08148575799714308, 0.08311262999995961, 0.08312660100273206, 0.08304894599859836, 0.08067626400224981, 0.07987127599699306, 0.07932421500299824, 0.08383102999869152, 0.0810573739981919, 0.08011853400239488, 0.08036924499901943, 0.0791581129997212, 0.08026161299858359, 0.08031306700286223, 0.08086750700022094, 0.08018900800016127, 0.0823558049996791, 0.08011431999693741, 0.08017183500123792, 0.0784261869994225, 0.08029931800047052, 0.0822193160020106, 0.08081573299932643]",tokens/s,12.283450812704455,, +float16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,6132.334592,,,,,s,1,90.84713750200171,90.84713750200171,0.0,90.84713750200171,90.84713750200171,90.84713750200171,90.84713750200171,[90.84713750200171],,kWh,0.0010448627633111328,0.0005694895313896468,0,0.0016143522947007795,,MB,6147.055616,,,,,s,10,8.072768113008351,0.8072768113008351,0.01385902257184664,0.8007011895006144,0.829717986601463,0.8317327158016269,0.833344499161758,"[0.7927945510018617, 0.8002993549998791, 0.7963087460011593, 0.8011030240013497, 0.7957770579996577, 0.8139620649999415, 0.796696580000571, 0.8128090200007136, 0.8292702690014266, 0.8337474450017908]",tokens/s,317.1155128158395,kWh,9.473915152244235e-06,5.163494840766349e-06,0.0,1.4637409993010584e-05,tokens/kWh,17489432.90665773,MB,6188.433408,,,,,s,10,53.7149477189887,5.3714947718988695,0.04258194419988249,5.369725200000175,5.424924088697298,5.42939343334765,5.432968909067931,"[5.326287449999654, 5.323044418997597, 5.32039523200001, 5.332673065000563, 5.4338627779980015, 5.359667009000987, 5.398701096997684, 5.379783390999364, 5.42393090099722, 5.416602376997616]",tokens/s,11.728578854732639,kWh,6.363549738532589e-05,3.4683819158336127e-05,0.0,9.8319316543662e-05,tokens/kWh,640769.3036802461,,s,630,53.707255257988436,0.08524961152061657,0.001960640013488116,0.08494156800043129,0.08686101159946702,0.08742505384998367,0.08882091418094205,"[0.08739604299989878, 0.08295146799719078, 0.0861445430018648, 0.08340475300065009, 0.08452410999962012, 0.08334187899890821, 0.08380206299989368, 0.08356838599866023, 0.08384912800102029, 0.08365385200158926, 0.08349663699846133, 0.08356656999967527, 0.08453640600055223, 0.08523240200156579, 0.08528943899727892, 0.08625086300162366, 0.08415058099853923, 0.08348421700065956, 0.08459205600229325, 0.08561749999716994, 0.08474178600226878, 0.085417389000213, 0.08501858099771198, 0.08347723000042606, 0.08517409900014172, 0.08473769600095693, 0.08348976799970842, 0.08402663200104143, 0.08365269599744352, 0.08475168600125471, 0.08435684700089041, 0.08479201899899635, 0.08497049799916567, 0.08579703900250024, 0.083531555999798, 0.08420031199784717, 0.08503054299944779, 0.08456968000245979, 0.08396531399921514, 0.08374051500140922, 0.08395625999764889, 0.08509176500228932, 0.08473786999820732, 0.08413565899900277, 0.0836625450028805, 0.08437009799672524, 0.08447512600105256, 0.08584112700191326, 0.08476495499780867, 0.0859762670006603, 0.08563942199907615, 0.08545720600159257, 0.08494240700019873, 0.08396770199760795, 0.08548586600227281, 0.08374107700001332, 0.08430495799984783, 0.08409452799969586, 0.08464002400069148, 0.08492111899977317, 0.08392974299931666, 0.08448514699921361, 0.08465275300113717, 0.08588808499916922, 0.086031599999842, 0.0851731100010511, 0.08465852299923426, 0.08405189299810445, 0.08484513900111779, 0.08460266800102545, 0.08422928100117133, 0.08311200099706184, 0.08366550000209827, 0.08348143799958052, 0.08267633099967497, 0.08307051000156207, 0.08386809699732112, 0.0834477770004014, 0.08591729000181658, 0.08703556199907325, 0.08350790300028166, 0.08567202699850895, 0.08491114999924321, 0.08461529100168264, 0.08455754199894727, 0.08392855000056443, 0.0843751530010195, 0.0833819030012819, 0.08577674299885985, 0.08428707200073404, 0.08442218799973489, 0.08416036599737708, 0.08321104000060586, 0.08380751900040195, 0.08495050700003048, 0.0836677459992643, 0.08418694099964341, 0.08416780800325796, 0.08394157599832397, 0.08470667399888043, 0.08378058500238694, 0.0847114079988387, 0.0852606510015903, 0.08412804499675985, 0.08389003200136358, 0.08398650099843508, 0.08406834700144827, 0.08429422400149633, 0.0860736270005873, 0.08780354499685927, 0.08508482900288072, 0.08607617399684386, 0.08435227200243389, 0.08428187399840681, 0.08400204899953678, 0.08401996199972928, 0.08286962800048059, 0.08463753500109306, 0.08524670699989656, 0.08392292500138865, 0.0847239559989248, 0.08494072900066385, 0.08446197100056452, 0.08492463899892755, 0.08527467699968838, 0.0836169920003158, 0.08598274500036496, 0.084880673999578, 0.08469388299999991, 0.08458530499774497, 0.0834452240014798, 0.08542315399972722, 0.08402466400002595, 0.08312726800068049, 0.08408310699815047, 0.08331761800218374, 0.08324500099843135, 0.08384327499879873, 0.08340663600029075, 0.08360173300025053, 0.08310700900256052, 0.08346741499917698, 0.08368826199875912, 0.08413266900242888, 0.08407439799702843, 0.08394507200137014, 0.08484145400143461, 0.08464441199976136, 0.0840624550000939, 0.08490187699862872, 0.08538035899982788, 0.08516875300119864, 0.08408184399741003, 0.08527035100269131, 0.08367466199706541, 0.08447144100136939, 0.08467495299919392, 0.08409685900187469, 0.08493765399907716, 0.08478334400206222, 0.08401996299653547, 0.08397763500033761, 0.08410438999999315, 0.08484239199970034, 0.08463233400107129, 0.08377087999906507, 0.0835057210024388, 0.08401747999960207, 0.0840683469978103, 0.08381746400118573, 0.0846620560005249, 0.08472730900030001, 0.08342629799881252, 0.08397033300207113, 0.08363040199765237, 0.08620587900077226, 0.08619838099912158, 0.08430289800162427, 0.08630718099811929, 0.08464190300219343, 0.08558937999987393, 0.08403123199968832, 0.08475818900114973, 0.08391350999954739, 0.0850873329982278, 0.08509741200032295, 0.08539479600221966, 0.0840312559994345, 0.08777449899935164, 0.08557094500065432, 0.08349974699740415, 0.08254025400310638, 0.08344378199763014, 0.08354318599958788, 0.08416274800038082, 0.08324937100042007, 0.08384287299850257, 0.08342762900065281, 0.08453308000025572, 0.08443926200197893, 0.08480587200028822, 0.08478736899996875, 0.08570363899707445, 0.08505516200239072, 0.0844831369977328, 0.08420412400300847, 0.0840929009973479, 0.08600204799950006, 0.0854141250019893, 0.08456782799839857, 0.0872741950006457, 0.08542922400010866, 0.0843222940020496, 0.08365195599981234, 0.08404119899932994, 0.083770254997944, 0.08515595700009726, 0.08661103499980527, 0.0858450570012792, 0.08661414699963643, 0.08573464499932015, 0.08476104700093856, 0.08358876499914913, 0.08477130599931115, 0.08670829200127628, 0.08323040200048126, 0.08370490000015707, 0.08514312300030724, 0.08380551200025366, 0.08419747700099833, 0.08430700899771182, 0.08432798600188107, 0.08454363299824763, 0.08466186599980574, 0.08687073699911707, 0.08413666700289468, 0.08428014399760286, 0.08634234800047125, 0.08547322300000815, 0.08428912499948638, 0.08413762500276789, 0.085729521997564, 0.08445712300090236, 0.0849270639992028, 0.08363128500059247, 0.08459883899922715, 0.08365692699953797, 0.08472320800137823, 0.08407905499916524, 0.08440860500195413, 0.08424740599730285, 0.08422458499990171, 0.08554574899972067, 0.082836818000942, 0.08377354999902309, 0.08260834599786904, 0.08392506900054286, 0.08373437400223338, 0.08499804899838637, 0.08497941500172601, 0.08503933099927963, 0.08493795899994439, 0.08401561899881926, 0.08423983000102453, 0.08376722100001643, 0.08377359799851547, 0.0848911280008906, 0.08454128399898764, 0.08664759200109984, 0.08682411699919612, 0.08616867499949876, 0.08531814100206248, 0.08512558299844386, 0.0861805300010019, 0.08501636199798668, 0.0840884110002662, 0.08492660000047181, 0.08425648400225327, 0.08600716299770284, 0.08610086800035788, 0.08507140499932575, 0.08501205100037623, 0.0841433710011188, 0.08543430899953819, 0.08538333799879183, 0.08745521700257086, 0.08721339899784653, 0.08734583900150028, 0.08626305000143475, 0.08531334899817011, 0.0844052139982523, 0.08430193200183567, 0.08312543799911509, 0.08488502399995923, 0.08481966000181274, 0.0845594530001108, 0.08522122800059151, 0.08461366199844633, 0.08486545599953388, 0.08467337600086466, 0.08556550499997684, 0.08507330299835303, 0.08561437000025762, 0.10035489700021571, 0.1094269000022905, 0.1109549100001459, 0.09389905099669704, 0.0854438670030504, 0.08617415499975323, 0.08541870199769619, 0.08502447100181598, 0.08560046099955798, 0.08561782299875631, 0.08527420999962487, 0.0852226540009724, 0.08557012100209249, 0.08467180699881283, 0.08540757700029644, 0.08600977099922602, 0.08499496700096643, 0.0858302630003891, 0.08572117500079912, 0.08888575499804574, 0.08493440700112842, 0.08483535699997446, 0.08605829199950676, 0.08604790399840567, 0.0861507879999408, 0.08553642900005798, 0.08488218999991659, 0.08588121800130466, 0.0853023720010242, 0.0842984279988741, 0.08411127300132648, 0.08388873399962904, 0.0842647629979183, 0.0849975130004168, 0.08482637999986764, 0.08604333500261419, 0.08587725200050045, 0.08827559799829032, 0.08581837800011272, 0.08474324100097874, 0.08516393899844843, 0.08413807200122392, 0.08339314499971806, 0.0841760980001709, 0.08425272599924938, 0.08381703100167215, 0.08454640099807875, 0.08506812600171543, 0.0850145619988325, 0.08511833000011393, 0.08467638799993438, 0.08465268899817602, 0.08438142000159132, 0.08422570699985954, 0.08470140400095261, 0.0839191960003518, 0.08557176400063327, 0.08470928299720981, 0.08584673299992573, 0.08368818999952055, 0.08536389900109498, 0.08396376300152042, 0.0846141189977061, 0.08570491000136826, 0.08469888399849879, 0.08377298699997482, 0.08455598400178133, 0.08488119899993762, 0.08490452499972889, 0.08526445800089277, 0.08554232799724559, 0.08701166800165083, 0.08479008699941915, 0.08444594699903973, 0.08438431800095714, 0.08643529999972088, 0.08416034800029593, 0.08420272799776285, 0.08418472300036228, 0.08448904900069465, 0.08592770000177552, 0.08456566499808105, 0.08750698800213286, 0.08450451899989275, 0.08513559499988332, 0.0844975969994266, 0.08520603199940524, 0.08447878199876868, 0.08405713899992406, 0.08511520799947903, 0.08557648500209325, 0.0844356230008998, 0.08447655199779547, 0.08390788600081578, 0.08528481300163548, 0.08430228099678061, 0.08478792300229543, 0.0847138270000869, 0.08497872799853212, 0.0847989329995471, 0.0848877230018843, 0.0858747050006059, 0.0850785589973384, 0.08686731700072414, 0.0876891309999337, 0.08599966300243977, 0.08675896999920951, 0.08685714699822711, 0.08468587800234673, 0.0844822139988537, 0.0844578129981528, 0.08695242300018435, 0.08693958200092311, 0.08860107800137484, 0.0851840719988104, 0.08679915299944696, 0.08476022399918293, 0.08876201400198624, 0.08652740699835704, 0.08559418500226457, 0.08446050199927413, 0.08486175999860279, 0.08532927900159848, 0.08685558700017282, 0.08502158799819881, 0.08628162300010445, 0.0858778190013254, 0.08731497999906424, 0.08485638899946935, 0.086905159001617, 0.08589382900026976, 0.08744879000005312, 0.08679162800035556, 0.08735793300002115, 0.08765870999923209, 0.08779229200081318, 0.0869887929984543, 0.08476195199909853, 0.08650436099924264, 0.08469685999807552, 0.08576136600095197, 0.08490359600182273, 0.08472186699873419, 0.08388218999971286, 0.08501477100071497, 0.0846658819973527, 0.08454771700053243, 0.08517004300301778, 0.08513665599821252, 0.08568701200056239, 0.08570216899897787, 0.08540571200137492, 0.08507753099911497, 0.08429890299885301, 0.08642569000221556, 0.08483997199800797, 0.08537240200166707, 0.08528505999856861, 0.08652179500131751, 0.08566974600034882, 0.08508525999786798, 0.08567214700087789, 0.08636846299850731, 0.08596676200249931, 0.08518232699861983, 0.08658486000058474, 0.08646928700181888, 0.08698765599910985, 0.08649995800078614, 0.08480400999906124, 0.08559564199822489, 0.08597064200148452, 0.08527393300028052, 0.08550440600083675, 0.08468663199892035, 0.08474202399884234, 0.0860650060021726, 0.08587682899815263, 0.08465481900202576, 0.08520800299811526, 0.08431787800145685, 0.08451428699845565, 0.08517351100090309, 0.08544124699983513, 0.08675759000107064, 0.08578717399723246, 0.08579823200125247, 0.08583008099958533, 0.08588918000168633, 0.08867089599880273, 0.08604656800162047, 0.08487707399763167, 0.08567836300062481, 0.08471518900114461, 0.08553731899883132, 0.08444567900005495, 0.08439929200176266, 0.08408888599660713, 0.08387239000148838, 0.0842868690015166, 0.08448115699866321, 0.08838370999728795, 0.0864607170005911, 0.08479758300018148, 0.08469502699881559, 0.08531074600250577, 0.08628967299955548, 0.08670784100104356, 0.08583583899962832, 0.08884497200051555, 0.08580692999748862, 0.08632393200241495, 0.08505307700033882, 0.08473874799892656, 0.08361036099813646, 0.08399533299962059, 0.08417467500112252, 0.08526552200055448, 0.0848360429990862, 0.0851289740021457, 0.08592036299887695, 0.08607439000115846, 0.08724840699869674, 0.08655305999855045, 0.08636162600305397, 0.08686031099932734, 0.08521010699769249, 0.0841520110006968, 0.08549035900068702, 0.08613997399879736, 0.0850102889999107, 0.08619201400142629, 0.08627459299896145, 0.08406551000007312, 0.08427256499999203, 0.08410645999902044, 0.08444025600329041, 0.08475811399694066, 0.08620376399994711, 0.08795844400083297, 0.08737284500239184, 0.08694790799927432, 0.08779372500066529, 0.0863728109980002, 0.08537170500130742, 0.0857967349984392, 0.08635387300091679, 0.08805585699883522, 0.08545662400138099, 0.08641266899940092, 0.08714205900105298, 0.08773483199911425, 0.08631400100057363, 0.08705420299884281, 0.08568793700032984, 0.0869789780008432, 0.0869062159981695, 0.0844751440017717, 0.08662418399762828, 0.08772725600283593, 0.08885720499893068, 0.0887613170016266, 0.08759703999749036, 0.08594062100019073, 0.0872584739990998, 0.08557346000088728, 0.08634366099795443, 0.08689042300102301, 0.08670183600042947, 0.08714973400128656, 0.08582229999956326, 0.08558058400012669, 0.08725117199719534, 0.08616658900064067, 0.08628817600038019, 0.087097317998996, 0.08590830300090602, 0.08692191000227467, 0.08709619199726149, 0.08864032900237362, 0.08691777599960915, 0.08546469299835735, 0.08791307900173706, 0.08795460699911928, 0.08666495999932522, 0.08581447400138131, 0.08680151799853775, 0.08650661900173873, 0.08590531499794452, 0.08650952600146411, 0.08537659899957362, 0.08623224699840648, 0.08497882400115486, 0.08484918199974345, 0.0856094100017799, 0.0841474869994272, 0.08470275199942989, 0.08535509400098817, 0.08413539799948921, 0.08437310199951753, 0.08549775099891121, 0.08581742699971073, 0.0855849700019462, 0.08510280699920258, 0.08484248199965805, 0.08533266100130277, 0.08458954699744936, 0.08496884000123828, 0.08641789000103017, 0.08618264799952158, 0.08513373499954469, 0.08634789600182557, 0.08535976199709694, 0.08561702100269031, 0.08523856099782279, 0.08765794000282767, 0.08821514599912916, 0.08702843799983384, 0.08609520299796714, 0.08515851100310101, 0.0858958819990221, 0.08498998199866037, 0.08512125600100262, 0.085435258999496, 0.08484941899951082, 0.08522696599902702, 0.08509543300169753]",tokens/s,11.730258732711789,, +bfloat16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cpu,,42,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,798.892032,,,,,s,1,8.25673309100057,8.25673309100057,0.0,8.25673309100057,8.25673309100057,8.25673309100057,8.25673309100057,[8.25673309100057],,kWh,6.981589805485302e-05,3.8051053237441425e-05,0,0.00010786695129229445,,MB,835.100672,,,,,s,15,0.980113280003934,0.0653408853335956,0.0004304203264354521,0.06535554299989599,0.06594069399907312,0.06613562309976259,0.06622356382002181,"[0.06624554900008661, 0.06526084600045579, 0.06537002400000347, 0.06549612800154136, 0.06531217800147715, 0.06487396100055776, 0.06497835099980875, 0.06608851199962373, 0.0657189669982472, 0.06535554299989599, 0.0654226100014057, 0.06541133800055832, 0.06453260600028443, 0.06509452599857468, 0.06495214100141311]",tokens/s,3917.9144679935234,kWh,7.947336912191851e-07,4.3314892811933023e-07,0.0,1.2278826193385153e-06,tokens/kWh,208488984.1814947,MB,845.55776,,,,,s,15,9.69449415899544,0.646299610599696,0.004648820202933521,0.645305369000198,0.6502607877984701,0.6528931953989741,0.6577738350804794,"[0.6473840580001706, 0.643205524998848, 0.6502341189989238, 0.6480869579991122, 0.64320374899944, 0.645305369000198, 0.6589939950008556, 0.6479464340009145, 0.6468643740008702, 0.6502785669981677, 0.6451608509996731, 0.644981885998277, 0.6406475260009756, 0.6438500900003419, 0.6383506579986715]",tokens/s,97.47801014693917,kWh,7.646574295335924e-06,4.167550876959612e-06,0.0,1.1814125172295539e-05,tokens/kWh,5332599.6704129055,,s,945,9.686770006996085,0.010250550271953528,0.00030107754393953324,0.01021268999829772,0.010527285199714242,0.010674218198982998,0.01143063995958073,"[0.012614911000127904, 0.010488073001397424, 0.010506089000045904, 0.010183323998717242, 0.010228111999822431, 0.009973854999770992, 0.010181717001614743, 0.010104883998792502, 0.010465457000464085, 0.010087819000546006, 0.010330246999728843, 0.010187489999225363, 0.010302475000571576, 0.010060483000415843, 0.010206827999354573, 0.009980635999454535, 0.010187216001213528, 0.010411506998934783, 0.010117768000782235, 0.010053935999167152, 0.010288319999745, 0.01014727600158949, 0.010214314999757335, 0.010238058999675559, 0.010211527000137721, 0.009993055999075295, 0.010248868000417133, 0.010099471001012716, 0.010271145998558495, 0.010202837000178988, 0.010524876999625121, 0.010224276000371901, 0.01050430299983418, 0.010317033000319498, 0.010209285001110402, 0.010168150998651981, 0.009982535000744974, 0.01030610799898568, 0.010224382000160404, 0.010299610001311521, 0.01021540100009588, 0.010244957999020698, 0.010524118999455823, 0.010406050001620315, 0.010282602999723167, 0.010684417999073048, 0.010036952000518795, 0.009976257999369409, 0.010110426001119777, 0.010199471998930676, 0.010117324000020744, 0.010429370000565541, 0.01032854199911526, 0.010476531000676914, 0.010093610000694753, 0.010245436998957302, 0.010032063000835478, 0.010385820998635609, 0.010118312000486185, 0.01022116900094261, 0.010109212000315893, 0.010056853998321458, 0.010220299000138766, 0.011798974999692291, 0.010182697000345797, 0.010248237998894183, 0.009916389000864001, 0.009935231999406824, 0.009805231000427739, 0.01007768700037559, 0.009885810999548994, 0.009851458999037277, 0.009678456000983715, 0.009934337000231608, 0.0101829129998805, 0.010181226998611237, 0.010107228001288604, 0.010213711999313091, 0.010151598000447848, 0.010028005999629386, 0.009954130000551231, 0.010078924999106675, 0.00992123600008199, 0.010109333001310006, 0.009946987998773693, 0.010095904999616323, 0.010307103000741336, 0.010029154000221752, 0.010266231000059634, 0.01023435199931555, 0.010387991000243346, 0.010184202999880654, 0.010676152000087313, 0.010219908999715699, 0.010159933000977617, 0.010177614998610807, 0.010146153001187486, 0.010268960999383125, 0.010328115000447724, 0.010402964000604698, 0.010246247999020852, 0.01019202600036806, 0.010464598999533337, 0.01036925100015651, 0.010301730999344727, 0.010396009000032791, 0.010228187000393518, 0.010333616000934853, 0.01031623699964257, 0.010251469000650104, 0.010285069000019575, 0.010485690998393693, 0.010456478999913088, 0.010348955000154092, 0.01015333100076532, 0.010304339999493095, 0.010128068999620155, 0.010336417000871734, 0.00999861700074689, 0.010252518999550375, 0.00995897199936735, 0.010234147999653942, 0.01007098300033249, 0.010148219000257086, 0.01006372600022587, 0.010317246998965857, 0.01171000899921637, 0.010537039999690023, 0.010185386001467123, 0.010357165998357232, 0.010128577001523809, 0.010154102999877068, 0.010307292999641504, 0.010004939998907503, 0.010277050001604948, 0.010004046998801641, 0.010389990000476246, 0.010284604999469593, 0.01079478400060907, 0.010166355999899679, 0.009920264999891515, 0.010084527999424608, 0.010645435000697034, 0.010499658999833628, 0.010605326000586501, 0.010184664999542292, 0.01036285800000769, 0.010664796000128263, 0.010790062000523903, 0.011076083999796538, 0.010399124999821652, 0.010418677999041392, 0.010192223000558442, 0.010242757000014535, 0.010047891999420244, 0.0101988110000093, 0.010135630000149831, 0.009923962999891955, 0.010232155000267085, 0.009914169000694528, 0.010182472999076708, 0.010931164000794524, 0.013365945000259671, 0.010078544999487349, 0.010009996000007959, 0.010153429000638425, 0.010035772000264842, 0.010229097999399528, 0.010307393999028136, 0.010154057999898214, 0.009970569000870455, 0.009966170000552665, 0.010146085000087623, 0.010090984998896602, 0.010321637000743067, 0.01018628599922522, 0.009989960999519099, 0.010190482000325574, 0.010033819000454969, 0.01018846199986001, 0.010577068000202416, 0.010085108000566834, 0.01023166599952674, 0.009870148000118206, 0.010343407000618754, 0.009979264999856241, 0.010070379999888246, 0.009981076998883509, 0.010101589999976568, 0.011561911998796859, 0.010438737999720615, 0.01035888800106477, 0.010209473999566399, 0.010236415000690613, 0.010095550000187359, 0.009975716999178985, 0.010108169000886846, 0.010299483999915537, 0.010175388999414281, 0.010184503000346012, 0.010223429999314249, 0.01015222400019411, 0.010035114999482175, 0.010328521000701585, 0.010062445000585285, 0.011001815999406972, 0.010557968998909928, 0.01055394300055923, 0.010294678000718704, 0.010556045999692287, 0.010064091999083757, 0.010258744001475861, 0.010161453999899095, 0.010289552999893203, 0.010208831999989343, 0.010348900999815669, 0.010387075999460649, 0.010368142000515945, 0.010527852000450366, 0.010278364999976475, 0.010262026999043883, 0.010286475000611972, 0.010052637999251601, 0.010401813000498805, 0.010111699000844965, 0.010236032998363953, 0.010081940999953076, 0.010277811001287773, 0.010045973000160302, 0.010305484998752945, 0.010833693000677158, 0.010530157000175677, 0.010100145998876542, 0.010355459000493283, 0.01021116400079336, 0.010642925999491126, 0.010152176999326912, 0.010227718001260655, 0.010163744998862967, 0.009985156000766438, 0.010000996999224299, 0.01011384400044335, 0.01022768700022425, 0.010202837000178988, 0.010147646000405075, 0.010225725998679991, 0.010090572000990505, 0.01026135800020711, 0.010080840998853091, 0.010385061001215945, 0.010018801998739946, 0.010278823001499404, 0.011338454998622183, 0.010161087000597036, 0.009905945000355132, 0.010323548000087612, 0.009751038000104018, 0.010289459998602979, 0.010231410000415053, 0.010100663001139765, 0.010442629998578923, 0.010440085001391708, 0.01009553599942592, 0.010115380999195622, 0.010042571000667522, 0.009876338999674772, 0.01003679700079374, 0.010027337999417796, 0.010226963999230065, 0.009963743001208059, 0.010207172999798786, 0.010171911999350414, 0.010043293999842717, 0.01027707899993402, 0.00997393500074395, 0.010274661999574164, 0.009975189001124818, 0.010385150999354664, 0.010259160000714473, 0.010006802000134485, 0.010060128999612061, 0.01064099100040039, 0.010438236999107176, 0.010051498000393622, 0.010283948999131098, 0.010049321999758831, 0.010297991000697948, 0.010138814999663737, 0.01021537199994782, 0.010257234000164317, 0.010535212000831962, 0.010291741999026272, 0.010171019999688724, 0.010803002000102424, 0.010089428000355838, 0.010221889000604278, 0.010208020999925793, 0.009976977000405896, 0.010336797999116243, 0.009937171000274247, 0.010409765000076732, 0.009942237998984638, 0.010378763001426705, 0.00995121899904916, 0.010231033000309253, 0.01000113600093755, 0.010593793998850742, 0.010051557001133915, 0.010494384998310125, 0.01002994500049681, 0.010159143001146731, 0.010118109999893932, 0.010459007999088499, 0.00998046700078703, 0.00996189199941, 0.011496610999529366, 0.010222971000985126, 0.010280344999046065, 0.010428094999952009, 0.010276388000420411, 0.010319764000087162, 0.010120008999365382, 0.010156157000892563, 0.010029024999312242, 0.01025732600101037, 0.01023738699950627, 0.0101156330001686, 0.010301214999344666, 0.010137139001017204, 0.010204768999756197, 0.01044112099953054, 0.010251526000502054, 0.010430031999931089, 0.010346983999625081, 0.01060501500069222, 0.010398029999123537, 0.010510321999390726, 0.010398811000413843, 0.010540340001170989, 0.010056535998955951, 0.010439838999445783, 0.01008580300003814, 0.010255629000312183, 0.009932462000506348, 0.01026331899993238, 0.012240160000146716, 0.0101519849995384, 0.009990316000767052, 0.009980929999073851, 0.009985538001274108, 0.009949022000000696, 0.009984913000153028, 0.009945830999640748, 0.01011191599900485, 0.009939530000337982, 0.00999401300032332, 0.00999727599992184, 0.009994381000069552, 0.009879122000711504, 0.010023338998507825, 0.010088205001011374, 0.01002058899939584, 0.010058113999548368, 0.010078958001031424, 0.010052809999251622, 0.010066728000310832, 0.010214743000688031, 0.00992709499951161, 0.01031472499926167, 0.009967449999749078, 0.010044757000287063, 0.00996621000012965, 0.01028666000092926, 0.011077092998675653, 0.010422266001114622, 0.010407458999907249, 0.010002719000112847, 0.01010313199913071, 0.011342685000272468, 0.010383775999798672, 0.010240311001325608, 0.011017064998668502, 0.010232669999822974, 0.010480631000973517, 0.01040821400056302, 0.010675140998500865, 0.011080815000241273, 0.010534465000091586, 0.010351016000640811, 0.010197823999988032, 0.010186203000557725, 0.010437353999805055, 0.010266440998748294, 0.010272576000716072, 0.010445362000609748, 0.010257093999825884, 0.0103466299988213, 0.010732230000940035, 0.010198945999945863, 0.010369538998929784, 0.0104141490010079, 0.010035175999291823, 0.010636610000801738, 0.010725574999014498, 0.01027157300086401, 0.010815722000188543, 0.010319421000531293, 0.010852961999262334, 0.010473856000317028, 0.010921917999439756, 0.0105077620009979, 0.010514473999137408, 0.00995207300002221, 0.010561126999164117, 0.010079685000164318, 0.01033038400055375, 0.010129128999324166, 0.01044269600060943, 0.01021672199931345, 0.010760226999991573, 0.010413805001007859, 0.010400070999821764, 0.010324574999685865, 0.010377923001215095, 0.010430225998788956, 0.010704590000386816, 0.010635000999172917, 0.01061535500048194, 0.010453480001160642, 0.010621195999192423, 0.010579497000435367, 0.010386154999650898, 0.010294723999322741, 0.010447618000398506, 0.010185360999457771, 0.01044047700088413, 0.01015023000036308, 0.010091168000144535, 0.010332658999686828, 0.010968509999656817, 0.010222958000667859, 0.011495382999783033, 0.010417052000775584, 0.010321354999177856, 0.010287714001606219, 0.010235975998512004, 0.010143198000150733, 0.010133527001016773, 0.010182578000240028, 0.010025628998846514, 0.010119821001353557, 0.010041502999229124, 0.010349259000577149, 0.01027384400003939, 0.009936315998857026, 0.010198637000939925, 0.010441313999763224, 0.01042484599929594, 0.010384342000179458, 0.010209957999904873, 0.01031331699959992, 0.010403486001450801, 0.010471058998518856, 0.010404475000541424, 0.010168979999434669, 0.01028198400126712, 0.010362442999394261, 0.010441013000672683, 0.010329662000003736, 0.01029334899976675, 0.010313171998859616, 0.010297104001438129, 0.010342486999434186, 0.010346887000196148, 0.009919600999637623, 0.010216741000476759, 0.010442245999001898, 0.010328477999792085, 0.01024313600100868, 0.01035118700019666, 0.01006250899990846, 0.010477320000063628, 0.010277998999299598, 0.01042913099990983, 0.010501775001102942, 0.011018143999535823, 0.010398930000519613, 0.010284320998835028, 0.01024640899959195, 0.010321671001292998, 0.010232364998955745, 0.010164588000407093, 0.01021797700013849, 0.010101202000441845, 0.010080033000122057, 0.0100914089998696, 0.010006581998823094, 0.009940042000380345, 0.009923759000230348, 0.009988478999730432, 0.010100159001012798, 0.009955853000064963, 0.010277719999066903, 0.010461740999744507, 0.0113041719996545, 0.010398440001154086, 0.010510326999792596, 0.010466301999258576, 0.01044157900105347, 0.01044367799840984, 0.010276547000103164, 0.010191689001658233, 0.010003111998230452, 0.010201018001680495, 0.010149386998818954, 0.00989065199973993, 0.010181843001191737, 0.010078963001433294, 0.010119770999153843, 0.010278518999257358, 0.010494782000023406, 0.010103888998855837, 0.010154364001209615, 0.010061583001515828, 0.010029637996922247, 0.009801947002415545, 0.010051489000034053, 0.010205226997641148, 0.009995691001677187, 0.01028921300166985, 0.0103101679997053, 0.010199426000326639, 0.009963535998394946, 0.010339327000110643, 0.01024870000037481, 0.010080981999635696, 0.010543274998781271, 0.010299510002369061, 0.010225996000372106, 0.010268083999108057, 0.010106197998538846, 0.010284428000886692, 0.010069026000564918, 0.010022241000115173, 0.010044900998764206, 0.010000302001571981, 0.010472147998370929, 0.010282076000294182, 0.010128604000783525, 0.010185490998992464, 0.010191943998506758, 0.010233283002889948, 0.010229634997813264, 0.01058562000253005, 0.01021268999829772, 0.01048644399998011, 0.010293055998772616, 0.010645031001331517, 0.010364376001234632, 0.010554077998676803, 0.010222589000477456, 0.010854453998035751, 0.010392094001872465, 0.010292652001226088, 0.010290488997270586, 0.010157303000596585, 0.010114929002156714, 0.011375803998816991, 0.011062361001677345, 0.010221633998298785, 0.010226168000372127, 0.010031009998783702, 0.010151097001653397, 0.0101860430004308, 0.010138800000277115, 0.0101550289982697, 0.010252773001411697, 0.009975483000744134, 0.01029880799978855, 0.01032622799903038, 0.01025561300048139, 0.010486011997272726, 0.01026313300098991, 0.010465580999152735, 0.010336020000977442, 0.010358879000705201, 0.010530294999625767, 0.010580312999081798, 0.010391039002570324, 0.010081218999403063, 0.010049402000731789, 0.009953359996870859, 0.010175612002058187, 0.01041554699986591, 0.01009884800077998, 0.010422176997963106, 0.010236823000013828, 0.01053102700097952, 0.010284072999638738, 0.010619561999192229, 0.010543436001171358, 0.010314219998690533, 0.010471335001057014, 0.01057691700043506, 0.01075048699931358, 0.010370889998739585, 0.010607657000946347, 0.010403763000795152, 0.010512340999412118, 0.010258105001412332, 0.010408239999378566, 0.010259520000545308, 0.010258435999276116, 0.01024995899933856, 0.01045910399989225, 0.010200327000347897, 0.010118540001712972, 0.010079305997351184, 0.010216423001111252, 0.010339780001231702, 0.010563911997451214, 0.010257173002173658, 0.01028016599957482, 0.009967894999135751, 0.010199461001320742, 0.010120766997715691, 0.00995010300175636, 0.010121459999936633, 0.009920076001435518, 0.010040741999546299, 0.011403119999158662, 0.010258943999360781, 0.010355843998695491, 0.010440134999953443, 0.010016394000558648, 0.010024491002695868, 0.009951970998372417, 0.010081858999910764, 0.0099448780019884, 0.010245035999105312, 0.01026743599868496, 0.009826944999076659, 0.010294869000063045, 0.009787062001123559, 0.010059298001579009, 0.009781125998415519, 0.010111472998687532, 0.00995292300285655, 0.009816540998144774, 0.009706683002150385, 0.009738325999933295, 0.009875792999082478, 0.01008825499957311, 0.010143609000806464, 0.010297945998900104, 0.010336472001654329, 0.010311384998203721, 0.010477920000994345, 0.01005778100079624, 0.010299937999661779, 0.01026757499857922, 0.009967948000848992, 0.010243578999507008, 0.010097967999172397, 0.010416820001410088, 0.010112641997693572, 0.010428794001200004, 0.010148605000722455, 0.010460405999765499, 0.0103937899984885, 0.010618419000820722, 0.010328939999453723, 0.010375662001024466, 0.010435158001200762, 0.010472474998096004, 0.010196618000918534, 0.010323758000595262, 0.01047719700000016, 0.010433927000121912, 0.010382876000221586, 0.010618748998240335, 0.010251986001094338, 0.010341150999011006, 0.010537465001107194, 0.01018441099949996, 0.010492118999536615, 0.009964534001483116, 0.010405114997411147, 0.010347778999857837, 0.01019558600091841, 0.010264533000736265, 0.0101221170007193, 0.010389087998191826, 0.011436487999162637, 0.010488387000805233, 0.010436067001137417, 0.01025043600020581, 0.010222452998277731, 0.010529402999964077, 0.010367331000452396, 0.010110754999914207, 0.010057868999865605, 0.009875013998680515, 0.010480886001460021, 0.01039961599963135, 0.01034369600165519, 0.010570263999397866, 0.0103932359997998, 0.010392044998297933, 0.010100337000039872, 0.009931653999956325, 0.010007480999774998, 0.010290650003298651, 0.010557541998423403, 0.010248271999444114, 0.010276879998855293, 0.010001546001149109, 0.010147083001356805, 0.01018769399888697, 0.010141208000277402, 0.010253806998662185, 0.01017248300195206, 0.01040208199992776, 0.010311054000339936, 0.010131497998372652, 0.010184437000134494, 0.010071605000121053, 0.010330466000596061, 0.010166059000766836, 0.010343399000703357, 0.010070757998619229, 0.010442383998451987, 0.010051602999737952, 0.010463736001838697, 0.01004395400013891, 0.01019357200129889, 0.010186239996983204, 0.009979612001188798, 0.009907283001666656, 0.009744405997480499, 0.010254135999275604, 0.009947061000275426, 0.009887390002404572, 0.010007778997533023, 0.010137155000847997, 0.009981326998968143, 0.010247795002214843, 0.010343009998905472, 0.010683623000659281, 0.010153292998438701, 0.010363667002820875, 0.010082764998514904, 0.01014043900067918, 0.010005124000599608, 0.01019238899971242, 0.01029968099828693, 0.01142319700011285, 0.010822050000570016, 0.010085834001074545, 0.0100901209989388, 0.00994226799957687, 0.010174404000281356, 0.010848992998944595, 0.010115249002410565, 0.010155518000829034, 0.010003990999393864, 0.010032663998572389, 0.010100281000632094, 0.010004864001530223, 0.010153155999432784, 0.010023120998084778, 0.010003829000197584, 0.009862856000836473, 0.009988671001337934, 0.0100230929965619, 0.010011846003180835, 0.009971159997803625, 0.010141261001990642, 0.009977402998629259, 0.010135289001482306, 0.00997768999877735, 0.010083961999043822, 0.010203670000919374, 0.010304047998943133, 0.010349177002353827, 0.010179518998484127, 0.010526434998610057, 0.01011032800306566, 0.010460721998242661, 0.01019016600184841, 0.010365807997004595, 0.009887208001600811, 0.010670527000911534, 0.01008033499965677, 0.01016499199977261, 0.01000231500074733, 0.010006385997257894, 0.009936659000231884, 0.009989219000999583, 0.010018494998803362, 0.009949290000804467, 0.010045087001344655, 0.009913666999636916, 0.009978515998227522, 0.009931668002536753, 0.010115470999153331, 0.009941477997926995, 0.00997765200008871, 0.00986326500060386, 0.010270448001392651, 0.010668012000678573, 0.01095195499874535, 0.010505950001970632, 0.010307487998943543, 0.010130328999366611, 0.009993331001169281, 0.00987909600007697, 0.01010918199972366, 0.010024961000453914, 0.011447565000707982, 0.010354877998906886, 0.011166018000949407, 0.010138818000996253, 0.010240256997349206, 0.010047724001196912, 0.01050289499835344, 0.01013411700114375, 0.010273713000060525, 0.010157265998714138, 0.010134841002582107, 0.010147201999643585, 0.010258417998556979, 0.010263563999615144, 0.01034822000292479, 0.010305707997758873, 0.010163939001358813, 0.009984692998841638, 0.01000207600009162, 0.01001061200076947, 0.01012785700004315, 0.009985994998714887, 0.010031204001279548, 0.010000056998251239, 0.010130116999789607, 0.010156164000363788, 0.010279774000082398, 0.009918899002514081, 0.010017953998612938, 0.010013199000241002, 0.010226640999462688, 0.009988752000936074, 0.01006871199933812, 0.010601761998259462, 0.010375232002843404, 0.010735527997894678, 0.010295947002305184, 0.010289624999131775, 0.010213503999693785, 0.010310852001566673, 0.010179032997257309, 0.01039501600098447, 0.010374530000262894, 0.010116179000760894, 0.010184263999690302, 0.00993653800105676, 0.010162103997572558, 0.010241479001706466, 0.010284031999617582, 0.010266368997690734, 0.010154749001230812, 0.01021984299950418, 0.010095120000187308, 0.010458124001161195, 0.009892645000945777, 0.010299887999281054, 0.009810180999920703, 0.010166848998778732, 0.01009318400247139, 0.010164250998059288, 0.009884271999908378, 0.00997526599894627, 0.010145999000087613, 0.011201712000911357, 0.01050422599655576, 0.010271894003381021, 0.009985507997043896, 0.009917627001414075, 0.010593538001558045, 0.010052825997263426, 0.009912075001921039, 0.009949594998033717, 0.010042579000582919, 0.010137394001503708, 0.009923833997163456, 0.009798334001970943, 0.009906144001433859, 0.01017601799685508, 0.010226734000752913, 0.010176824001973728, 0.01002331299969228, 0.009963882999727502, 0.010148189998290036, 0.01007715300147538, 0.009794023000722518, 0.0100745999989158, 0.009662045998993563, 0.010088005001307465, 0.009947586000635056, 0.010188243999436963, 0.010074159999930998, 0.010025561998190824, 0.009970008002710529, 0.010023408998677041, 0.009990285001549637, 0.01008602199726738, 0.010088758001074893, 0.009852058999967994, 0.010574566000286723, 0.009998003999498906, 0.010098205999383936, 0.010012881000875495, 0.010101424999447772, 0.009955772999092005, 0.009845259002759121, 0.01014934699924197, 0.01005554799849051, 0.010164621002331842, 0.010357103998103412, 0.010635319998982595, 0.010119203001522692, 0.010220423999271588, 0.010071910000988282, 0.010155315001611598, 0.010061119999591028, 0.010213354998995783, 0.010203262998402352, 0.010113160002219956, 0.010269736998452572, 0.0100755540006503, 0.010263361000397708, 0.010317184998712037, 0.01017443000091589, 0.010260284001560649, 0.010382638000010047, 0.010153502997127362]",tokens/s,97.55573832324829,, +bfloat16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cpu,,42,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cpu,,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +bfloat16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cpu,,42,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,12675.555328,,,,,s,1,159.70505922200027,159.70505922200027,0.0,159.70505922200027,159.70505922200027,159.70505922200027,159.70505922200027,[159.70505922200027],,kWh,0.0018568492267708406,0.0010120535320007513,0,0.002868902758771592,,MB,12898.725888,,,,,s,10,20.571253644002354,2.0571253644002354,0.011200779817389877,2.0574909434999427,2.070192047000637,2.0760838385006535,2.080797271700667,"[2.054039360000388, 2.056634958000359, 2.0583469289995264, 2.059023277999586, 2.0588079760000255, 2.0460298319994763, 2.0688827600006334, 2.043233387001237, 2.0442795340004523, 2.08197563000067]",tokens/s,124.44550265639158,kWh,2.429793235048755e-05,1.324313725576786e-05,0.0,3.754106960625541e-05,tokens/kWh,6819198.352231901,MB,13027.762176,,,,,s,10,81.82274617299663,8.182274617299663,0.03665982580145618,8.180775010499929,8.224108223999064,8.23759191599911,8.248378869599145,"[8.151513246000832, 8.12987049000003, 8.129500327999267, 8.177928581999367, 8.205390242999783, 8.179163675000382, 8.251075607999155, 8.194805806999284, 8.221111847999055, 8.182386345999475]",tokens/s,7.699570467459015,kWh,9.674973843319308e-05,5.273238056033515e-05,0.0,0.00014948211899352824,tokens/kWh,421455.0905766031,,s,630,81.81856641899867,0.12987074034761695,0.0022911200549274205,0.12948964099996374,0.1329902128010872,0.13397532615081217,0.1369667184401624,"[0.13219707800089964, 0.1294600069995795, 0.1295117699992261, 0.1281607740002073, 0.13102802600042196, 0.12746862700078054, 0.12699771899860934, 0.12968346300112898, 0.1279192879992479, 0.12827947400000994, 0.13017008799943142, 0.12739852000049723, 0.12692242000048282, 0.12676236299921584, 0.1264761250004085, 0.12751361899972835, 0.12685818200043286, 0.1286368830005813, 0.13006684699939797, 0.12862791800034756, 0.1288690420005878, 0.1273576299990964, 0.12781822500073758, 0.1288364960000763, 0.13206311899921275, 0.13371546000053058, 0.13373530599892547, 0.13037574000009045, 0.1294951569998375, 0.1294060910004191, 0.1334942579997005, 0.12847179399977904, 0.12946015400120814, 0.12644962499871326, 0.1294178460011608, 0.1277282440005365, 0.1288177859987627, 0.1289441250009986, 0.12912724600028014, 0.1319482099988818, 0.13001189100032207, 0.12915475299996615, 0.12984745200083125, 0.1302862019983877, 0.1285632910003187, 0.13075032599954284, 0.1337802420002845, 0.1296521849999408, 0.13032300000122632, 0.1304019489998609, 0.12868838999929721, 0.12802062099945033, 0.12941339800090645, 0.1309989620003762, 0.13380638399939926, 0.12951988900022116, 0.1308289970002079, 0.12811295800020162, 0.12907940900004178, 0.13017567899987625, 0.12863048599865579, 0.12758119300087856, 0.12777750299937907, 0.1318833410005027, 0.13256101200022385, 0.13217026299935242, 0.1298184860006586, 0.12680122699930507, 0.12537878500006627, 0.1255562619990087, 0.1280915940005798, 0.12777605799965386, 0.12677198900018993, 0.1277283350009384, 0.12783285200021055, 0.1259415029999218, 0.12755482799911988, 0.12717430600059743, 0.13049270199917373, 0.13160540100034268, 0.1277243759996054, 0.12839253800120787, 0.12821829399945273, 0.12955102500018256, 0.1267661780002527, 0.12843687199892884, 0.12883567300013965, 0.1265408210001624, 0.12737923100030457, 0.13070792100006656, 0.13142000600055326, 0.1295258689988259, 0.13141591100065853, 0.12782207000054768, 0.1287719140000263, 0.1284324829994148, 0.13118437400044058, 0.12948252800015325, 0.1318172789997334, 0.13373236599909433, 0.1294431970000005, 0.12870943800044188, 0.12996688400016865, 0.12854099899959692, 0.129140715000176, 0.12874448299953656, 0.1298328960001527, 0.13657588900059636, 0.1302302799995232, 0.13035008100086998, 0.12786595199941075, 0.12793792000047688, 0.1281031719991006, 0.1291262459999416, 0.1309598379993986, 0.13071971099998336, 0.12887806600156182, 0.1279107969985489, 0.13092290700114972, 0.12918468099996971, 0.12762777299940353, 0.12723901900062629, 0.12734871999964525, 0.1278097729991714, 0.12793214400153374, 0.12706776599952718, 0.13353459800055134, 0.1304305169996951, 0.12712601299972448, 0.13051814900063619, 0.1265334259987867, 0.12975888199980545, 0.12794397300058336, 0.1266028650006774, 0.12598323999918648, 0.1255630560008285, 0.1263450999995257, 0.12884653599940066, 0.12789458199949877, 0.13387243600118381, 0.1288862559995323, 0.1298221030010609, 0.12882314899979974, 0.12795955000001413, 0.1299133829998027, 0.1314863379993767, 0.12834948300042015, 0.12870690599993395, 0.1274195879996114, 0.12637658300081966, 0.1266018609985622, 0.12586116100101208, 0.1298339089989895, 0.1268529450007918, 0.12585854200005997, 0.12810707000062393, 0.12620361399967805, 0.12848340299933625, 0.12754658499943616, 0.12655345000166562, 0.12690646299961372, 0.1266674870003044, 0.1279555769997387, 0.12907978399925923, 0.1328189209998527, 0.12976027100012288, 0.12910755700067966, 0.12851432400020713, 0.13157995199981087, 0.13800848299979407, 0.13429904399890802, 0.13217879200055904, 0.1289653100011492, 0.1295250479997776, 0.12867812299919024, 0.13063887799944496, 0.13262702800057014, 0.12909010800103715, 0.13283256699833146, 0.13299010900118446, 0.12789754100049322, 0.12774265299958643, 0.12699354399956064, 0.1271763650001958, 0.12830944400047883, 0.12903120599912654, 0.12734319499941193, 0.1282707130012568, 0.13350563399944804, 0.1337291069994535, 0.13193327699991642, 0.13087840000116557, 0.13737956099976145, 0.13212635600029898, 0.12873653899987403, 0.12739175699971383, 0.12787095100065926, 0.12744510199991055, 0.12692270499974256, 0.13120254899877182, 0.12961712200012698, 0.13034182800038252, 0.12896932100011327, 0.12699264799994125, 0.12842970299971057, 0.12831042200014053, 0.13042926100024488, 0.1271863490001124, 0.1288156339996931, 0.12888300200029335, 0.1312679910006409, 0.12718304699956207, 0.12884319799923105, 0.12764649000018835, 0.12794859200039355, 0.1292405759995745, 0.12753976800013334, 0.12795933800043713, 0.12986645199998748, 0.12923719900027208, 0.12864110400005302, 0.1324309470001026, 0.13065265600016573, 0.12839712999993935, 0.1287539739987551, 0.1273425870003848, 0.12841569800002617, 0.1295905210008641, 0.1264791569992667, 0.1296794490008324, 0.1319670729990321, 0.12979338799959805, 0.13529816700065567, 0.13087219900080527, 0.13092742799926782, 0.12828543099931267, 0.12745958699997573, 0.12972861300113436, 0.12985645499975362, 0.12806804200045008, 0.12998980299926188, 0.13080106599954888, 0.12878125800125417, 0.12871857699974498, 0.1291572999998607, 0.13136648799991235, 0.13156780499957677, 0.13088225499996042, 0.13010897300046054, 0.13538298700041196, 0.1319044699994265, 0.13585792499907257, 0.13407992899919918, 0.13143703800051298, 0.1305697660009173, 0.13056273499933013, 0.12988820799910172, 0.1293816390007123, 0.12632026800019958, 0.127245588999358, 0.1271166490005271, 0.12837111400040158, 0.12895767999907548, 0.12777952699980233, 0.12746798000080162, 0.13011563600048248, 0.1297010529997351, 0.12918860299942025, 0.13006635300007474, 0.1268372240010649, 0.12856079199991655, 0.12855892699917604, 0.12985425000078976, 0.1286015449986735, 0.12825949800026137, 0.13037870500011195, 0.12673654100035492, 0.13354198200067913, 0.14316609900015465, 0.13341286399918317, 0.13090021800053364, 0.12980606599921884, 0.13468727500003297, 0.13076271700083453, 0.131319097999949, 0.1287576300001092, 0.13100350799868465, 0.1297868230012682, 0.1283263289988099, 0.12844692200087593, 0.12967391699930886, 0.13502918400081398, 0.12761099499948614, 0.1277222349999647, 0.1322107150008378, 0.1308455660000618, 0.13231803999951808, 0.13054794799973024, 0.129429094000443, 0.12998207799864758, 0.12914312900102232, 0.12835880999955407, 0.12854538800092996, 0.13117118199988909, 0.13162217500030238, 0.13026602499849105, 0.130070388000604, 0.13029462899976352, 0.13183608399958757, 0.13052382900059456, 0.1312217230006354, 0.13412892499945883, 0.13369067100029497, 0.1292015109993372, 0.12956343699988793, 0.1351484340011666, 0.13154167199900257, 0.1317753100011032, 0.12911422599972866, 0.13006173699977808, 0.12791435799954343, 0.12930223799958185, 0.13040180900134146, 0.13122004799879505, 0.12850136599990947, 0.12795740500041575, 0.13133990899950732, 0.12761127800149552, 0.12884496599872364, 0.12887305700132856, 0.1285005749996344, 0.1288353059990186, 0.13214401799996267, 0.13054859200019564, 0.12751037800080667, 0.1269137250001222, 0.12822335499913606, 0.13419378500111634, 0.127878116998545, 0.1298984400000336, 0.131992408001679, 0.1293154709983355, 0.12800604100084456, 0.1281818209990888, 0.1283294630011369, 0.12887208400024974, 0.1300570549992699, 0.1275827740009845, 0.1282006139990699, 0.1279822750002495, 0.12839042199993855, 0.1283708539995132, 0.13140529700103798, 0.1295858949997637, 0.12786949799919967, 0.13200881000011577, 0.13145474799966905, 0.13014581200150133, 0.12745678599821986, 0.1289277170017158, 0.13041685299867822, 0.13209300100061228, 0.13089790599951812, 0.12993931399978464, 0.12914926600024046, 0.1276671970008465, 0.1279206669987616, 0.13160540100034268, 0.1291061930005526, 0.12808925799981807, 0.13104385699989507, 0.13243148299989116, 0.13305134099937277, 0.13405950900050811, 0.12837315700016916, 0.12854116699963924, 0.13172510500044154, 0.134258514000976, 0.1373506969994196, 0.1330986640004994, 0.13014034399930097, 0.13311759600037476, 0.13115482899956987, 0.12977148500067415, 0.13075303499863367, 0.1291071590003412, 0.13117255100041803, 0.13022271800036833, 0.13195713300046918, 0.12823095199928503, 0.1314124519994948, 0.13368429299953277, 0.1280479170000035, 0.12912641200091457, 0.12894042900006752, 0.12972638000064762, 0.1292644079985621, 0.12948412500009, 0.1328993850002007, 0.1330083470002137, 0.13575020800089987, 0.13307753299886826, 0.1290073290001601, 0.1280947390005167, 0.1302427179998631, 0.13130003799960832, 0.1327821140002925, 0.1332436239990784, 0.13305378700169967, 0.1313832879986876, 0.13006048500028555, 0.12957753800037608, 0.13324633499905758, 0.13511217000086617, 0.13187481000022672, 0.130458390000058, 0.1309919510003965, 0.12938722799844982, 0.1299315910000587, 0.13009234099990863, 0.13262697600112006, 0.13179812700036564, 0.13155553399883502, 0.130781093999758, 0.12925372400059132, 0.12937277300079586, 0.12979287799862504, 0.13503893800043443, 0.12984543700076756, 0.1296075659993221, 0.13226090900025156, 0.13003431300057855, 0.13528087699887692, 0.13196973300000536, 0.12864226100100495, 0.12914826099950005, 0.12885423000079754, 0.12834157099860022, 0.128447037001024, 0.1298996530003933, 0.12778945899844985, 0.1323950709993369, 0.13026674600041588, 0.1283848749990284, 0.12935944400123844, 0.1335131010000623, 0.13272382499962987, 0.12992199799919035, 0.12619028300105128, 0.12700065100034408, 0.12771696699928725, 0.12760643100045854, 0.12822539300032076, 0.12993471099980525, 0.1296313069997268, 0.12969191999945906, 0.13079997400018328, 0.13374575799934973, 0.1285195290001866, 0.1294252219995542, 0.13005583400081377, 0.12979363500016916, 0.12900732799971593, 0.12890838900057133, 0.12801369999942835, 0.13005345500096155, 0.12667135699848586, 0.12712590299997828, 0.12993220000134897, 0.12848306100022455, 0.12717611899824988, 0.12854184500065458, 0.12857790599991858, 0.12935582300087844, 0.13537225700019917, 0.12890739399881568, 0.1303179429996817, 0.13574431400047615, 0.1290349890005018, 0.12981452500025625, 0.1371263529999851, 0.12998363800033985, 0.12825455899837834, 0.13005286100087687, 0.13104675099930319, 0.13158056900101656, 0.13003853900045215, 0.12754894199861155, 0.1301889200003643, 0.13373056499949598, 0.12848615500115557, 0.129052819998833, 0.1315969759998552, 0.13552616500055592, 0.13556914000037068, 0.1320864570006961, 0.13024737899831962, 0.12860798400106432, 0.12956522300009965, 0.13205780899988895, 0.13000399699922127, 0.12897184299981745, 0.12786298200080637, 0.12927052499981073, 0.13592716899984225, 0.13476033699953405, 0.132329636000577, 0.13021639900034643, 0.12823343699892575, 0.12884377599948493, 0.12870155400014482, 0.1290585930000816, 0.12889015700056916, 0.1292000459998235, 0.12800017699919408, 0.1277829410009872, 0.1280966329995863, 0.12840001000040502, 0.13009554500058584, 0.12879741599863337, 0.129905312000119, 0.12647821900100098, 0.12825559900011285, 0.13126895599998534, 0.1284780519999913, 0.12992006099921127, 0.13027903600050195, 0.13139841500014882, 0.1282748180001363, 0.12870134999866423, 0.13086327400014852, 0.13137360300061118, 0.13299114700021164, 0.135204595999312, 0.1326514610009326, 0.13805465800032835, 0.12844219599901407, 0.1287186389999988, 0.13248146400110272, 0.12981248599862738, 0.13129154099988227, 0.12841264600137947, 0.12780669099993247, 0.1268639299996721, 0.1309067509992019, 0.1313243410004361, 0.13235998500022106, 0.13143001699972956, 0.13042659299935622, 0.13192998100021214, 0.13352617600139638, 0.13035512299939, 0.12851713400050357, 0.1293803399985336, 0.13084336100109795, 0.1301561029995355, 0.1330640090000088, 0.12849760300014168, 0.13186916499944346, 0.1333543800010375, 0.13332253599946853, 0.13198518399985915, 0.13001457499922253, 0.1304852370012668, 0.12957621100031247, 0.130549716999667, 0.13153149800018582, 0.14064775999941048, 0.13593682900136628, 0.1293848360000993, 0.13030680699921504, 0.12818223599970224, 0.12863554900104646, 0.12905479799883324, 0.1320876040008443, 0.1312660589992447, 0.1294243930005905, 0.12732347000019217, 0.1276766390001285, 0.1285048720001214, 0.12771583099856798, 0.1312339450014406, 0.12958694400003878, 0.12943000199993548, 0.12891079899964097, 0.13050280700008443, 0.12697141299940995, 0.128162427999996, 0.12817594800071674, 0.12914484799875936, 0.12875916499979212, 0.12822584100103995, 0.12692041099944618, 0.131960918999539, 0.1302603980002459, 0.12903011100024742, 0.13279105100082234, 0.12777081499916676, 0.1309531690003496, 0.13024662799944053, 0.1306845710005291, 0.12924117900001875, 0.12914142800036643, 0.1279118379989086, 0.12931485599983716, 0.12795709100100794, 0.12854972199966141, 0.1288249770004768, 0.13185234699994908, 0.12789663200055656, 0.12960799099892029, 0.1277768769996328, 0.12714469300044584, 0.12748505399940768, 0.13080593000086083, 0.13115474799997173, 0.13343974499912292, 0.1320354170002247, 0.13225041199984844, 0.12825284100108547, 0.13004708399967058, 0.12995121899984952, 0.13056779200087476, 0.1308510529997875, 0.13089072599905194, 0.1326281569999992, 0.129531937000138, 0.1312905579998187, 0.12898651900104596, 0.13072083800034306]",tokens/s,7.699963804959932,, +float32-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cpu,,42,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float32-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cpu,,42,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,24784.900096,,,,,s,1,61.02686414100026,61.02686414100026,0.0,61.02686414100026,61.02686414100026,61.02686414100026,61.02686414100026,[61.02686414100026],,kWh,0.0006901925983298573,0.0003761804603315801,0,0.0010663730586614375,,MB,25631.019008,,,,,s,10,15.961028011002782,1.5961028011002782,0.030233379538160264,1.5840821319998213,1.6283443715006798,1.6524122952508151,1.6716666342509234,"[1.579503900000418, 1.5906680370007962, 1.578171344999646, 1.585258442999475, 1.5736175580004783, 1.6764802190009505, 1.5829058210001676, 1.573124137999912, 1.6229959440006496, 1.5983026060002885]",tokens/s,160.39067146773104,kWh,1.8630973835139078e-05,1.0154414463790743e-05,0.0,2.8785388298929824e-05,tokens/kWh,8893400.96237359,MB,25636.585472,,,,,s,10,106.21130797499973,10.621130797499973,0.0952491986418871,10.597710490999816,10.753395898599956,10.797170641299726,10.832190435459543,"[10.585181579999698, 10.610239401999934, 10.621331709000515, 10.569498233000559, 10.500062295999669, 10.563363115999891, 10.628448117999142, 10.548569959000815, 10.840945383999497, 10.743668178000007]",tokens/s,5.931571807290904,kWh,0.00012500059010639046,6.813020514447285e-05,0.0,0.00019313079525086328,tokens/kWh,326203.8035838223,,s,630,106.20689972099899,0.1685823805095222,0.003902743850471036,0.16797043299902725,0.17398343269942415,0.17547980515082598,0.18072319818982577,"[0.17257069399965985, 0.1738807910005562, 0.17510875599873543, 0.16496041000027617, 0.16476559399961843, 0.16375816199979454, 0.16599752600086504, 0.1654400049992546, 0.16621476500040444, 0.16442809299951477, 0.1651982160001353, 0.1674881369999639, 0.16461957600040478, 0.16792505600096774, 0.16923021599905042, 0.1697901980005554, 0.17103009299898986, 0.17528148400015198, 0.17824452300010307, 0.16805688900058158, 0.16597613299927616, 0.1709476430005452, 0.16926663999947777, 0.16685815100026957, 0.16976633400008723, 0.16609648499979812, 0.1633536350000213, 0.16560250700058532, 0.1641485250002006, 0.1635278459998517, 0.1668488189989148, 0.16653471700010414, 0.1655577430010453, 0.17148317200008023, 0.17530296099903353, 0.16695429700121167, 0.16660765300002822, 0.1683176469996397, 0.17094270000052347, 0.16714854799829482, 0.17186488900006225, 0.1693601140013925, 0.16674276099911367, 0.1685211599997274, 0.16850343599980988, 0.16533377000087057, 0.1683824449992244, 0.16498091900029976, 0.1709483110007568, 0.16586191799979133, 0.16790138299984392, 0.1674075350001658, 0.1674901080004929, 0.1699328599988803, 0.1653688719998172, 0.1679286510006932, 0.16650838200075668, 0.16773832999933802, 0.16520881299948087, 0.16611604100035038, 0.16644536800049536, 0.17543584200029727, 0.1654795069989632, 0.16831072599961772, 0.1646302540011675, 0.16634072099986952, 0.1646925850000116, 0.1638224859998445, 0.1719941169994854, 0.1790654149990587, 0.1652205980008148, 0.16859149400079332, 0.1734789579986682, 0.1701611389999016, 0.16836220300137938, 0.16834073300015007, 0.17069961199922545, 0.16521697099960875, 0.16385403000094811, 0.16852382899924123, 0.16727809400072147, 0.16633071599972027, 0.1692664189995412, 0.16926927600070485, 0.172985544999392, 0.1646104680003191, 0.1635234549994493, 0.17108594899946183, 0.17075933100022667, 0.16787683300026401, 0.16498617100114643, 0.1648567229985929, 0.17085863400097878, 0.16589529299926653, 0.16212574699966353, 0.16562554800111684, 0.16377164000004996, 0.16848016899893992, 0.16613848199995118, 0.16679781400125648, 0.1697943330000271, 0.18089868099923478, 0.17153056200004357, 0.16795999799978745, 0.1696724190005625, 0.17164855199916929, 0.16560105400094471, 0.17018005600039032, 0.16488644399942132, 0.17022911999993084, 0.16964667899992492, 0.1712908539993805, 0.17012017700108117, 0.16636220500004129, 0.17002394299925072, 0.17585267499998736, 0.16666354900007718, 0.1665789210001094, 0.1659154810004111, 0.16548432200033858, 0.1670937579983729, 0.1700506060005864, 0.16712918899975193, 0.16561001100126305, 0.17527828199854412, 0.17050411900163454, 0.17473784600042563, 0.17216934200041578, 0.17133723100050702, 0.17204975899949204, 0.17060342199874867, 0.1726018080007634, 0.1678177859994321, 0.17633762500008743, 0.1694987549999496, 0.16525806299978285, 0.16324243100098101, 0.1706359820000216, 0.1693128560000332, 0.1729503679998743, 0.16809146400009922, 0.1654207920000772, 0.16324430699933146, 0.16366315699997358, 0.1689807140010089, 0.1685912519988051, 0.17201271500016446, 0.16678769500140334, 0.16886616699957813, 0.16608160400028282, 0.16750357999990229, 0.16844003099868132, 0.16225953300090623, 0.16470427499916696, 0.16350227300063125, 0.16885412100054964, 0.16636990299957688, 0.163367424000171, 0.16230056499989587, 0.16258840800037433, 0.16310801799954788, 0.1657776960000774, 0.17186662200037972, 0.16627667099965038, 0.16687891199944715, 0.16433067100115295, 0.16527447499902337, 0.16904500700002245, 0.16600474799997755, 0.16926212799990026, 0.16932907400041586, 0.16692399100065813, 0.17397777199948905, 0.17185404299925722, 0.16925189600078738, 0.17219862700039812, 0.16938580899841327, 0.16898633300115762, 0.1732823299989832, 0.1742542730007699, 0.16675096899962227, 0.16940365400114388, 0.16768205799962743, 0.17285719199935556, 0.17186707300061244, 0.1700675489992136, 0.17525320300046587, 0.17130707499927667, 0.1683110720005061, 0.1700561050001852, 0.16569904900097754, 0.16987082099876716, 0.1691646600011154, 0.16592569899876253, 0.1702379700000165, 0.1674439850012277, 0.1657277679987601, 0.17088175100070657, 0.16700349100028689, 0.1668087079997349, 0.17092671699901985, 0.16552543500074535, 0.16458599699944898, 0.1645771399998921, 0.1654935540009319, 0.1659814840004401, 0.169817587999205, 0.17358629800037306, 0.17322526499992819, 0.16693719000068086, 0.1650931659987691, 0.16763035300027695, 0.16889523999998346, 0.16300399099964125, 0.1675468780013034, 0.17101946199909435, 0.1690738090001105, 0.17533939100030693, 0.17264216199873772, 0.1658127830014564, 0.16585945399856428, 0.16384432300037588, 0.16290733400091995, 0.16199061299994355, 0.1674959560004936, 0.17206005499974708, 0.1607808799999475, 0.16597581699898, 0.17224988599991775, 0.16606262599998445, 0.17188221500146028, 0.16966622599829861, 0.16529545600133133, 0.16503305699916382, 0.1699680180008727, 0.1686992539998755, 0.16613488199982385, 0.1656029160003527, 0.16725131099883583, 0.1651938129998598, 0.16632053900139, 0.16709187800006475, 0.1668719859990233, 0.17519755600005738, 0.1690173910010344, 0.16700703099922976, 0.16491007399963564, 0.1691891320006107, 0.17288253100014117, 0.16799351099871274, 0.16538746500009438, 0.16771957800119708, 0.17307663599967782, 0.16515246400012984, 0.16733045799992396, 0.16494860999955563, 0.16626288299994485, 0.168615275000775, 0.1740343789988401, 0.1690582880000875, 0.16906685899994045, 0.16790780300107144, 0.16469085899916536, 0.16559603799942124, 0.16433240300102625, 0.16212135899877467, 0.16364957400037383, 0.16374409899981401, 0.16408414200122934, 0.16285958399930678, 0.1629002310000942, 0.1646380099991802, 0.16498378300093464, 0.16644827399977657, 0.17082478600059403, 0.17260249699938868, 0.17558151299999736, 0.1643110559998604, 0.1692300200011232, 0.1671030299985432, 0.1694675870003266, 0.16672860700055026, 0.16818420699928538, 0.16633616299986898, 0.17305311800009804, 0.16636416700021073, 0.1697446090001904, 0.16352971600099409, 0.1648876029994426, 0.16641310900013195, 0.16458179499932157, 0.1653423820007447, 0.16700611699889123, 0.17565562600066187, 0.1686646380003367, 0.16606056700038607, 0.16535050299899012, 0.1671247009999206, 0.16914910299965413, 0.16769916800149076, 0.1676913039991632, 0.16368797600080143, 0.1642122979992564, 0.16404556999987108, 0.1623201069996867, 0.1634031049998157, 0.16318055800002185, 0.16626648100100283, 0.1673982039992552, 0.16670704100033618, 0.16411646800042945, 0.1626523279992398, 0.16462850700008858, 0.16560799599938036, 0.16714287800095917, 0.17191235000063898, 0.1644882909986336, 0.16955836300076044, 0.16686824200041883, 0.1700889379990258, 0.1757959070000652, 0.16983465300108946, 0.17139751599825104, 0.16666419700050028, 0.16376628000034543, 0.1677964209993661, 0.16461792800146213, 0.16561449699838704, 0.1716273510010069, 0.1644083840001258, 0.1673450590005814, 0.16798086799826706, 0.16915960400001495, 0.16544647400041868, 0.1692381499997282, 0.16275337800107081, 0.1652135129988892, 0.1703842570004781, 0.16361703000075067, 0.16631902399967657, 0.16631907299961313, 0.16702308499952778, 0.16772933200081752, 0.17025384299995494, 0.16871405900019454, 0.16845128199929604, 0.16770948300109012, 0.17079453599944827, 0.1673314389990992, 0.1687284700001328, 0.16763698100112379, 0.16764743400017323, 0.16375592599979427, 0.1689675969992095, 0.16666581600111385, 0.16854361199875711, 0.16506784400007746, 0.1663582420005696, 0.16750914799922612, 0.16630405300020357, 0.16786831600074947, 0.16850813099881634, 0.16878574800102797, 0.16684451999935845, 0.1659179939997557, 0.16748372100119013, 0.1717662589999236, 0.16548765299921797, 0.168548997000471, 0.16605393200006802, 0.1650440030007303, 0.16721025899823871, 0.1684878830001253, 0.1685819040012575, 0.16466115299954254, 0.16858421299912152, 0.17230356000072788, 0.1674362720004865, 0.18071461499857833, 0.16856331500093802, 0.1684183960005612, 0.16516419399886217, 0.16839452800013532, 0.1743534430006548, 0.17354532100034703, 0.16555330499977572, 0.16410722999899008, 0.16779171800044423, 0.1648151469998993, 0.16587715199966624, 0.16716152600019996, 0.16398972000024514, 0.16739327700088324, 0.18201837400010845, 0.1784945039999002, 0.16973737999978766, 0.16575707499941927, 0.1698789010006294, 0.1704197769995517, 0.1635464270002558, 0.16410208599882026, 0.1657129610011907, 0.16433041100026458, 0.16492567199929908, 0.16539017899958708, 0.16389084900038142, 0.1631596189999982, 0.1650795420009672, 0.16321890599829203, 0.16318714000044565, 0.16860195400113298, 0.1671843779986375, 0.1667969259997335, 0.16959170300106052, 0.16529967099995702, 0.16821502299899294, 0.1699210540009517, 0.1714836209994246, 0.16463292899970838, 0.1671010030004254, 0.16635269000107655, 0.1658895759992447, 0.16761810100069852, 0.1736431729987089, 0.17490778200044588, 0.17591390400048112, 0.17432175199974154, 0.16701417500007665, 0.16719017700052063, 0.16687862199978554, 0.17308740499902342, 0.1807267040003353, 0.17321966300005442, 0.1725214899997809, 0.1693634210005257, 0.1684374750002462, 0.16656171199974779, 0.1744728239991673, 0.1709994810007629, 0.16614716599906387, 0.16518790400004946, 0.17234401600035198, 0.1696646830005193, 0.17499287700047716, 0.17076385599830246, 0.17333917100040708, 0.16619563499989454, 0.1704074799999944, 0.1692908670011093, 0.163688448999892, 0.1667808199999854, 0.16252631899988046, 0.16295217699916975, 0.16316421100054868, 0.16102571500050544, 0.16237634499884734, 0.16775059400060854, 0.1641663200007315, 0.16538502499861352, 0.16743075500016857, 0.163351840001269, 0.16432512599931215, 0.16549536300044565, 0.16532912900038355, 0.1641469549995236, 0.1650385789998836, 0.16265823299909243, 0.16368909000084386, 0.16382756500024698, 0.16625796999869635, 0.16638058700118563, 0.16240733999984514, 0.16248426600031962, 0.16647196399935638, 0.16646776600100566, 0.16288839599837956, 0.16471679800088168, 0.16552352699909534, 0.16901109700120287, 0.16876262400000996, 0.17114968199894065, 0.17798424400098156, 0.1700613399989379, 0.1660964820002846, 0.16745212500063644, 0.17028100899915444, 0.16948500299986335, 0.16840496600161714, 0.17698935699991125, 0.1716476369983866, 0.16825997000159987, 0.16600328399908904, 0.16799805899972853, 0.16417031600030896, 0.16916571799993108, 0.168709014000342, 0.16928895999990345, 0.16956304499944963, 0.16966835700077354, 0.16710278099890274, 0.1676172570005292, 0.17459656699975312, 0.17101142899991828, 0.17191137300142145, 0.17665877099898353, 0.17178663900085667, 0.1772507610003231, 0.16737858199849143, 0.17347889200027566, 0.16733493200081284, 0.17566860199985967, 0.17054322399962984, 0.16996530700089352, 0.17021442099940032, 0.1705912189991068, 0.18686335700112977, 0.1895163049994153, 0.17729385900020134, 0.17424666700026137, 0.16847493699970073, 0.16900656899997557, 0.17103093600053398, 0.1697886980000476, 0.17323808499895677, 0.1696843230001832, 0.16772152200064738, 0.1714475320004567, 0.16842195999925025, 0.16801629399924423, 0.17340953600069042, 0.169133587000033, 0.16636045999985072, 0.17022626000107266, 0.16815919199871132, 0.17610955000054673, 0.17371257900049386, 0.16880227499859757, 0.1760808730014105, 0.17116786499900627, 0.17012573200008774, 0.17214072800015856, 0.1744314179995854, 0.16666684400115628, 0.16890760899877932, 0.1720069090006291, 0.17237324900088424, 0.16943217900006857, 0.1707596279993595, 0.17154187099913543, 0.16905614400093327, 0.1694066380005097, 0.17277456499869004, 0.17745217100127775, 0.1712009200000466, 0.17272097199929703, 0.17121282100015378, 0.18293309199907526, 0.17394386399973882, 0.17047197300053085, 0.16739963200052443, 0.1697491060003813, 0.17123841799912043, 0.16684948300098768, 0.16980355299892835, 0.1712071140009357, 0.18081419999907666, 0.17504082699997525, 0.17017781600043236, 0.16795746599927952, 0.16824630900009652, 0.16701713700058463, 0.16820087899941427, 0.16935781500069425, 0.16656688500006567, 0.1662721100001363, 0.1681277739990037, 0.16812473599929945, 0.17004513900064921, 0.17476199500015355, 0.17448567999963416, 0.1684055850000732, 0.16844529199988756, 0.16545638699972187, 0.17242777200044657, 0.17513796800085402, 0.17891412799872342, 0.17502327200054424, 0.16787594699962938, 0.16783639500135905, 0.1723073769990151, 0.1673826080004801, 0.16689367799881438, 0.16757790300107445, 0.17027259899987257, 0.16828513899963582, 0.17131683600018732, 0.17467702099929738, 0.1745973180004512, 0.17568369600121514, 0.17099710999900708, 0.1763992070009408, 0.1733838009986357, 0.17404666699985682, 0.17698723900139157, 0.17423198699907516, 0.16879198100104986, 0.16793288799999573, 0.17403914499845996, 0.16891961800138233, 0.16843913499906193, 0.1734747160007828, 0.16602589299873216, 0.1715097460000834, 0.17859293000037724, 0.17426908500056015, 0.16762392400050885, 0.172240026999134, 0.1737197290003678, 0.17136012399896572, 0.1755157750012586, 0.1711768610002764, 0.17495472999871708, 0.16767336700104352, 0.17043456200008222, 0.16424893200019142, 0.1623269749998144, 0.1630622510001558, 0.16641706399968825, 0.16675635100000363, 0.16987189399878844]",tokens/s,5.9318180048093225,, +float32-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cpu,,42,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float32-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cpu,,42,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,6213.902336,,,,,s,1,16.82828445300038,16.82828445300038,0.0,16.82828445300038,16.82828445300038,16.82828445300038,16.82828445300038,[16.82828445300038],,kWh,0.000163032265474313,8.885757816504003e-05,0,0.000251889843639353,,MB,6334.312448,,,,,s,10,3.714737233998676,0.3714737233998676,0.007708497202829009,0.3687407044990323,0.37846695830048704,0.3842126676501721,0.3888092351299201,"[0.3665656239991222, 0.37282725800105254, 0.36448245300016424, 0.3631408780001948, 0.36830376199941384, 0.377190134000557, 0.3770356740005809, 0.36605542699908256, 0.36917764699865074, 0.38995837699985714]",tokens/s,689.146994454928,kWh,4.359083559325965e-06,2.3758070505436826e-06,0.0,6.734890609869648e-06,tokens/kWh,38011010.843271114,MB,6339.305472,,,,,s,10,28.279212945000836,2.827921294500084,0.02247434674107151,2.826415186000304,2.8522092179009633,2.865909620450111,2.8768699424894293,"[2.8244233060013357, 2.837737545998607, 2.801933495000412, 2.828407065999272, 2.879610022999259, 2.8068749949998164, 2.8299959569994826, 2.8037580190011795, 2.8491646840011526, 2.817307854000319]",tokens/s,22.2778477330774,kWh,3.368288653400898e-05,1.8358372070625973e-05,0.0,5.2041258604634965e-05,tokens/kWh,1210577.9469828005,,s,630,28.274672424999153,0.044880432420633574,0.0011727571334431803,0.044669613499536354,0.04645259969965991,0.04720746805087401,0.048416777320253464,"[0.047616452000511345, 0.04500083600032667, 0.04357341499962786, 0.043362002999856486, 0.04322311299983994, 0.042846592999921995, 0.04392330500013486, 0.04348152100101288, 0.044620249998843065, 0.04498082500140299, 0.0453546759999881, 0.04601072699915676, 0.04340224100087653, 0.043908796998948674, 0.04398424400096701, 0.043568470999161946, 0.043551380000280915, 0.04423656499966455, 0.04522772000018449, 0.044490427999335225, 0.04415977100143209, 0.04481037799996557, 0.04337102899989986, 0.04456848799964064, 0.04384122800001933, 0.04326177499933692, 0.04438445500090893, 0.045511026999520254, 0.044399152999176295, 0.04383762700126681, 0.04411357800017868, 0.04470384500018554, 0.04608506499971554, 0.04820492699946044, 0.045500889000322786, 0.04591168099977949, 0.04525337499944726, 0.04500620300132141, 0.046451473999695736, 0.0445720949992392, 0.04504755200105137, 0.044833989999460755, 0.04413523599941982, 0.04331343200101401, 0.043706337999537936, 0.045341250999626936, 0.04654720799953793, 0.046105741001156275, 0.04553804399984074, 0.04525115300020843, 0.0473070199986978, 0.0460285670014855, 0.04545643299934454, 0.043768130999524146, 0.044516317000670824, 0.04584303099909448, 0.044293450000623125, 0.04377764300079434, 0.04471628599821997, 0.04572423500030709, 0.04633483700126817, 0.04673706499852415, 0.04542559400033497, 0.046644086000014795, 0.044958062999285175, 0.0447727640002995, 0.04516596400026174, 0.044043490999683854, 0.04324137199910183, 0.04393077200074913, 0.04335965099926398, 0.043639449000693276, 0.043105388000185485, 0.04306403199916531, 0.042949644001055276, 0.04300856100053352, 0.04333893999864813, 0.043781715001387056, 0.04416690299876791, 0.04543620100048429, 0.04453693100003875, 0.04393080999943777, 0.04379871099990851, 0.045369749999736086, 0.043690006001270376, 0.043397629999162746, 0.04423723100080679, 0.044988999999986845, 0.04453871700025047, 0.04523036199861963, 0.04505776500081993, 0.0447732669999823, 0.04425030099992, 0.0456214789992373, 0.04683704900162411, 0.0466670059995522, 0.0461573069987935, 0.04535368700089748, 0.04435379399910744, 0.04491087099995639, 0.045863898001698544, 0.045639761998245376, 0.047324234001280274, 0.045036053999865544, 0.044684595999569865, 0.044349017000058666, 0.044135853999250685, 0.044772612000087975, 0.04483277200051816, 0.04506300499997451, 0.04917057500097144, 0.044209136998688336, 0.045123855999918305, 0.04480135599987989, 0.04535061800015683, 0.045780256999933044, 0.046059622000029776, 0.04642762100047548, 0.04646273099933751, 0.04516118100036692, 0.04627738700037298, 0.04605393500060018, 0.04584991599949717, 0.04725398899972788, 0.047288244000810664, 0.04804658699868014, 0.049363452000761754, 0.044561243999851285, 0.04582825800025603, 0.04382458499821951, 0.0443800930006546, 0.043861596001079306, 0.043374021999625256, 0.04394494199914334, 0.04375234799954342, 0.04368905800038192, 0.043211239000811474, 0.04328156699921237, 0.04279735999989498, 0.04284189000100014, 0.04320875499979593, 0.044043034000424086, 0.04359093599850894, 0.044128993000413175, 0.04504454600100871, 0.04590145600013784, 0.04417490899868426, 0.04446301400093944, 0.04406306599958043, 0.04480440000043018, 0.04554258099960862, 0.044505821000711876, 0.04390623399922333, 0.04361586300001363, 0.04425946499941347, 0.04377483200005372, 0.04474071399999957, 0.04378078500121774, 0.046860935999575304, 0.04596698100067442, 0.04318188699835446, 0.044209389001480304, 0.045781712999087176, 0.04413629100054095, 0.044456390000050305, 0.04464512699996703, 0.04439176599953498, 0.0441274879995035, 0.04409353099981672, 0.043738043001212645, 0.043838949999553734, 0.043807694000861375, 0.044661903999440256, 0.04510758200012788, 0.04529131500021322, 0.04349542799900519, 0.044439369999963674, 0.04409418500108586, 0.045046250999803306, 0.04499474099975487, 0.04762166799991974, 0.04539595700043719, 0.04427091700017627, 0.045748794998871745, 0.043954894001217326, 0.04388725699936913, 0.04372094099926471, 0.04510145300082513, 0.04523786499885318, 0.04643167499853007, 0.04577179400075693, 0.04636061100063671, 0.04580371199881483, 0.045194824000645895, 0.04536520799956634, 0.043440948000352364, 0.04507415999978548, 0.04417739600103232, 0.04703462499855959, 0.043921451000642264, 0.04445573399971181, 0.04325968100056343, 0.04470007999952941, 0.047195142000418855, 0.04695714799890993, 0.04513313400093466, 0.043782266999187414, 0.044133297000371385, 0.04564505199959967, 0.04517240099994524, 0.045253673000843264, 0.04452277999916987, 0.04410246100087534, 0.04400945199995476, 0.04550692699922365, 0.04539466200003517, 0.04522977700071351, 0.045736608999504824, 0.047044975000972045, 0.04553346799912106, 0.047217553001246415, 0.04499868199854973, 0.0443508890002704, 0.04541611000058765, 0.043911437000133446, 0.04461076199913805, 0.0453706250009418, 0.04533038899899111, 0.04531006599972898, 0.044204761001310544, 0.04436992799855943, 0.04392899200138345, 0.043864700999620254, 0.04448939000030805, 0.04434292999940226, 0.0445096769999509, 0.044925919000888825, 0.04412773100011691, 0.04293470799893839, 0.04351413699987461, 0.04305361099977745, 0.045043508000162547, 0.04439962899959937, 0.04485096500138752, 0.04382878099931986, 0.04482187699977658, 0.04430674899958831, 0.04377599500003271, 0.04413172600106918, 0.044373880000421195, 0.04681522999999288, 0.045326546000069357, 0.04826671199953125, 0.04689284400046745, 0.04407089599953906, 0.046759374999965075, 0.04656292300023779, 0.04622854499939422, 0.04558829000052356, 0.04413304499939841, 0.04523638300088351, 0.04738904499936325, 0.045700429000135045, 0.04616213200097263, 0.04518755999924906, 0.046226712000134285, 0.046427421999396756, 0.04656732200055558, 0.046678916000018944, 0.04688075299964112, 0.04542857300111791, 0.046035879999180906, 0.04732274900015909, 0.04684094299955177, 0.04569142199943599, 0.04686767000021064, 0.048435016000439646, 0.04749292000087735, 0.04714211999998952, 0.045780328999171616, 0.046226415999626624, 0.04550256699985766, 0.0459384770001634, 0.04555208800047694, 0.044529254999361, 0.04514690799987875, 0.04433821300153795, 0.04458023899860564, 0.04568173200095771, 0.045479757000066456, 0.045595049999974435, 0.043600003999017645, 0.045481436000045505, 0.04501094700026442, 0.04677142000036838, 0.04450549300054263, 0.043972870998914004, 0.046170073001121636, 0.04510295999898517, 0.0455861360005656, 0.044520085999465664, 0.045718892999502714, 0.04363406600168673, 0.04357824499857088, 0.044179726000947994, 0.04424170200036315, 0.043913711999266525, 0.046036470999752055, 0.047290892000091844, 0.04494983900076477, 0.04512871099905169, 0.04589567699986219, 0.045994632000656566, 0.04556326599958993, 0.04570114300076966, 0.04774125799849571, 0.04578750800101261, 0.04334687899972778, 0.04384132899940596, 0.043192336001084186, 0.04290550999940024, 0.042963136000253144, 0.043875827999727335, 0.04520089899961022, 0.048774865999803296, 0.04543073800050479, 0.044857430999400094, 0.04402583000046434, 0.04457668100076262, 0.04403756399915437, 0.04403238499980944, 0.04455529500046396, 0.04496170499987784, 0.045008520999545, 0.04365143200084276, 0.04308293699978094, 0.04411744999924849, 0.043522147001567646, 0.04373956599920348, 0.04390758199951961, 0.044766166000044905, 0.04497773100047198, 0.04632655800014618, 0.0450204249991657, 0.045101082001565374, 0.043428384999060654, 0.043758892999903765, 0.04360940500009747, 0.043998880000799545, 0.04501832200003264, 0.044473632999142865, 0.04373732299973199, 0.04329045400118048, 0.044107664998591645, 0.043826301000081, 0.043061373000455205, 0.04683971099984774, 0.043908244000704144, 0.043306253999617184, 0.047287948000303004, 0.044141991998913, 0.0453430940015096, 0.044891526998981135, 0.04477433099964401, 0.04344732000026852, 0.04442057800042676, 0.04483424499994726, 0.044416860000637826, 0.043777939999927185, 0.04492282100000011, 0.04476164900006552, 0.04460688699873572, 0.04469355100081884, 0.04458655800044653, 0.04484125299859443, 0.04565162600010808, 0.045896157000242965, 0.045501365000745864, 0.04797606200008886, 0.04505432099904283, 0.04526974100008374, 0.04634475199964072, 0.04537933900064672, 0.04467222699895501, 0.04628723499990883, 0.044941569000002346, 0.04442410600131552, 0.043997152999509126, 0.043868466000276385, 0.044569698999112006, 0.044426589000067906, 0.04446099500091805, 0.044950194998818915, 0.04413266900155577, 0.04520468800001254, 0.04618582599869114, 0.04837212399979762, 0.04677715100115165, 0.04703412699927867, 0.04744066700004623, 0.04578249199948914, 0.04478908900091483, 0.047149957999863545, 0.04421162099970388, 0.04316146200108051, 0.0434486839985766, 0.044631014001424774, 0.04424092899898824, 0.043345198000679375, 0.04383762299949012, 0.04383811200023047, 0.04507937000016682, 0.04409302899875911, 0.04365639300158364, 0.044119488999058376, 0.04406042400114529, 0.04433237699959136, 0.04519354299918632, 0.044231899999431334, 0.04550415900121152, 0.044998703999226564, 0.04474387300069793, 0.0438151590005873, 0.04552281499854871, 0.04550584699973115, 0.044762879000700195, 0.044287885999438004, 0.043823630001497804, 0.04380009099986637, 0.044707626999297645, 0.045027366000795155, 0.04534876299840107, 0.044449839000662905, 0.04436632500073756, 0.04366319799919438, 0.045132793000448146, 0.04577172800054541, 0.04476003899981151, 0.046123727999656694, 0.043338780000340194, 0.044951902998946025, 0.047161848000541795, 0.04679249099899607, 0.046209579000787926, 0.04459291199964355, 0.043911722999837366, 0.04513304200008861, 0.04457628700038185, 0.044602142999792704, 0.04390366200095741, 0.0454823159998341, 0.04485618400030944, 0.04562004999934288, 0.04371250899930601, 0.04273009100143099, 0.04433099399830098, 0.043541279001146904, 0.043966085999272764, 0.04377923399988504, 0.043701500000679516, 0.04280589200061513, 0.04543544799889787, 0.045661591000680346, 0.04389560799972969, 0.04376732099990477, 0.043770419999418664, 0.04331467900010466, 0.043633942001179093, 0.044303258999207173, 0.04512871100087068, 0.045372635999228805, 0.04558861099940259, 0.04465039900060219, 0.04346453399921302, 0.04380403800132626, 0.04411039799924765, 0.04450779299986607, 0.045492147999539156, 0.04411537900159601, 0.04507789099989168, 0.044295916999544716, 0.04347221100033494, 0.04361064200020337, 0.04507467699841072, 0.04547391699998116, 0.0441201200010255, 0.04432321099920955, 0.04486545500003558, 0.04517505500007246, 0.044366465001075994, 0.043926574999204604, 0.04478537000068172, 0.04344700500041654, 0.044083593998948345, 0.04328669999995327, 0.04565236599955824, 0.045325773000513436, 0.04483955599971523, 0.04422610599976906, 0.04370740400008799, 0.04416654600026959, 0.0446135719994345, 0.044348458000968094, 0.04568325900072523, 0.047674342000391334, 0.0443852770004014, 0.04437119099929987, 0.04564468300122826, 0.044289919000220834, 0.04409401699922455, 0.043973200999971596, 0.04447685000013735, 0.044869201999972574, 0.04407060300036392, 0.04332717499892169, 0.043816286000947, 0.04416693100029079, 0.04485899999963294, 0.043876902000192786, 0.04337613200004853, 0.04459147600027791, 0.044617453999308054, 0.04385736400035967, 0.044858567000119365, 0.048981419999108766, 0.04433096400134673, 0.04418831999828399, 0.04390436700123246, 0.04403583699968294, 0.04435709599965776, 0.04349026300042169, 0.04516388099909818, 0.045746099000098184, 0.04623492200153123, 0.049220050999792875, 0.04741909599943028, 0.044899823999003274, 0.043914109000979806, 0.04528372099957778, 0.046014924000701285, 0.04562237199934316, 0.04474581000067701, 0.044538905000081286, 0.0444300259987358, 0.04485911800111353, 0.044458565998866106, 0.04417612300130713, 0.04538698699980159, 0.04545462500027497, 0.04363614099929691, 0.045006009999269736, 0.0463808570002584, 0.046183790000213776, 0.044836389999545645, 0.04794932300137589, 0.04806096199899912, 0.04795960199953697, 0.046793327001068974, 0.04621167400000559, 0.04483726699982071, 0.04537595299916575, 0.044782608001696644, 0.04484624099859502, 0.045149608000429, 0.04643761899933452, 0.047873535000690026, 0.046298392000608146, 0.04850015199917834, 0.04633296300016809, 0.044369657000061125, 0.043621857001198805, 0.0451516229986737, 0.046930106000218075, 0.04447718200026429, 0.044015516999934334, 0.044782684999518096, 0.045394360000500455, 0.04534589000104461, 0.0447100439996575, 0.04476653599886049, 0.04510821200165083, 0.04500446699967142, 0.0451597289993515, 0.044436393000069074, 0.04420728299919574, 0.04451526200136868, 0.044550904000061564, 0.044015645999024855, 0.044667000000117696, 0.044210457999724895, 0.04404329200042412, 0.04444051900100021, 0.044663697999567376, 0.04397423999944294, 0.043961821000266355, 0.04498613500072679, 0.04442280699913681, 0.043594955999651575, 0.044235240000489284, 0.04433271199923183, 0.04509943400080374, 0.046129827000186197, 0.04415917999904195, 0.044334727001114516, 0.04573800899925118, 0.044559083000422106, 0.04390699999930803, 0.04494740600057412, 0.04442783100057568, 0.044000291998600005, 0.044159071001558914, 0.044204292998983874, 0.04436176899980637, 0.04652045600050769, 0.044367705999320606, 0.04492141199989419, 0.04446433200064348, 0.044413777999579906, 0.04415121500096575, 0.045376805999694625, 0.04418687500037777, 0.044348080999043304, 0.04562299200006237, 0.04442778799966618, 0.043795040000986774, 0.044147171000076924, 0.04556468199916708, 0.043944248000116204, 0.0441491199999291, 0.04552024700024049]",tokens/s,22.281425246256195,, +float32-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gptj,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cpu,,42,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 564, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTJForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +bfloat16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cpu,,42,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,800.27648,,,,,s,1,7.250024639997719,7.250024639997719,0.0,7.250024639997719,7.250024639997719,7.250024639997719,7.250024639997719,[7.250024639997719],,kWh,5.721620516041337e-05,3.1183561531503536e-05,0,8.839976669191691e-05,,MB,834.506752,,,,,s,17,1.1208413999956974,0.06593184705857044,0.0012142121645627763,0.06581297099910444,0.06758569099984016,0.06782723380092648,0.06798058676140499,"[0.06478042399976403, 0.06745661099921563, 0.06655471200065222, 0.06474048499876517, 0.06389624300209107, 0.0663358710007742, 0.06438100000013947, 0.06581297099910444, 0.06420548999813036, 0.06569095499799005, 0.06564425000033225, 0.06708905999767012, 0.06647115499799838, 0.06641139500061399, 0.0655725420001545, 0.06801892500152462, 0.06777931100077694]",tokens/s,3882.7973342318596,kWh,7.885553023935346e-07,4.2978040299061075e-07,0.0,1.2183357053841454e-06,tokens/kWh,210122709.91375265,MB,846.782464,,,,,s,17,9.356281086991658,0.5503694757053916,0.00833010417346717,0.5495066299990867,0.5642630127986195,0.5667418311983056,0.5669455758393451,"[0.5520696709972981, 0.550958735999302, 0.5433079629983695, 0.5409999229996174, 0.5424858739970659, 0.5461325489995943, 0.5412751139992906, 0.5408520019991556, 0.5456455460007419, 0.5465740070030733, 0.5495066299990867, 0.5525149640016025, 0.5501940040012414, 0.5574365169995872, 0.5626529139990453, 0.566996511999605, 0.5666781609979807]",tokens/s,114.46855754355715,kWh,6.433827144829188e-06,3.5065853545587428e-06,0.0,9.94041249938793e-06,tokens/kWh,6337765.158526284,,s,1071,9.348910220996913,0.00872914119607555,0.0002896541047236104,0.008657130001665791,0.009060186999704456,0.0092222054990998,0.009913969300396274,"[0.009949040999345016, 0.009209806001308607, 0.009188401996652829, 0.009028280001075473, 0.008821491999697173, 0.008920778000174323, 0.008724571001948789, 0.008726179999939632, 0.00990794799872674, 0.008826427001622505, 0.008897640000213869, 0.008859830999426777, 0.008781008997175377, 0.008879062002961291, 0.008907682997232769, 0.008888497002772056, 0.00883001199690625, 0.008813815002213232, 0.008840276001137681, 0.008710250000149244, 0.008672761996422196, 0.00864500100215082, 0.008745056999032386, 0.008635148999019293, 0.008573785002226941, 0.008555660999263637, 0.00863283400030923, 0.008702154998900369, 0.008659248000185471, 0.008640445001219632, 0.008642221000627615, 0.008681670999067137, 0.008875932999217184, 0.008699395999428816, 0.008691394999914337, 0.008594793001975631, 0.008556255997973494, 0.008451508001598995, 0.008483307999995304, 0.008513974000379676, 0.008568075998482527, 0.008690923001267947, 0.008615789000032237, 0.008752554000238888, 0.008712421000382164, 0.008603534999565454, 0.008615560000180267, 0.00856458499765722, 0.008460005003144033, 0.008494958998198854, 0.008447252999758348, 0.008431721998931607, 0.008591413003159687, 0.008797222999419319, 0.008674577999045141, 0.008618833999207709, 0.008840115999191767, 0.008946607002144447, 0.008704801999556366, 0.008953934000601294, 0.008681483999680495, 0.008632969998870976, 0.008888583000953076, 0.010385957000835333, 0.009621646000596229, 0.009467271996982163, 0.008745659000851447, 0.008651641001051757, 0.00869301399870892, 0.008501206000801176, 0.008520300001691794, 0.008522128999175038, 0.008754790000239154, 0.009420037997188047, 0.009252963001927128, 0.008605119001003914, 0.008579910998378182, 0.008637173999886727, 0.008915320999221876, 0.010024144001363311, 0.009119737998844357, 0.00855114400110324, 0.008693916999618523, 0.008491052998579107, 0.008763354002439883, 0.008728075998078566, 0.008793513999989955, 0.008585351999499835, 0.00891639200199279, 0.008499186998960795, 0.008598831002018414, 0.008624981997854775, 0.008573566999984905, 0.008389156002522213, 0.008348874998773681, 0.008625029000540962, 0.0085005729997647, 0.008418147001066245, 0.008639922998554539, 0.008393722000619164, 0.008546630000637379, 0.008602502999565331, 0.008823591997497715, 0.008460397002636455, 0.008481388998916373, 0.0084534759989765, 0.00866547899931902, 0.008414048003032804, 0.008443352999165654, 0.010514985999179771, 0.00839002700013225, 0.008483251000143355, 0.008591401001467602, 0.00865031099965563, 0.008557969998946646, 0.008669661001476925, 0.008612671997980215, 0.008804405999399023, 0.008591105000959942, 0.008581190999393584, 0.00845975500124041, 0.008661259998916648, 0.00859723499888787, 0.00857390300006955, 0.008614099002443254, 0.008716237000044202, 0.00974530399980722, 0.008922763001464773, 0.008757949999562697, 0.008684473999892361, 0.009396981000463711, 0.009007728996948572, 0.008627302002423676, 0.008554441999876872, 0.008567307999328477, 0.008442074999038596, 0.008592571000917815, 0.008351945998583687, 0.00856693199966685, 0.008342859000549652, 0.008355346002645092, 0.00854643499769736, 0.008523787000740413, 0.008487538998451782, 0.008469268002954777, 0.008808913997199852, 0.00874524100072449, 0.008591482001065742, 0.008611175999249099, 0.008571467998990556, 0.008554465999623062, 0.00840569200227037, 0.00843326500034891, 0.008579130997532047, 0.008452969002973987, 0.008477772000333061, 0.008387587997276569, 0.008391933002712904, 0.008519521998096025, 0.008437807999143843, 0.008619162999821128, 0.008526331999746617, 0.009318694999819854, 0.008621120003226679, 0.008640119998744922, 0.009027207997860387, 0.008719066001503961, 0.008608575000835117, 0.008552520997909596, 0.008578621000197018, 0.008526224999513943, 0.008571736001613317, 0.008678769998368807, 0.008524763001332758, 0.00842347900106688, 0.008408269000938162, 0.00865501999942353, 0.008480136999423848, 0.008602382000390207, 0.00843758699920727, 0.008787681999820052, 0.00856680799915921, 0.008684389002155513, 0.008786594997218344, 0.00860625800123671, 0.008502343000145629, 0.008385578999877907, 0.008662770000228193, 0.008481559998472221, 0.009910947999742348, 0.009208125000441214, 0.00961105399983353, 0.008748068998102099, 0.008529190999979619, 0.008544712000002619, 0.008543381001800299, 0.008434047998889582, 0.008645421999972314, 0.008638629002234666, 0.008516874997440027, 0.008366229001694592, 0.00842832099806401, 0.008441923000646057, 0.008629452000604942, 0.008610459000919946, 0.008512125998095144, 0.008527129000867717, 0.00851395700010471, 0.008503126999130473, 0.008392216001084307, 0.008593370999733452, 0.008560682999814162, 0.008625910999398911, 0.008564805000787601, 0.00847208900086116, 0.008529994000127772, 0.008547737001208588, 0.008440879999398021, 0.008316123999975389, 0.008568437999201706, 0.008405126998695778, 0.008480183001665864, 0.008507221999025205, 0.008505547000822844, 0.008531821000360651, 0.008534982000128366, 0.008538799000234576, 0.008647253998788074, 0.008579453999118414, 0.008452697002212517, 0.008608577998529654, 0.008693797000887571, 0.008572325998102315, 0.008459720000246307, 0.008481701999698998, 0.008410623002419015, 0.00845506899713655, 0.008454065002297284, 0.008527366000635084, 0.008665958997880807, 0.0087277830025414, 0.008612652000010712, 0.008535476998076774, 0.008497121001710184, 0.008531005998520413, 0.008505001998855732, 0.008501195999997435, 0.00853365300281439, 0.008441283000138355, 0.0085206029980327, 0.008540547998563852, 0.008646134003356565, 0.009569906997057842, 0.008773246001510415, 0.0086028029982117, 0.008554534000722924, 0.008315128001413541, 0.008582654998463113, 0.008517562000633916, 0.008485957001539646, 0.00856363499769941, 0.008552888000849634, 0.008440137000434333, 0.008624998001323547, 0.008344601999851875, 0.008659600000100909, 0.008675966997543583, 0.008531024999683723, 0.008373879001737805, 0.008606711999163963, 0.008469397998851491, 0.00851510200300254, 0.00879550099853077, 0.008586673000536393, 0.008571602000301937, 0.008712626000487944, 0.008670522998727392, 0.008469095999316778, 0.008410753998759901, 0.00872793100279523, 0.008513084998412523, 0.008543202002329053, 0.009277609999116976, 0.008530772000085562, 0.008713908999197884, 0.008520693001628388, 0.008637224997073645, 0.008584444000007352, 0.008475947000988526, 0.00866651400065166, 0.008499572999426164, 0.008503919001668692, 0.008533745996828657, 0.008548953002900817, 0.008655965997604653, 0.008627236002212157, 0.008556388998840703, 0.008646615999168716, 0.00870143100109999, 0.008718720000615576, 0.008652279997477308, 0.008617635001428425, 0.008611339999333723, 0.008503140001266729, 0.008450523000647081, 0.008441193000180647, 0.008827435998682631, 0.008384528999158647, 0.00839094299954013, 0.008651736003230326, 0.00856470899816486, 0.008666649999213405, 0.008809790000668727, 0.008755082999414299, 0.008569421999709448, 0.009723630999360466, 0.008812050000415184, 0.008807158999843523, 0.00847452200105181, 0.008533038999303244, 0.008526069999788888, 0.008485182999720564, 0.008369944000151008, 0.008473941001284402, 0.008408942001551623, 0.008387484998820582, 0.008394845001021167, 0.00843361699662637, 0.008679032001964515, 0.008564241001295159, 0.008529077997081913, 0.008494363002682803, 0.008523714997863863, 0.00873294499979238, 0.008786311002040748, 0.008759618998738006, 0.008534939999663038, 0.008601178000390064, 0.008366735000890912, 0.008635754998977063, 0.00891514700197149, 0.00881126499734819, 0.008467728999676183, 0.008567029002733761, 0.009004722996905912, 0.008532073003152618, 0.00853090899909148, 0.008486098999128444, 0.00876126699949964, 0.008560058999137254, 0.008626578000985319, 0.008829897000396159, 0.008703006998985074, 0.00855577300171717, 0.008742912999878172, 0.008968148998974357, 0.008686871002282714, 0.008450377998087788, 0.008680805000039982, 0.008534024000255158, 0.008477575000142679, 0.00873869000133709, 0.008535700999345863, 0.008614734000730095, 0.008624081998277688, 0.008772987999691395, 0.008484944999509025, 0.008503872999426676, 0.008694118001585593, 0.00900198399904184, 0.008695253000041703, 0.008688199002790498, 0.0086475669995707, 0.008677562000229955, 0.008630368000012822, 0.009259370999643579, 0.008904867998353438, 0.009301312002207851, 0.009564107000187505, 0.008979563997854711, 0.008616544000688009, 0.008565492000343511, 0.008782632001384627, 0.008585282997955801, 0.00849898399974336, 0.008484920999762835, 0.008445515999483177, 0.00854157800131361, 0.008530104001692962, 0.008652902997710044, 0.008484279002004769, 0.008409423000557581, 0.008587773998442572, 0.008497717000864213, 0.008412841998506337, 0.008565739000914618, 0.008533595999324461, 0.008599593998951605, 0.008916804003092693, 0.008630635998997604, 0.008511100000760052, 0.008530183997208951, 0.008788366001681425, 0.008600282999395858, 0.008388559999730205, 0.008679836999363033, 0.00842168000235688, 0.008480727999994997, 0.00842521999948076, 0.008592869999120012, 0.008383355001569726, 0.008494006997352699, 0.009637105002184398, 0.008482774999720277, 0.00858483900083229, 0.008561332997487625, 0.008485015001497231, 0.008484967998811044, 0.00851935299942852, 0.008616867002274375, 0.008472579000226688, 0.008437731998128584, 0.008427519001998007, 0.008559205998608377, 0.008508658000209834, 0.00846922100026859, 0.008512443000654457, 0.008495158999721752, 0.00845497399859596, 0.008450256998912664, 0.008635051002784166, 0.008760206997976638, 0.008757454001170117, 0.008637262999400264, 0.008533207001164556, 0.008545682998374104, 0.00855879300070228, 0.008517572001437657, 0.008582760998251615, 0.008548515001166379, 0.008401856997807045, 0.009761714998603566, 0.009066603000974283, 0.008616712999355514, 0.008534967000741744, 0.009921019001922105, 0.00845400499747484, 0.008587790001911344, 0.008344562000274891, 0.00843420899764169, 0.008404270000028191, 0.008386981000512606, 0.008660294999572216, 0.008604751001257682, 0.008415055999648757, 0.008484182999382028, 0.008659964001708431, 0.009108496000408195, 0.008587522999732755, 0.008590973997343099, 0.008472813002299517, 0.008523000000423053, 0.008495875998050906, 0.008493926001392538, 0.008422066999628441, 0.008456879000732442, 0.008490723997965688, 0.00847844400050235, 0.008481138000206556, 0.008393047999561531, 0.008488228999340208, 0.008435168001597049, 0.008414911000727443, 0.008441254998615477, 0.008384114000364207, 0.008536115998140303, 0.00858098900062032, 0.008581019999837736, 0.008334514001035132, 0.008441522000794066, 0.008554543001082493, 0.008445755996945081, 0.008440915000392124, 0.008412847000727197, 0.008452710000710795, 0.008534177999536041, 0.008498087998304982, 0.008512600001267856, 0.008548131001589354, 0.00872003900076379, 0.008541736999177374, 0.008928965999075444, 0.00855267600127263, 0.008672624997416278, 0.008701929000380915, 0.008596804000262637, 0.00855847300044843, 0.008483909001370193, 0.00864456099952804, 0.00844726799914497, 0.0084695740006282, 0.008494088000588818, 0.008480248998239404, 0.00878365899916389, 0.009970267001335742, 0.008878757998900255, 0.008739409000554588, 0.008699779999005841, 0.008614963000582065, 0.008649416999105597, 0.008882795002136845, 0.00856437000038568, 0.008635139998659724, 0.008847960998537019, 0.008863114002451766, 0.008741923000343377, 0.008638392999273492, 0.008531543000572128, 0.008493944998917868, 0.008376658999623032, 0.008564333002141211, 0.008526077999704285, 0.008616431998234475, 0.008460678000119515, 0.008739362001506379, 0.008788719998847228, 0.008638250001240522, 0.008513634998962516, 0.008536623998224968, 0.008667437003168743, 0.008472430999972858, 0.008621860997664044, 0.008492626999213826, 0.00843423300102586, 0.008607112999015953, 0.008279179000965087, 0.00860043700231472, 0.008387578996917, 0.008692929000972072, 0.008473412999592256, 0.008417389999522129, 0.008603075002611149, 0.008603764999861596, 0.008611182998720324, 0.008835653999994975, 0.00854243199864868, 0.008487581002555089, 0.008481279997795355, 0.008598511001764564, 0.008714359999430599, 0.008576904998335522, 0.00867235400073696, 0.008398710000619758, 0.008434753999608802, 0.008583204002206912, 0.008729387998755556, 0.008544263000658248, 0.008644912999443477, 0.008699096000782447, 0.009034140999574447, 0.00878594899768359, 0.008831537001242395, 0.008815500001219334, 0.008884363000106532, 0.00882665300014196, 0.009071225998923182, 0.00856252900121035, 0.010277663001033943, 0.008879124001396121, 0.00878319599723909, 0.008721453999896767, 0.008603109999967273, 0.008478084000671515, 0.008624321002571378, 0.008675285000208532, 0.0087622309983999, 0.009106567998969695, 0.008735316001548199, 0.00884358199982671, 0.008705772001121659, 0.008614843998657307, 0.00850751499820035, 0.008473002002574503, 0.008418708999670343, 0.008518423997884383, 0.008529103000910254, 0.008406189001107123, 0.008699682999576908, 0.008740058001421858, 0.008690335998835508, 0.008854293999320362, 0.008642986002087127, 0.008496685997670284, 0.008804044999124017, 0.008786717000475619, 0.008700612001121044, 0.008578036999097094, 0.008806718000414548, 0.008560721998946974, 0.008549817001039628, 0.008622638000815641, 0.008886758998414734, 0.008657130001665791, 0.008693877000041539, 0.0084848949991283, 0.008500872001604876, 0.008506835998559836, 0.008554132000426762, 0.009027945001435, 0.00854728899867041, 0.008557061999454163, 0.008510760999342892, 0.008565325999370543, 0.00860665400250582, 0.008575783998821862, 0.008566123000491643, 0.008644686000479851, 0.008680939998157555, 0.00857338400237495, 0.008556152999517508, 0.008603514997957973, 0.0085862520027149, 0.008621799999673385, 0.008592322999902535, 0.008833396997943055, 0.008646106001833687, 0.008576169999287231, 0.008647713999380358, 0.008536110999557422, 0.008605759001511615, 0.01040124699648004, 0.009158054002909921, 0.008914546000596602, 0.008726566997211194, 0.00872382200032007, 0.008482920999085763, 0.008581426001910586, 0.008733440001378767, 0.008699958998477086, 0.008680017999722622, 0.008673940999869956, 0.008634427002107259, 0.008818743997835554, 0.00861816700125928, 0.008619461998023326, 0.008732558002520818, 0.008587446998717496, 0.008527327998308465, 0.00882552199982456, 0.008640867999929469, 0.008656130001327256, 0.008581633999710903, 0.008866595999279525, 0.00860708300024271, 0.008601288001955254, 0.008901694000087446, 0.008536231998732546, 0.008490053998684743, 0.008572689002903644, 0.008605418999650283, 0.008540558999811765, 0.009003751998534426, 0.00866696299999603, 0.008735659001104068, 0.008686561999638798, 0.008990826998342527, 0.008706615000846796, 0.008682071002112934, 0.00930852999954368, 0.008581158999731997, 0.008672420000948478, 0.008636464997834992, 0.008714419000170892, 0.008554500000172993, 0.008649537001474528, 0.008496174999891082, 0.00859111299723736, 0.008466024002700578, 0.008488545998261543, 0.008567913002480054, 0.008622857996670064, 0.008657617003336782, 0.00870448399655288, 0.008650108000438195, 0.008643963003123645, 0.008932171997003024, 0.008794830002443632, 0.008668279999255901, 0.008803150998573983, 0.008654290999402292, 0.008804701999906683, 0.008527513000444742, 0.008669992999784881, 0.00979507800002466, 0.008865069001330994, 0.008840598999086069, 0.008691879000252811, 0.008661402000143426, 0.008646449001389556, 0.008530785999028012, 0.008453637001366587, 0.008529724000254646, 0.008522083000571001, 0.008400244998483686, 0.008410796999669401, 0.0085342540005513, 0.008448550001048716, 0.008464548998745158, 0.008732446000067284, 0.00852884299820289, 0.00848110200240626, 0.008526179997716099, 0.008524623000994325, 0.008680820999870775, 0.008705049000127474, 0.00880322599914507, 0.008862030001182575, 0.009004439001728315, 0.008982804996776395, 0.00894481500290567, 0.009041992998390924, 0.008680027000082191, 0.008715150001080474, 0.008995878000860102, 0.008580650999647332, 0.008669183000165503, 0.008826344997942215, 0.008651490999909583, 0.008700682999915443, 0.008724003000679659, 0.009024328999657882, 0.008771253000304569, 0.008727016000193544, 0.00924037699951441, 0.00880029700056184, 0.008916499999031657, 0.00881334600126138, 0.008818807000352535, 0.008706214997800998, 0.00872014900232898, 0.008725965999474283, 0.00879120700119529, 0.00890092799818376, 0.008619135001936229, 0.008671011997648748, 0.008682192001288058, 0.008845110998663586, 0.009004646999528632, 0.009026150000863709, 0.008760959000937873, 0.008719682999071665, 0.008702470000571338, 0.009120151000388432, 0.00876321399846347, 0.008910881999327103, 0.009129011999903014, 0.009890003999316832, 0.009142865001194878, 0.009106478999456158, 0.008928048999223392, 0.009822387000895105, 0.009069424002518645, 0.008994061998237157, 0.008966956000222126, 0.008736995998333441, 0.008624092002719408, 0.008587066000472987, 0.008421368998824619, 0.008476118000544375, 0.008474226997350343, 0.008576488002290716, 0.008574385999963852, 0.008481264998408733, 0.008671464001963614, 0.008636832997581223, 0.008700975002284395, 0.0085653669993917, 0.008599318000051426, 0.008540745999198407, 0.00864217100024689, 0.008549729998776456, 0.008600487002695445, 0.008520879000570858, 0.008564251998905092, 0.008774050998908933, 0.008680424998601666, 0.008611362001829548, 0.008840825001243502, 0.009197149996907683, 0.008784974001173396, 0.00880318800045643, 0.00880865599901881, 0.00867671600281028, 0.008670162998896558, 0.008580257999710739, 0.008794969999144087, 0.008619800999440486, 0.0086598439993395, 0.008687679001013748, 0.008646902002510615, 0.00865220999912708, 0.008470839999063173, 0.008608447998994961, 0.0088077489999705, 0.008828144000290195, 0.008854722000251058, 0.008769075000600424, 0.008579214998462703, 0.008608470001490787, 0.008734410999750253, 0.008488643001328455, 0.008488121999107534, 0.008568959001422627, 0.0084639819979202, 0.00866597999993246, 0.008795707999524893, 0.00868669700139435, 0.008536825000192039, 0.008820369999739341, 0.009764455000549788, 0.009008950997667853, 0.0088553720015625, 0.00884648699866375, 0.008889742999599548, 0.008790997002506629, 0.008816338999167783, 0.008633298999484396, 0.008613246001914376, 0.008541582999896491, 0.008570986999984598, 0.008537227997294394, 0.008574955001677154, 0.00866867600052501, 0.009757010997418547, 0.009788131999812322, 0.009086807000130648, 0.009398087000590749, 0.008808318001683801, 0.008673463999002706, 0.008849066998664057, 0.008635299000161467, 0.008672533000208205, 0.008862995000527008, 0.008592767000664026, 0.008558278001146391, 0.009050284999830183, 0.00862879700071062, 0.008679310998559231, 0.008724758998141624, 0.00911636599994381, 0.00870566300000064, 0.008686975001182873, 0.008692657000210602, 0.008941648000472924, 0.008698264999111416, 0.008837319001031574, 0.009016619998874376, 0.009380037001392338, 0.008941363001213176, 0.00881824999669334, 0.008681696002895478, 0.008802174997981638, 0.00857613200059859, 0.008519678998709423, 0.008547117002308369, 0.008605132999946363, 0.008610177999798907, 0.008641148000606336, 0.008746314997551963, 0.008834272000967758, 0.008920521999243647, 0.008866601001500385, 0.008703240000613732, 0.008722670998395188, 0.00881135700183222, 0.008919298998080194, 0.008977569999842672, 0.008841162001772318, 0.008901844998035813, 0.009034383001562674, 0.008919263000279898, 0.009082605000003241, 0.009946987996954704, 0.008977694000350311, 0.008751927001867443, 0.008692004998010816, 0.008584383002016693, 0.00913128599859192, 0.00875151500076754, 0.008715594998648157, 0.008681882001837948, 0.008768233998125652, 0.008619741001893999, 0.008547475998057052, 0.00868330000230344, 0.00862960299855331, 0.00875954500224907, 0.009103090997086838, 0.008737382002436789, 0.008863151997502428, 0.008784730001934804, 0.008780821000982542, 0.008633673998701852, 0.008831823000946315, 0.008787364000454545, 0.00868822799748159, 0.00910531599947717, 0.008627602001070045, 0.008749923999857856, 0.0087207260003197, 0.008979604001069674, 0.008767662999161985, 0.008775091999268625, 0.008958682999946177, 0.00878830200235825, 0.008861571997840656, 0.00911957700009225, 0.00922618199911085, 0.009254168002371443, 0.0091419969976414, 0.00932247700256994, 0.009097190999455051, 0.00923674599835067, 0.009218228999088751, 0.00953735200164374, 0.009125562999543035, 0.008962122999946587, 0.009208884999679867, 0.00961473800271051, 0.008863508999638725, 0.00896747299702838, 0.008966917001089314, 0.0088236470000993, 0.008772465000220109, 0.008712588001799304, 0.008703776999027468, 0.008775878999585984, 0.00884446000054595, 0.008872777998476522, 0.009031436002260307, 0.00898192999738967, 0.009089899001992308, 0.008997871998872142, 0.009068295999895781, 0.008891877001588, 0.009977458998037037, 0.009060186999704456, 0.00903945300160558, 0.008833536001475295, 0.008875637999153696, 0.008724831997824367, 0.00880994900217047, 0.008896392999304226, 0.008782142998825293, 0.008769225001742598, 0.008853098999679787, 0.008889871998690069, 0.009089453000342473, 0.008950565999839455, 0.00906674500220106, 0.009067413997399854, 0.009000693000416504, 0.009029439002915751, 0.008831601997371763, 0.009062547000212362, 0.0090100480010733, 0.009136595999734709, 0.009288591998483753, 0.009060350002982887, 0.009265589997085044, 0.008996784999908414, 0.009106663001148263, 0.009070405001693871, 0.008974102998763556, 0.008998000997962663, 0.008777811002801172, 0.008994408999569714, 0.008944986999267712, 0.009211990000039805, 0.009375794001243776, 0.008953724998718826, 0.008989940000901697, 0.009091463998629479, 0.009053162000782322, 0.008927014998334926, 0.00898617200073204, 0.009040286000526976, 0.008876878000592114, 0.008814673998131184, 0.009167888001684332, 0.008840940998197766, 0.008717323002201738, 0.008897678999346681, 0.008959663999121403, 0.009010409001348307, 0.008975025000836467, 0.009058687999640824, 0.00871403599740006, 0.008835645003273385, 0.00887802999932319, 0.009125361997575965, 0.009373977001814637, 0.008867369000654435, 0.008784213998296764, 0.009094532000744948, 0.008838320998620475, 0.008858003002387704, 0.009016284999233903, 0.010126794997631805, 0.009033194000949152, 0.00884330999906524, 0.008725252002477646, 0.008992605999083025, 0.008970957998826634, 0.00883921599961468, 0.008659751001687255, 0.00878520499827573, 0.008651546002511168, 0.008737719999771798, 0.00887012900057016, 0.008953436998126563, 0.008749462002015207, 0.008820863997243578, 0.009208700001181569, 0.008809194001514697, 0.008898903997760499, 0.009159158002148615, 0.008989590998680796, 0.008890049000910949, 0.008919873998820549, 0.009118189998844173, 0.009001883001474198, 0.008843130999593996, 0.009350808999442961, 0.008965981000073953, 0.00894272000005003, 0.008865514002536656, 0.008909503998438595, 0.008940922998590395, 0.008877055999619188, 0.008951329000410624, 0.00929916400127695, 0.008826414999930421, 0.00905255500038038, 0.008971242998086382, 0.00914397300221026, 0.009429023000848247, 0.00937587699809228, 0.009141409002040746, 0.009019465000164928, 0.008854256997437915, 0.008811900999717182, 0.008897353000065777, 0.009087411999644246, 0.008912217999750283, 0.008898797001165804, 0.009020675999636296, 0.008760133001487702, 0.009203964000334963, 0.008995144999062177, 0.008818818998406641, 0.009082219999982044, 0.00900225100122043, 0.008933842000260483, 0.008870433000993216, 0.009135241998592392, 0.008847854001942324, 0.008913707999454346, 0.009233140997821465, 0.009210333002556581, 0.009080840998649364]",tokens/s,114.5588068216356,, +float32-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neo,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cpu,,42,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,1143.107584,,,,,s,1,4.0587995480000245,4.0587995480000245,0.0,4.0587995480000245,4.0587995480000245,4.0587995480000245,4.0587995480000245,[4.0587995480000245],,kWh,1.947744179235037e-05,1.0614373952206124e-05,0,3.0091815744556492e-05,,MB,1203.48672,,,,,s,19,0.833764943999995,0.043882365473683946,0.000625064505694988,0.043716924999898765,0.04444004560100438,0.04532229999913397,0.04556897200065577,"[0.043716924999898765, 0.04528803999892261, 0.043691532000593725, 0.04563064000103623, 0.043213367000134895, 0.04334531999847968, 0.04298494499926164, 0.04380180299995118, 0.04411552599958668, 0.04413793300045654, 0.04341831600140722, 0.04390394400070363, 0.04402314300023136, 0.04352137699970626, 0.04380141900037415, 0.043608129999483936, 0.04366751399902569, 0.04366702299921599, 0.04422804700152483]",tokens/s,5833.778494769659,kWh,5.307851733495375e-07,2.8928982561102767e-07,0.0,8.20074998960565e-07,tokens/kWh,312166570.526448,MB,1204.588544,,,,,s,19,9.488216687999738,0.49937982568419675,0.004064820446410168,0.49825668199991924,0.5046494797999912,0.5055243601995244,0.5058301960400786,"[0.5014008569996804, 0.49825668199991924, 0.502270850000059, 0.4981829160005873, 0.49739025800045056, 0.49920912900051917, 0.5041972380004154, 0.5004290640008549, 0.5059066550002171, 0.5054818829994474, 0.4982358249999379, 0.5026601829995343, 0.5044413790001272, 0.4979378279986122, 0.4900455039987719, 0.4938750399996934, 0.4962062970007537, 0.4968013930010784, 0.4952877069990791]",tokens/s,126.15647801487404,kWh,5.869766659874607e-06,3.1991635989531544e-06,0.0,9.06893025882776e-06,tokens/kWh,6946795.068654912,,s,1197,9.482040092991156,0.007921503837085344,0.000215854106431482,0.007875440998759586,0.008170755600440316,0.00831041279998317,0.00878448019946518,"[0.008867268999892985, 0.008320558999912464, 0.007989735999217373, 0.007939490000353544, 0.00792609500058461, 0.008046635999562568, 0.008119583999359747, 0.008004776000234415, 0.007941302001199801, 0.0078408599983959, 0.00783658600084891, 0.007866990999900736, 0.0078118159999576164, 0.007839796000553179, 0.007834373998775845, 0.007893190000686445, 0.008024778000617516, 0.007762755998555804, 0.007691059001444955, 0.007737030999123817, 0.007834590000129538, 0.00792883400026767, 0.007801435000146739, 0.007908100000349805, 0.007979038999110344, 0.008051679000345757, 0.008038769999984652, 0.007965931999933673, 0.007894882999607944, 0.007912652999948477, 0.007802341000569868, 0.00783173799936776, 0.00792384800115542, 0.007860345000153757, 0.0078113159997883486, 0.007832723998944857, 0.00783181900078489, 0.008036361999984365, 0.008261105000201496, 0.007931871999971918, 0.007760114998745848, 0.007716725000136648, 0.007853389000956668, 0.007861914999011788, 0.00786592900112737, 0.00770129999909841, 0.007762709999951767, 0.007949722999910591, 0.00780911699985154, 0.007866642001317814, 0.007906777000243892, 0.007878239999627112, 0.007861824000428896, 0.00801151099949493, 0.00803524799994193, 0.008020866000151727, 0.0082817709990195, 0.008364680001250235, 0.00831300399840984, 0.008303728000100818, 0.008002067001143587, 0.008053281999309547, 0.008136831000228995, 0.008806838999589672, 0.008338704999914626, 0.008309765000376501, 0.008004985000297893, 0.008009651999600464, 0.008107555000606226, 0.008122298000671435, 0.00811691799935943, 0.007954946000609198, 0.008036147999519017, 0.007995818999916082, 0.008032334000745323, 0.008006131000001915, 0.007995657999344985, 0.007829350999600138, 0.007908338000561344, 0.007768392999423668, 0.007955961000334355, 0.008006246000149986, 0.008042608998948708, 0.008033642001464614, 0.008038504998694407, 0.007980555999893113, 0.00792389100024593, 0.007906655000624596, 0.00799240599917539, 0.007910752001407673, 0.00776776899874676, 0.007760997999866959, 0.007717855000009877, 0.007684412999878987, 0.007752229999823612, 0.007827193001503474, 0.007785173998854589, 0.007929131001219503, 0.007906005999757326, 0.007738033000350697, 0.007810197999788215, 0.00793658699876687, 0.007942218000607681, 0.007917447999716387, 0.007782475000567501, 0.007703764000325464, 0.0076397739994717995, 0.007695162999880267, 0.007786639000187279, 0.008112433999485802, 0.008089229000688647, 0.007733898999504163, 0.0077549990000989055, 0.007783908000419615, 0.007766922999508097, 0.007852978000300936, 0.0077692249997198815, 0.0077604739999515004, 0.007810578999851714, 0.007917019000160508, 0.007801558000210207, 0.007762331999401795, 0.007708740000452963, 0.007603501000630786, 0.0077078699996491196, 0.007794725999701768, 0.00878432699937548, 0.008231842999521177, 0.008074885001406074, 0.007972100998813403, 0.007858430999476695, 0.00784501199996157, 0.007816974000888877, 0.007740160999674117, 0.007889556000009179, 0.007877318999817362, 0.0078936120007711, 0.008373053000468644, 0.008796393000011449, 0.0085854829994787, 0.008166469999196124, 0.00828150400047889, 0.00782333099959942, 0.00798604200099362, 0.007956894998642383, 0.007999031000508694, 0.007854778999899281, 0.008032163001189474, 0.00803059600002598, 0.00806863699835958, 0.007951387000503018, 0.007801219000612036, 0.007856455998989986, 0.007823836000170559, 0.008035036000364926, 0.008022289999644272, 0.007946409001306165, 0.007977846998983296, 0.007901295999545255, 0.007847102000596351, 0.00772907400096301, 0.007873758999267011, 0.007876765999753843, 0.007934770999781904, 0.007838465000531869, 0.007838501000151155, 0.007820011000148952, 0.007895828999608057, 0.007837970000764471, 0.007729767999990145, 0.007809589998942101, 0.007992635000846349, 0.00802072899932682, 0.007869687000493286, 0.00795740899957309, 0.008043416999498731, 0.008203389001209871, 0.008227480999266845, 0.00812512000084098, 0.007961319999594707, 0.007931971000289195, 0.00788744300007238, 0.007825233999028569, 0.007806907000485808, 0.007812953999746242, 0.007807685000443598, 0.00772141099878354, 0.007730786001047818, 0.007754812999337446, 0.00881278300039412, 0.008322318999489653, 0.008193097999537713, 0.008218811000915593, 0.008043465999435284, 0.007984694999322528, 0.007890515000326559, 0.007783930999721633, 0.007744723001451348, 0.007981494998603011, 0.00785738800004765, 0.007891229000961175, 0.00790095999946061, 0.0077319690008152975, 0.007913616000223556, 0.00770572799956426, 0.007926295000288519, 0.007745003998934408, 0.007946605001052376, 0.008132416998705594, 0.007984650001162663, 0.008034376998693915, 0.008303540000270004, 0.007917078999525984, 0.007853270000850898, 0.007795662999342312, 0.007836758000848931, 0.00780620200021076, 0.0077893769994261675, 0.007798963000823278, 0.007812340998498257, 0.008127571001750766, 0.008119350999550079, 0.007857943999624695, 0.007745887000055518, 0.007778518000122858, 0.007786682999721961, 0.007681971999772941, 0.007757338000374148, 0.007896253999206237, 0.007937958000184153, 0.0077635960005864035, 0.007761574999676668, 0.007767061999402358, 0.007693873001699103, 0.007857493999836151, 0.007919279998532147, 0.007779654000842129, 0.007845829999496345, 0.007756412000162527, 0.007737658001133241, 0.007831758999600424, 0.007851065000068047, 0.00798823599870957, 0.007712969001659076, 0.007729209999524755, 0.0077863980004622135, 0.007816985998942982, 0.007896807001088746, 0.008039153999561677, 0.008070631000009598, 0.008046458999160677, 0.007874295999499736, 0.008937960999901406, 0.007969973001308972, 0.0078061519998300355, 0.007892360999903758, 0.007962009998664143, 0.007952961001137737, 0.00793606200022623, 0.007800085999406292, 0.007630719999724533, 0.007636857000761665, 0.007632830998772988, 0.007793078000759124, 0.007900029999291291, 0.007763260000501759, 0.007616947999849799, 0.0076981269994576, 0.007743856000161031, 0.0076365519998944364, 0.007588685000882833, 0.007566357999166939, 0.007628964000105043, 0.007717524000327103, 0.007851104999645031, 0.007654001001355937, 0.00776447199859831, 0.007806558000083896, 0.007819615000698832, 0.007765263000692357, 0.007835529999283608, 0.0077380300008371705, 0.007818102998498944, 0.007819145001121797, 0.007771825999952853, 0.007665148999876692, 0.007751466999252443, 0.007862270000259741, 0.00789068700032658, 0.007949667999127996, 0.007847352000680985, 0.007859896000809385, 0.008015127999897231, 0.007977904999279417, 0.008425852000073064, 0.008022942000025068, 0.008808754000710906, 0.007947090000016033, 0.007889257998613175, 0.00791453900092165, 0.008019922999665141, 0.007887711999501335, 0.007976292001330876, 0.007815186998413992, 0.007811537001543911, 0.007969239999511046, 0.00786995099952037, 0.007805874000041513, 0.007971856999574811, 0.008118077001199708, 0.008222884998758673, 0.008214508001401555, 0.008106481998765958, 0.008132150000164984, 0.007879175000198302, 0.008788157001617947, 0.008126683998852968, 0.007968575000631972, 0.008085379000476678, 0.008034015998418909, 0.007990892001544125, 0.008045165999646997, 0.008000292000360787, 0.007780308998917462, 0.007802517000527587, 0.007675255999856745, 0.00767652300055488, 0.007658233998881769, 0.00779833900014637, 0.007701271000769339, 0.007783955999911996, 0.007772765000481741, 0.007753986999887275, 0.007782176999171497, 0.007833039000615827, 0.00785000199903152, 0.007959395001307712, 0.007825592998415232, 0.007738179001535173, 0.007805395998730091, 0.007937970000057248, 0.008010019000721513, 0.008023046999369399, 0.00801098100055242, 0.007831951999833109, 0.007895217000623234, 0.00791832199865894, 0.007784424000419676, 0.007815113000106066, 0.007877387999542407, 0.007926603000669274, 0.008080263000010746, 0.008029743999941275, 0.008053629000642104, 0.007960863998960122, 0.007882346000769758, 0.008045378999668173, 0.0081290899997839, 0.008176671999535756, 0.008177147001333651, 0.007958223999594338, 0.007911987999250414, 0.007944349001263618, 0.007959429000038654, 0.007865097999456339, 0.007869269998991513, 0.007828835001419066, 0.007857130998672801, 0.00782092600093165, 0.007816178998837131, 0.00785557500057621, 0.007990158001121017, 0.008054101999732666, 0.007955680999657488, 0.007902721999926143, 0.007837404000383685, 0.007772168999508722, 0.007872553000197513, 0.008371830999749363, 0.007847098999263835, 0.007681131000936148, 0.007666691999475006, 0.007823558000382036, 0.007937881999168894, 0.007714305000263266, 0.007684485000936547, 0.007554404999609687, 0.0075885810001636855, 0.007744425000055344, 0.007778650999171077, 0.007861410000259639, 0.007782586000757874, 0.007734236000032979, 0.007746794999548001, 0.007838546000130009, 0.0077713409991702065, 0.007807212999978219, 0.007845334001103765, 0.007830590999219567, 0.007874586999605526, 0.007740558001387399, 0.007754786998702912, 0.007784856001308071, 0.00783174699972733, 0.007833448000383214, 0.008212943999751587, 0.008343886000147904, 0.008569931000238284, 0.0085543269997288, 0.008880194998710067, 0.008534444001270458, 0.00834032899911108, 0.008187235000150395, 0.008060744999966118, 0.00812768200012215, 0.008123896001052344, 0.00785748799899011, 0.007792984000843717, 0.0079273209994426, 0.00784973500049091, 0.007776309999826481, 0.007766258999254205, 0.00776549799957138, 0.007954366001285962, 0.008178170999599388, 0.008002181999472668, 0.007939891000205535, 0.007976300999871455, 0.007973018000484444, 0.008049800000662799, 0.008132750999720884, 0.008044729998800904, 0.00789218499994604, 0.008021499001188204, 0.008415831998718204, 0.008821186000204762, 0.008294040000691894, 0.008316774999912013, 0.008467772000585683, 0.008176295999874128, 0.008110278999083675, 0.008684877000632696, 0.008033668998905341, 0.008016491001399118, 0.007640516998435487, 0.007694020001508761, 0.008467871999528143, 0.00857608399928722, 0.008128655999826151, 0.008407303001149558, 0.00784826099879865, 0.007940121000501676, 0.007916665999800898, 0.008078560000285506, 0.008109974000035436, 0.008079743000052986, 0.00804498599973158, 0.008124195001073531, 0.008039573000132805, 0.008142802998918341, 0.0080398520003655, 0.008049001000472344, 0.007951497000249219, 0.007964058999277768, 0.008004662999155698, 0.007970877000843757, 0.007950545999847236, 0.008049415999266785, 0.007987376000528457, 0.007811216999471071, 0.007912558001407888, 0.007967322999320459, 0.007849531000829302, 0.007841139999072766, 0.007897440000306233, 0.007801801999448799, 0.007796587000484578, 0.007915604999652714, 0.00806784499945934, 0.00786948900167772, 0.008084171999144019, 0.0079751680004847, 0.007851626000046963, 0.007805769999322365, 0.007754902999295155, 0.007753260000754381, 0.007721600999502698, 0.007717263000813546, 0.007730375000392087, 0.007696229999055504, 0.007770821999656619, 0.007710874000622425, 0.007676557999729994, 0.007731011999567272, 0.007747151001240127, 0.0077652939999097725, 0.007681058999878587, 0.007912426999610034, 0.00796158900084265, 0.007845934998840676, 0.007833022000340861, 0.007956978999573039, 0.007874271999753546, 0.007848006000131136, 0.008506345000569127, 0.008052717999817105, 0.008275456999399466, 0.008295815001474693, 0.007829261998267611, 0.007838436000383808, 0.007829377000234672, 0.00780547199974535, 0.007843520001188153, 0.007785781999700703, 0.007753207999485312, 0.007666163999601849, 0.007726466001258814, 0.007812425999873085, 0.007967252999151242, 0.00797217599938449, 0.008025079001527047, 0.00792254499901901, 0.007885779999924125, 0.008209626999814645, 0.0079308280000987, 0.007825074000720633, 0.007920948999526445, 0.007794933000695892, 0.008362545999261783, 0.00804113599951961, 0.008254944999862346, 0.00797238200175343, 0.007869075998314656, 0.007738167001662077, 0.007987938999576727, 0.007942969999930938, 0.007978928999364143, 0.007926030000817263, 0.008444793998933164, 0.008061509999606642, 0.00780094900073891, 0.007944218999909936, 0.008037396999498014, 0.008038533000217285, 0.008215679001295939, 0.008120589998725336, 0.008261797000159277, 0.007938894999824697, 0.008002290000149515, 0.007966500001202803, 0.008033473999603302, 0.007987601000422728, 0.0080040579996421, 0.007933347998914542, 0.007997200000318117, 0.008036774001084268, 0.007945906998429564, 0.008314313001392293, 0.008512999998856685, 0.008417293000093196, 0.008324098000230151, 0.00830196300012176, 0.008155948000421631, 0.008252227000411949, 0.008174599999620114, 0.007975889999215724, 0.007849174000511994, 0.008746997000343981, 0.008182206998753827, 0.007943366001200047, 0.007988610999746015, 0.008008627999515738, 0.008047220000662492, 0.00824885899965011, 0.008197917999495985, 0.008083048000116833, 0.008108641999569954, 0.00806409500000882, 0.0078116220010997495, 0.007897220999439014, 0.00785805200030154, 0.007931850999739254, 0.007775431999107241, 0.007943266000438598, 0.007838888999685878, 0.007814696000423282, 0.00864860299952852, 0.007897670000602375, 0.007967015999383875, 0.008049254000070505, 0.008070773999861558, 0.008014046999960556, 0.007967056000779849, 0.008019213999432395, 0.007847156000934774, 0.007909669999207836, 0.00802067200129386, 0.008148588998665218, 0.008176232000550954, 0.00814206799987005, 0.008390626000618795, 0.008294451999972807, 0.008094638000329724, 0.00811021399931633, 0.007935292000183836, 0.007974293999723159, 0.008016603000214673, 0.00815374299963878, 0.008054489000642207, 0.008037280000280589, 0.007936041998618748, 0.007894606000263593, 0.0078545779997512, 0.007913883000583155, 0.007835316999262432, 0.007766355000057956, 0.007729615001153434, 0.007813182999598212, 0.0080506640006206, 0.007889196000178345, 0.007930421999844839, 0.008013909999135649, 0.008078340000793105, 0.008133932999044191, 0.007949980999910622, 0.008077115000560298, 0.00813682399893878, 0.007928565000838717, 0.007909445999757736, 0.007905788999778451, 0.008598309999797493, 0.00805225699878065, 0.007825333001164836, 0.007879490000050282, 0.0077741059994878015, 0.007787615999404807, 0.0077605879996553995, 0.007896862000052352, 0.007773979999910807, 0.00760440000158269, 0.007693280998864793, 0.007714215000305558, 0.007635930000105873, 0.007820106000508531, 0.008013979999304865, 0.007708821000051103, 0.007839064001018414, 0.0077412019982148195, 0.007931378000648692, 0.00783956300074351, 0.008284730000013951, 0.00823648699952173, 0.008108132000415935, 0.008124415999191115, 0.008034755001062877, 0.007760503998724744, 0.007818955000402639, 0.007698441000684397, 0.007657956999537419, 0.00765997599955881, 0.007861932001105743, 0.007987889999640174, 0.00808468799914408, 0.008043040999837103, 0.008074967001448385, 0.008236478999606334, 0.008088769000096363, 0.00807801299924904, 0.008082259999355301, 0.008066752001468558, 0.008002558999578469, 0.008047838000493357, 0.007924083000034443, 0.007790948999172542, 0.007693543000641512, 0.007854026998757035, 0.00789477600119426, 0.007880115999796544, 0.007886790999691584, 0.007870757999626221, 0.007917132999864407, 0.007827040000847774, 0.007751162000204204, 0.0077548310000565834, 0.007916603999547078, 0.007987011000295752, 0.00798163800027396, 0.007872697999118827, 0.007924014000309398, 0.007916468999610515, 0.007743785999991815, 0.007784766999975545, 0.007832159999452415, 0.009161272999335779, 0.008502321999912965, 0.008289956000226084, 0.008084343999144039, 0.00791375700100616, 0.007770525000523776, 0.007695648999288096, 0.007707347000177833, 0.007652610998775344, 0.007640233001438901, 0.007576529000289156, 0.007759697999063064, 0.007624518999364227, 0.0076117480002722004, 0.007768737001242698, 0.007830394999473356, 0.007921144999272656, 0.008755341999858501, 0.008556775999750243, 0.007867315000112285, 0.007856448000893579, 0.00778366700069455, 0.00806262499827426, 0.007956938999996055, 0.00804106400028104, 0.0080686710007285, 0.007762422999803675, 0.0078012510002736235, 0.00763852399904863, 0.007863824001105968, 0.0077340099996945355, 0.007898903999375761, 0.00788345900036802, 0.007920662999822525, 0.007862652000767412, 0.007869687999118469, 0.00806664300034754, 0.00788147400089656, 0.007879965000029188, 0.00784626599852345, 0.007917234001070028, 0.0079105980003078, 0.00806681899848627, 0.007932043999971938, 0.007837622000806732, 0.007648245000382303, 0.007756115999654867, 0.008161828000083915, 0.008263274999990244, 0.008186114999261918, 0.008000474001164548, 0.00803824899958272, 0.007947090000016033, 0.007859794999603764, 0.00806740200096101, 0.00813634699989052, 0.008241913999881945, 0.007998265000423999, 0.008065264999459032, 0.008144813999024336, 0.00840782900013437, 0.008135310999932699, 0.008201783000913565, 0.009244062999641756, 0.008162750000337837, 0.007972637000420946, 0.00837618999867118, 0.007753587000479456, 0.007673308000448742, 0.007866855999964173, 0.00797349300046335, 0.008068639999692095, 0.007890798999142135, 0.007935067000289564, 0.00781722400097351, 0.007813146999978926, 0.007811128998582717, 0.00783081200097513, 0.007697854000070947, 0.0077175320002425, 0.007872454998505418, 0.007989802999873064, 0.007881275001636823, 0.007861857999159838, 0.007858283999667037, 0.007829862001017318, 0.007990567999513587, 0.007962288000271656, 0.007947423999212333, 0.008350760999746853, 0.008065601999987848, 0.007975107000675052, 0.00803695600006904, 0.00796230900050432, 0.008032517000174266, 0.008080570998572512, 0.008120918000713573, 0.008043413999985205, 0.008118318999549956, 0.008121320001009735, 0.007872616999520687, 0.008043620999160339, 0.008170714001607848, 0.00825384499876236, 0.008153320000928943, 0.008245622999311308, 0.008154384999215836, 0.007971358001668705, 0.007989491999978782, 0.007998438999493374, 0.007989468000232591, 0.008001774000149453, 0.008012204998522066, 0.008192123001208529, 0.008098947999314987, 0.007997243999852799, 0.008010243000171613, 0.007803247999618179, 0.007754010001008282, 0.007833550000214018, 0.00792728799933684, 0.007937001000755117, 0.007950831999551156, 0.007971790999363293, 0.008069946999967215, 0.007979605999935302, 0.008637008999357931, 0.008030815999518381, 0.008004143001016928, 0.008010031999219791, 0.0078704050010856, 0.007868486998631852, 0.008446339001238812, 0.008228634000261081, 0.008155682999131386, 0.008024532999115763, 0.00782267700014927, 0.008141320000504493, 0.007998467001016252, 0.008170817998689017, 0.008221785999921849, 0.008302738000566023, 0.008281288000944187, 0.00830076599959284, 0.008196214999770746, 0.008013978000235511, 0.007952291000037803, 0.007985512998857303, 0.00790087400127959, 0.007686659999308176, 0.007713316999797826, 0.00775260599948524, 0.0076485750014398945, 0.007707630998993409, 0.00781850699968345, 0.007827348999853712, 0.008152031999998144, 0.007834035000996664, 0.007815323000613716, 0.007740637998722377, 0.007857046999561135, 0.007823693000318599, 0.007768395000312012, 0.007902927000031923, 0.007840739999664947, 0.007862301001296146, 0.007855527999709011, 0.007835565000277711, 0.00776503499946557, 0.0077565640003740555, 0.007593836999149062, 0.007642466000106651, 0.007682024001042009, 0.007791275998897618, 0.007946117999381386, 0.007671123001273372, 0.007657104999452713, 0.007639120000021649, 0.007566817999759223, 0.007821381001122063, 0.007650513998669339, 0.00786787099968933, 0.007762386001559207, 0.0076987209995422745, 0.0077983339997445, 0.007833874000425567, 0.007742783000139752, 0.007829936999769416, 0.007898968000517925, 0.008721586998944986, 0.008306550000270363, 0.008108722000542912, 0.008030461000089417, 0.00795367299906502, 0.007825280999895767, 0.007898134001152357, 0.00774509599978046, 0.007759239999359124, 0.007572407999759889, 0.007596629000545363, 0.007538650999777019, 0.007529766000516247, 0.007652059999600169, 0.0077251650000107475, 0.007777786999213276, 0.007741075000012643, 0.00778965500103368, 0.007740855999145424, 0.007680389000597643, 0.00762605199997779, 0.007636288000867353, 0.007585044999359525, 0.007634571000380674, 0.007762383000226691, 0.00772942699950363, 0.007761245000438066, 0.007722661999650882, 0.007714547999057686, 0.007739771999695222, 0.007834594000087236, 0.007778852001138148, 0.007765373000438558, 0.007747818999632727, 0.007796459998644423, 0.0077265540003281785, 0.007801944000675576, 0.0077637310005229665, 0.007912590999694658, 0.007850357000279473, 0.007841132999601541, 0.007868226999562467, 0.007879816999775358, 0.007698065999647952, 0.0077359540009638295, 0.0076833959992654854, 0.007786251000652555, 0.007675139000639319, 0.007857949000026565, 0.007927475999167655, 0.007862790000217501, 0.007657285999812302, 0.007630633999724523, 0.007538801000919193, 0.007561308999356697, 0.007652992999283015, 0.007632814000317012, 0.0077328839997790055, 0.007890583001426421, 0.007757396999295452, 0.007756823999443441, 0.007770402000460308, 0.007734090999292675, 0.008706717999302782, 0.008147184000335983, 0.007978962999914074, 0.007871489000535803, 0.007767135999529273, 0.007762941000692081, 0.007767629998852499, 0.007759142999930191, 0.007711120000749361, 0.007763477000480634, 0.0077529809987026965, 0.00785099000131595, 0.007832016999600455, 0.00790350299939746, 0.00772984200011706, 0.007883415000833338, 0.007716902999163722, 0.007842384000468883, 0.007869686000049114, 0.007936836000226322, 0.007866836998800864, 0.00786306600093667, 0.00763957399976789, 0.007691736000197125, 0.0079106689991022, 0.007875837000028696, 0.007814659000359825, 0.007657014999495004, 0.007742961000985815, 0.0077297560001170496, 0.007728102000328363, 0.007688831999985268, 0.007764683999994304, 0.007757477998893592, 0.007785269001033157, 0.007974366999405902, 0.008103001000563381, 0.007944605000375304, 0.007904559999587946, 0.00785335199907422, 0.007839256000806927, 0.00793994599916914, 0.007926772999780951, 0.00820974400085106, 0.00801960299941129, 0.007763503999740351, 0.0078042060013103765, 0.007819450998795219, 0.007720340001469594, 0.007910489999630954, 0.007829893000234733, 0.007772928000122192, 0.007740876999378088, 0.00761578899982851, 0.007729405999270966, 0.007723847000306705, 0.007666540001082467, 0.007721307998508564, 0.007724219000010635, 0.0077166790015326114, 0.007754498999929638, 0.007850351999877603, 0.00792407199878653, 0.008472901001368882, 0.008048589999816613, 0.007904187999884016, 0.0077702910002699355, 0.007740024999293382, 0.007916488999399007, 0.007646769001439679, 0.007500721998439985, 0.007638542001586757, 0.007663555999897653, 0.007692612000028021, 0.007671411000046646, 0.007726357000137796, 0.007878477999838651, 0.007682340999963344, 0.007717643999058055, 0.007808813999872655, 0.00769390000095882, 0.007790927998939878, 0.007829064999896218, 0.007886139001129777, 0.007877669999288628, 0.007827437999367248, 0.0078063030014163814, 0.0077138010001363, 0.007740803999695345, 0.007862057998863747, 0.007937103000585921, 0.007825687000149628, 0.0077271459995245095, 0.00777829000071506, 0.007832409999537049, 0.00789874600013718, 0.00792478299990762, 0.0077652760010096245, 0.007895062999523361, 0.007888746000389801, 0.0077281959984247806, 0.0077410630001395475, 0.007885556000474025, 0.007964410000568023, 0.007922447999590077, 0.008008900000277208, 0.008033972999328398, 0.007931104000817868, 0.00832873000035761, 0.008296019999761484, 0.008237423999162274, 0.007920987000034074, 0.00803899799939245, 0.008045241000218084, 0.008014048000404728, 0.007873200000176439, 0.007746690000203671, 0.007817476000127499, 0.007996413000000757, 0.007960937000461854, 0.00792442499914614, 0.007919855999716674, 0.007864323000831064, 0.007826877999832504, 0.007862733998990734, 0.007937901000332204, 0.008628212000985513, 0.008025526998608257, 0.00779466499989212, 0.007713394001257257, 0.007690675000048941, 0.007875440998759586, 0.007809934000761132, 0.007731175999651896, 0.007710947000305168, 0.007757585999570438, 0.007800028999554343, 0.007769632000417914, 0.007561439000710379, 0.007584700999359484, 0.0076197520011191955, 0.007618863999596215, 0.00767751499915903, 0.007740453000224079, 0.007821655000952887, 0.007830857999579166, 0.007645102999958908, 0.007671544999539037, 0.007675651999306865, 0.007735383000181173, 0.007796577001499827, 0.007856117999835988, 0.007826604998626863, 0.007792686001266702, 0.007860879999498138, 0.007783544000631082, 0.007843520999813336, 0.007743950998701621, 0.007732838001174969, 0.007812359999661567, 0.008057050999923376, 0.008027478999792947, 0.007860730000174954, 0.007882407999204588, 0.007799485001669382, 0.0077911299995321315, 0.0076847450000059325, 0.0076371800005290424, 0.007775981999657233, 0.0076067310001235455, 0.00779806199898303, 0.00788307699986035, 0.007957334000820993, 0.007907303999672877, 0.008114954000120633, 0.008320417999129859, 0.008121600001686602, 0.00811696099844994, 0.008162388001437648, 0.008238910999352811, 0.008237306999944849, 0.008422543000051519, 0.008129920999635942, 0.008020237999517121, 0.008030689999941387, 0.008058999001150369, 0.008124269999825628, 0.008172149999154499, 0.008013931999812485, 0.008869110999512486, 0.008225899999160902, 0.008120021000650013, 0.007828539999536588, 0.007782774000588688, 0.007857426999180461, 0.007794127999659395, 0.007869720000599045, 0.00784440099960193, 0.007775500000207103, 0.007781112000884605, 0.00773063899941917, 0.007749145001071156, 0.007815594999556197, 0.007821764998880099, 0.008004870000149822, 0.007813203999830876, 0.007747299001493957, 0.00783268199847953, 0.007741192001049058, 0.007640411999091157, 0.007837515000574058, 0.007819208000000799, 0.007760489999782294, 0.007826932000170927, 0.007772543000100995, 0.007813174999682815, 0.007937976999528473, 0.007817595000233268, 0.00792059400009748, 0.007916771000964218, 0.007951335999678122, 0.0077916629998071585, 0.007873739999922691, 0.00783057400076359, 0.007774287998472573, 0.007827575000192155, 0.007752688001346542, 0.007850239999243058, 0.007749570000669337, 0.007754249998470186, 0.007723597000222071, 0.0077291650013648905, 0.007680633998461417, 0.00773988600121811, 0.0077428160002455115, 0.007722205999016296, 0.007787036000081571, 0.007815824999852339, 0.007823208999980125, 0.007863947999794618, 0.007953940001243609, 0.007866878999266191, 0.007869224000387476, 0.00791172400022333, 0.007959128999573295, 0.007959959000800154, 0.008042999999815947, 0.007979955998962396, 0.007933574999697157, 0.007857368000259157, 0.007820541000910453, 0.00797530099953292]",tokens/s,126.23865626604838,, float32-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cpu,,42,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): File ""/workspace/llm_perf/common/benchmark_runner.py"", line 106, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) @@ -9727,113 +10144,461 @@ ChildProcessError: Traceback (most recent call last): AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-eager-pytorch,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stablelm,stabilityai/stablelm-2-12b,stabilityai/stablelm-2-12b,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cpu_pytorch.py"", line 114, in benchmark_cpu_pytorch +float16-eager-pytorch,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stablelm,stabilityai/stablelm-2-12b,stabilityai/stablelm-2-12b,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cpu_pytorch.py"", line 114, in benchmark_cpu_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 174, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float16-eager-pytorch,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,26443.382784,,,,,s,1,416.0725446910001,416.0725446910001,0.0,416.0725446910001,416.0725446910001,416.0725446910001,416.0725446910001,[416.0725446910001],,kWh,0.004881660385143748,0.002660693515757511,0,0.007542353900901259,,MB,27043.291136,,,,,s,10,35.84458376100338,3.5844583761003377,0.007329127151613124,3.581920324500061,3.5933664820999183,3.59342874005024,3.593478546410497,"[3.5934909980005614, 3.5806673650004086, 3.578081989000566, 3.570503140001165, 3.5930893710010423, 3.582125721999546, 3.5900059239993425, 3.593352646999847, 3.5817149270005757, 3.5815516780003236]",tokens/s,71.41943723127054,kWh,4.222885850826313e-05,2.3016187507371003e-05,0.0,6.524504601563414e-05,tokens/kWh,3923669.544790524,MB,27315.351552,,,,,s,10,191.71148664899556,19.171148664899555,0.045699471895773496,19.166182255499734,19.235882381699593,19.243812118349616,19.250155907669633,"[19.156944663000104, 19.202089746999263, 19.111819742998705, 19.251741854999636, 19.23412021799959, 19.175419847999365, 19.133905310000046, 19.11193912899944, 19.18669035899984, 19.146815776999574]",tokens/s,3.286188068393975,kWh,0.0002262354940516688,0.00012330715505750746,0.0,0.00034954264910917624,tokens/kWh,180235.51678331124,,s,630,191.70293175499864,0.30429036786507724,0.0018862060580412715,0.30394128000079945,0.30653846689892816,0.3081794959993204,0.3102356020798288,"[0.3071339079997415, 0.3039268569991691, 0.3025128429999313, 0.3030618900011177, 0.3028014229985274, 0.3019031950007047, 0.30257120600072085, 0.30231144499884977, 0.3047235340000043, 0.3035496350003086, 0.3037798580007802, 0.3035140769989084, 0.30473717200038664, 0.30349296900021727, 0.3031198520002363, 0.30223485699934827, 0.3026049550007883, 0.30127563899986853, 0.30157310299910023, 0.3016757030000008, 0.30238696800006437, 0.30201786100042227, 0.3015998770006263, 0.3017297599999438, 0.3029425739987346, 0.30220787200050836, 0.3097000960005971, 0.3053774509990035, 0.30325844300023164, 0.30385756800023955, 0.3023264450002898, 0.31021991899979184, 0.3062353109999094, 0.304095462000987, 0.3034545939990494, 0.304184388000067, 0.30408804999933636, 0.30488553600116575, 0.3042747050003527, 0.3052109849995759, 0.30435396199936804, 0.3056618350001372, 0.3046495729995513, 0.30325742900095065, 0.30284192799990706, 0.30342087899953185, 0.30393571199965663, 0.30349379400104226, 0.30432189899875084, 0.30387068500021996, 0.3044745170009264, 0.3044783900004404, 0.304281220998746, 0.30435262200080615, 0.3031663660003687, 0.3057132529993396, 0.3031677889994171, 0.3046165600007953, 0.30329834499934805, 0.31053031100054795, 0.3069283929999074, 0.30631289099983405, 0.3083560579998448, 0.3099601960002474, 0.30365896900002554, 0.30428311600007873, 0.30325002999961725, 0.30206570999871474, 0.3032417680005892, 0.3028044910006429, 0.30219072500040056, 0.3037480479997612, 0.3034019009992335, 0.3028087679995224, 0.3039518900004623, 0.3015947940002661, 0.3033112920002168, 0.3034547150000435, 0.3023228959991684, 0.3039195050005219, 0.30988555399926554, 0.3029087830000208, 0.3025268410001445, 0.304003578999982, 0.3037717960014561, 0.3034793679998984, 0.30260862700015423, 0.3040192689986725, 0.3036694020011055, 0.30422015099975397, 0.3045588280001539, 0.303696613998909, 0.3058660469996539, 0.3066598460009118, 0.30643868100014515, 0.30505161300061445, 0.3056245999996463, 0.3037997859992174, 0.30484869400061143, 0.3047657219995017, 0.3051213190010458, 0.30393826899853593, 0.30445818299995153, 0.3056851920009649, 0.3061037080005917, 0.30417158999989624, 0.3059934759985481, 0.30597544600095716, 0.3052456740006164, 0.30446598199887376, 0.304912695000894, 0.3040380839993304, 0.30489457799922093, 0.31170258200108947, 0.3047725269989314, 0.3058464810001169, 0.3063942250009859, 0.30507593899892527, 0.30452177100050903, 0.3080648489994928, 0.3061719080014882, 0.3070284819987137, 0.3064574250001897, 0.3045297570006369, 0.30587915599971893, 0.3076493959997606, 0.309199006000199, 0.30197966399828147, 0.3033080010009144, 0.3019802779999736, 0.30310353099957865, 0.30167388300105813, 0.30080179299875454, 0.30394411900124396, 0.3090539909990184, 0.30344218900063424, 0.30300476899901696, 0.30278881400045066, 0.3026891789995716, 0.3026805610006704, 0.30275334699945233, 0.30164107699965825, 0.3024530880011298, 0.3029965120003908, 0.30252654399919265, 0.3018854060010199, 0.3022042449993023, 0.30271324800014554, 0.30203671299932466, 0.3019465670004138, 0.3048901799993473, 0.3017107870000473, 0.3030862610012264, 0.3032648779990268, 0.30316766100077075, 0.3029588609988423, 0.3036780930015084, 0.3025255249995098, 0.30376672099919233, 0.3026055690006615, 0.30276281199985533, 0.30329071300002397, 0.3020725999995193, 0.30294859199966595, 0.30085602500003006, 0.30239806400095404, 0.3033662859998003, 0.3106209540001146, 0.30315211199922487, 0.3027512620010384, 0.3037636509998265, 0.3034309550002945, 0.3029242800002976, 0.3040668369994819, 0.3041110390004178, 0.30279789799897117, 0.30367870699956256, 0.3040857230007532, 0.30382603900034155, 0.3036548359996232, 0.3055684699993435, 0.303714249001132, 0.3026670399995055, 0.30504507399928116, 0.30426826500115567, 0.30346402999930433, 0.3039348599995719, 0.3037278360006894, 0.30348297400087176, 0.30950051399850054, 0.30279291800070496, 0.3041378130001249, 0.3028259620004974, 0.3072156759990321, 0.3051922200011177, 0.3038096169984783, 0.30326598700048635, 0.30468586000097275, 0.30483968399857986, 0.30671483900005114, 0.3090727760009031, 0.3098011370002496, 0.303593364000335, 0.3031097629991564, 0.30606986400016467, 0.30373424500066903, 0.3043320519991539, 0.3036697820007248, 0.30420030699860945, 0.30410157600090315, 0.304636742999719, 0.30415662999985216, 0.30533455800105, 0.30542951399911544, 0.3043979819995002, 0.30566189999990456, 0.3049073890015279, 0.3064194989983662, 0.30393844100035494, 0.30636387300000933, 0.30532961599965347, 0.3118024380000861, 0.3048845710000023, 0.305132166000476, 0.3066085509999539, 0.3041491029998724, 0.30488976200103934, 0.3045865229996707, 0.3049149269991176, 0.3058463010002015, 0.3056846060007956, 0.3051623409992317, 0.30586545200094406, 0.30559523800002353, 0.30479139399903943, 0.3054030399998737, 0.30520066800090717, 0.3077267219996429, 0.30685654700027953, 0.3059312030000001, 0.306103744000211, 0.30559365899898694, 0.3064166580006713, 0.30569289199956984, 0.30572045400003844, 0.30701942599989707, 0.30571903599957295, 0.3060324310008582, 0.30682470799911243, 0.3077705949999654, 0.30743252000138455, 0.3063038479995157, 0.30876606600031664, 0.30421788199964794, 0.30487888700008625, 0.304530364999664, 0.304191950999666, 0.3061068349998095, 0.305866648999654, 0.30420380100076727, 0.30358054299904325, 0.30448782200073765, 0.30396707699947, 0.30378431500139413, 0.3035885219996999, 0.30462202799935767, 0.30575387899989437, 0.3035869690011168, 0.30352741799833893, 0.3040374280008109, 0.30443981700045697, 0.30446703499910655, 0.30337889200018253, 0.30415278200052853, 0.3036101130001043, 0.31023782000011124, 0.3051217279989942, 0.30408917299973837, 0.3027903310012334, 0.3071485990003566, 0.3045454069997504, 0.3075414550003188, 0.3053073539995239, 0.30472302899943315, 0.30568266400041466, 0.3050216850006109, 0.305744938999851, 0.30641166399982467, 0.3047887089996948, 0.30528842599960626, 0.30465976300001785, 0.3076309530006256, 0.30723267599933024, 0.3056978250006068, 0.30652790000021923, 0.3061542750001536, 0.30450233400006255, 0.3052837709983578, 0.30549265000081505, 0.30438526699981594, 0.3051304060008988, 0.3046704389998922, 0.30587790199933806, 0.305211655999301, 0.30582402200161596, 0.3050026799992338, 0.3043625650007016, 0.30759193599988066, 0.31111400699955993, 0.3048560320003162, 0.3065619469998637, 0.3046590039994044, 0.305204563001098, 0.30615207699884195, 0.30527903599977435, 0.30840713799989317, 0.3018237220003357, 0.30143443499946443, 0.3043430210000224, 0.30393834999995306, 0.3034137089998694, 0.30247743900144997, 0.3020826729989494, 0.3028766560000804, 0.30363115100044524, 0.30283857399990666, 0.30211022599905846, 0.3034588140017149, 0.30337308999878587, 0.3090060699996684, 0.30367535799996404, 0.304038625999965, 0.3025829320013145, 0.3028827050002292, 0.30382159299915656, 0.30435478999970655, 0.3045443669998349, 0.30382548099987616, 0.30415216800065537, 0.3040650309994817, 0.30309074400065583, 0.3024819749989547, 0.30229242800123757, 0.3038300909993268, 0.3034871039999416, 0.30403829900023993, 0.30193810799937637, 0.30156624900155293, 0.3019941779984947, 0.3027257930007181, 0.30279468499975337, 0.3060970839997026, 0.30428133600071305, 0.3047387799997523, 0.30598383100004867, 0.30208902500089607, 0.3032517639985599, 0.3037736610003776, 0.30223546299930604, 0.3028673759999947, 0.30340071900172916, 0.30582628099909925, 0.3097822029994859, 0.30445412500012026, 0.30597398999998404, 0.3031032390008477, 0.30709261300034996, 0.30822156199974415, 0.30498238600011973, 0.3040002079997066, 0.30621250699914526, 0.30972251900129777, 0.3100124129996402, 0.30746895399897767, 0.30763929700151493, 0.30646571299985226, 0.3056797359986376, 0.3056611360007082, 0.30770400400069775, 0.3041033299996343, 0.3027220710009715, 0.3027192270001251, 0.309242623998216, 0.30391503400096553, 0.3023700590001681, 0.3050273600001674, 0.3028769149987056, 0.3022433020014432, 0.3028958409995539, 0.3026022539997939, 0.3009949709994544, 0.30217691200050467, 0.30339289699986693, 0.304207070999837, 0.30305328400027065, 0.30387588200028404, 0.3032120930001838, 0.30296895500032406, 0.3036315699991974, 0.30458899199948064, 0.30367828000089503, 0.30248664499958977, 0.3035238680004113, 0.3026806170000782, 0.3032152870000573, 0.3032942299996648, 0.3047138090005319, 0.30486556100004236, 0.30428716899950814, 0.30526818900034414, 0.304161498999747, 0.3047236759994121, 0.3037033970012999, 0.30497962399931566, 0.3039448830004403, 0.3102301719991374, 0.30448641300063173, 0.3032105610000144, 0.30426809699929436, 0.3037427060007758, 0.3030780440003582, 0.30323906499870645, 0.30319408099967404, 0.3026865540014114, 0.3025963079999201, 0.3029844579996279, 0.30289611499938474, 0.30230070400102704, 0.3057393549988774, 0.3033140530005767, 0.3044351270000334, 0.30327862599915534, 0.3050604840009328, 0.3034374239996396, 0.30353691799973603, 0.302935386000172, 0.30217411200101196, 0.30169431899958, 0.3038251389989455, 0.30246080700089806, 0.302132722999886, 0.3085780179990252, 0.3040854280006897, 0.30579221100015275, 0.30294481300006737, 0.3022818949993962, 0.303253575000781, 0.30197955899893714, 0.3026492600001802, 0.30352982899967174, 0.3031408150000061, 0.3030467120006506, 0.30404755999916233, 0.30528696200053673, 0.3032578599995759, 0.3037959650009725, 0.30333046600026137, 0.30418211599862843, 0.3043877620002604, 0.3038736730013625, 0.30336882099982176, 0.3028488939999079, 0.3036753010001121, 0.30444197799988615, 0.30302207800014, 0.30220265799835033, 0.30315436300043075, 0.30401704700125265, 0.3022796229997766, 0.30598990399994364, 0.30479422099961084, 0.30198419899898, 0.3020516590004263, 0.3029592680013593, 0.30269775799933996, 0.3023366069992335, 0.3023354670003755, 0.3021338510006899, 0.3021994509999786, 0.3026061549990118, 0.30240087200036214, 0.3032777660009742, 0.30331159599882085, 0.3037337839996326, 0.30326724800033844, 0.30248569499963196, 0.3041771360003622, 0.30376638599955186, 0.30348570100068173, 0.30307136100054777, 0.3023486469992349, 0.30209753300005104, 0.3030255430003308, 0.3012353030007944, 0.3029088969997247, 0.302289010000095, 0.3030771609992371, 0.30318832000011753, 0.30260597599954053, 0.30250898099984624, 0.30168785900059447, 0.30455251099920133, 0.308270095001717, 0.30380233399955614, 0.3065358579988242, 0.30344057900038024, 0.30162154600111535, 0.30159257499872183, 0.3038788600006228, 0.3012045070008753, 0.3024113709998346, 0.30266211399975873, 0.30216416499933985, 0.30507843799932743, 0.303988528001355, 0.3043352770000638, 0.30305960799887544, 0.30228713900032744, 0.3023118519995478, 0.3026115630000277, 0.30219082000076014, 0.3030412320003961, 0.30365486399932706, 0.3035577589998866, 0.310145363000629, 0.30449719699936395, 0.30427877900001477, 0.3023436679995939, 0.30800867000107246, 0.3079293310001958, 0.3047548819995427, 0.3044389469996531, 0.30331301000114763, 0.3059470919997693, 0.306084956999257, 0.30417437399955816, 0.30536087899963604, 0.3051035300013609, 0.30372757599980105, 0.30313271100021666, 0.3049493119997351, 0.30590195799959474, 0.30390774099942064, 0.30481824500020593, 0.3037258060012391, 0.3048097399987455, 0.3043334110006981, 0.3045616519993928, 0.3034695759997703, 0.3043970310009172, 0.3039207929996337, 0.3040317180002603, 0.3030412640000577, 0.3029629800003022, 0.3081280819988024, 0.3050551190008264, 0.3103016259992728, 0.30470996199983347, 0.30440754600022046, 0.3064451589998498, 0.3057579300002544, 0.3069333610001195, 0.3072845080005209, 0.30722127899935003, 0.30538110900124593, 0.3048966419992212, 0.3056088350003847, 0.31009881300087727, 0.3044546490000357, 0.30394592299853684, 0.3044848270001239, 0.30490589500004717, 0.30348448200129496, 0.30439824999848497, 0.3050631410005735, 0.3030040680005186, 0.30356067399952735, 0.30950715400103945, 0.30460050299916475, 0.3032100229993375, 0.30401359500137914, 0.30306150499927753, 0.3041821099996014, 0.3028111010007706, 0.30233202899944445, 0.30211140100072953, 0.3030771559988352, 0.3032807489998959, 0.30288984199978586, 0.3031511880017206, 0.30357551899942337, 0.3027297329990688, 0.3043640410014632, 0.3040234629988845, 0.302651773999969, 0.30261192500074685, 0.30164105699986976, 0.3018576579997898, 0.3009523460004857, 0.30081379399962316, 0.30222629200034135, 0.3022622459993727, 0.30430213200088474, 0.3029304659994523, 0.3037095500003488, 0.3037285529990186, 0.3040030840002146, 0.30296583000017563, 0.3032651609992172, 0.30354356700081553, 0.3092657279994455, 0.3046112350002659, 0.304800977000923, 0.30445797799984575, 0.3033229329994356, 0.3023557960004837, 0.30576182900040294, 0.30506530099955853, 0.303907702998913, 0.3042684280007961, 0.30499370799952885, 0.3061601130011695, 0.3047164689996862, 0.30355477999910363, 0.3034126210004615, 0.3035775600001216, 0.3042765699992742, 0.3038325610013999, 0.30556072799845424, 0.3043188170013309]",tokens/s,3.286334717119279,, +float16-eager-pytorch,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt2,openai-community/gpt2,openai-community/gpt2,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cpu_pytorch.py"", line 114, in benchmark_cpu_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 174, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float16-eager-pytorch,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,codegen,Salesforce/codegen-6B-nl,Salesforce/codegen-6B-nl,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cpu_pytorch.py"", line 114, in benchmark_cpu_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 174, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float16-eager-pytorch,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cpu_pytorch.py"", line 114, in benchmark_cpu_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking + context_stack.enter_context(energy_tracker.track()) + File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context + result = _cm_type.__enter__(cm) + File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ + return next(self.gen) + File ""/workspace/optimum_benchmark/trackers/energy.py"", line 174, in track + self.emission_tracker.start_task() + File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task + if self._scheduler: +AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float32-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cpu,,42,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,28892.696576,,,,,s,1,70.74499230299989,70.74499230299989,0.0,70.74499230299989,70.74499230299989,70.74499230299989,70.74499230299989,[70.74499230299989],,kWh,0.0008070688300270844,0.0004398825798870292,0,0.0012469514099141136,,MB,29343.924224,,,,,s,10,17.60650355799953,1.760650355799953,0.012717061771106623,1.7565232184997512,1.7740552504003062,1.7817817926998942,1.7879630265395645,"[1.7467989820015646, 1.7543419970006653, 1.7723382410003978, 1.789508334999482, 1.7686037460007356, 1.7507506499987358, 1.7464263879992359, 1.7646887819992116, 1.75758004599993, 1.7554663909995725]",tokens/s,145.40081689512178,kWh,2.0775009377361663e-05,1.1323023498067578e-05,0.0,3.209803287542924e-05,tokens/kWh,7975566.6334296055,MB,29358.768128,,,,,s,10,118.04324531600105,11.804324531600106,0.08261022840095958,11.808232546500221,11.877726705600072,11.92205097479964,11.957510390159296,"[11.844335371999478, 11.854265794998355, 11.96637524399921, 11.828923614000814, 11.754196735999358, 11.65284812200116, 11.736662926001372, 11.750219160001507, 11.867876868000167, 11.787541478999628]",tokens/s,5.3370271065785575,kWh,0.00013957839608749964,7.607565571913965e-05,0.0,0.0002156540518066393,tokens/kWh,292134.55287400464,,s,630,118.039158890002,0.18736374426984445,0.003763427079279641,0.18688544199994794,0.19202031139957398,0.19396479465031008,0.2000851361601963,"[0.19178188499972748, 0.18689998499939975, 0.191624729999603, 0.18677885500073899, 0.18637572599982377, 0.18863128699922527, 0.19270476100064116, 0.18695323999963875, 0.1865806480000174, 0.18645300200114434, 0.18402437299846497, 0.18506512600106362, 0.1887104220004403, 0.18688150299931294, 0.18773363100081042, 0.18823305999831064, 0.18276148500081035, 0.1869435810003779, 0.18957231800050067, 0.18570121399898198, 0.18791963300100178, 0.18656690299940237, 0.18732884599921817, 0.18258238400085247, 0.18422659099996963, 0.18469222699968668, 0.18712904499989236, 0.186343253000814, 0.188709388999996, 0.18877780299953884, 0.19092775299941422, 0.18746987000122317, 0.18604822199995397, 0.19672175699997752, 0.188204265999957, 0.18846861400015769, 0.1842746049987909, 0.18762317500113568, 0.19076526399840077, 0.18697365100160823, 0.18647935199987842, 0.18657247999908577, 0.19398204000026453, 0.18645088099947316, 0.18503634800072177, 0.1869801680004457, 0.1869638619991747, 0.18796277299952635, 0.18803756300076202, 0.19008199400013837, 0.1914310039992415, 0.19052958900101657, 0.191630448000069, 0.18984631399871432, 0.19264048500008357, 0.18840088300021307, 0.19179045500095526, 0.18784236299870827, 0.1890660300014133, 0.18617595199975767, 0.18556063899995934, 0.1871440260001691, 0.19116625300011947, 0.19251974899998459, 0.18570074899980682, 0.19430037999882188, 0.18806600800053275, 0.1886140009992232, 0.19167467200168176, 0.19134867599859717, 0.1899928960010584, 0.19938202599951182, 0.19354805699913413, 0.19353262600088783, 0.18771992300025886, 0.19092241599901172, 0.18603434299984656, 0.18478194900126255, 0.18404409200047667, 0.19077371999992465, 0.18515880799895967, 0.18572969900014868, 0.18964646499989612, 0.19501600700095878, 0.18674412600012147, 0.184992312999384, 0.18787544899896602, 0.1913403600010497, 0.18994606299929728, 0.18740019500000926, 0.18448108499978844, 0.18938325000090117, 0.18856350100031705, 0.1880520830000023, 0.18684931099960522, 0.18273063900051056, 0.18755083499854663, 0.18230548500105215, 0.1865094089989725, 0.1914284869999392, 0.1848823110012745, 0.18843907799964654, 0.1927882990003127, 0.1916588629992475, 0.1880511720009963, 0.1876671569989412, 0.1863669859994843, 0.18586912300088443, 0.18673058199965453, 0.18356018200029212, 0.18457310600024357, 0.18369282300045597, 0.1916108039986284, 0.18813804000092205, 0.1882118929988792, 0.190923535001275, 0.18306074099928082, 0.18311951299983775, 0.18596622000040952, 0.188400917000763, 0.19186441700003343, 0.19065683699955116, 0.18561479199888709, 0.18370500500168419, 0.18740292799884628, 0.18624669800010452, 0.19301123399964126, 0.18561445900013496, 0.19041300599928945, 0.18650508999962767, 0.1856512680005835, 0.18711961999906634, 0.1876834400009102, 0.18396059299993794, 0.18399405399941315, 0.18459270400126115, 0.18627349799862714, 0.18235129300046538, 0.18327848199987784, 0.1814024329996755, 0.18475091600157612, 0.1849850159997004, 0.18458894099967438, 0.1875207720004255, 0.18709481499899994, 0.1899245010008599, 0.1966650379999919, 0.18752368300010858, 0.1856475059994409, 0.1881584919992747, 0.1915197510006692, 0.18989553100072953, 0.18539061099909304, 0.18687497599967173, 0.1879871010005445, 0.18789892200038594, 0.19201064899971243, 0.18725207500028773, 0.18829237899990403, 0.18709645499984617, 0.18838579099974595, 0.18931575799979328, 0.19647019600051863, 0.1925801199995476, 0.19532279200029734, 0.19173921500077995, 0.19258151199937856, 0.19827325999904133, 0.2004623660013749, 0.20157953399939288, 0.20037232200047583, 0.19252867300019716, 0.1944085939994693, 0.19439577999946778, 0.19155731700084289, 0.18747899399932066, 0.19311858800028858, 0.1900944970002456, 0.19101687699912873, 0.18979857799968158, 0.19155430300088483, 0.19651052200060803, 0.1914136300001701, 0.19175766499938618, 0.1949321729989606, 0.19087347500135365, 0.194361673999083, 0.18991754300077446, 0.19018258200048876, 0.19780535400059307, 0.19323061499926553, 0.18736166100097762, 0.1876471279992984, 0.1902961209998466, 0.1901625230002537, 0.1883992679995572, 0.18412425400128996, 0.18858447399907163, 0.1823438810006337, 0.191463801998907, 0.18519402800120588, 0.18734952199883992, 0.1850769089996902, 0.18618140700164076, 0.18809602599867503, 0.18740488100047514, 0.18541304099926492, 0.18933574600123393, 0.18666740299886442, 0.1844750620002742, 0.18583374500121863, 0.1899280279994855, 0.1885835129996849, 0.19231765100084885, 0.186573970999234, 0.1878648180008895, 0.18791857899850584, 0.18400802700125496, 0.18527362900022126, 0.18461752699840872, 0.18564769600016007, 0.19020239699966623, 0.18322732000160613, 0.1864922389995627, 0.1911881899995933, 0.18272932800027775, 0.18517941199934285, 0.18507329100066272, 0.18379265300063707, 0.1934005650000472, 0.1875430319996667, 0.19430117299998528, 0.18644608599970525, 0.18512010300037218, 0.18296757399912167, 0.18434811000042828, 0.18559960899983707, 0.18912179099970672, 0.18709196800045902, 0.18534484200063162, 0.19255133199840202, 0.1907488180004293, 0.1884181780005747, 0.1875913229996513, 0.18939215300088108, 0.1919655129986495, 0.18370618000153627, 0.18670610299886903, 0.19530344800114108, 0.19210727299832797, 0.1891916520016821, 0.1864875569990545, 0.1969135039998946, 0.18711707199872762, 0.18514407200018468, 0.18965478500103927, 0.18865077199916414, 0.19002111299960234, 0.1851264020006056, 0.19741520500065235, 0.1891702319990145, 0.18493726099950436, 0.1868669060004322, 0.17980584100041597, 0.18308379799964314, 0.18450503700114496, 0.18488059299852466, 0.18727711400060798, 0.18544404900057998, 0.1839975549992232, 0.1855776219999825, 0.18240487900038715, 0.1895763470001839, 0.18652835200009577, 0.18266309200043906, 0.1847722069996962, 0.1871705169996858, 0.19193951600027503, 0.19274891599889088, 0.1940329010012647, 0.18634386399935465, 0.18394083800012595, 0.1826718369993614, 0.18508661099986057, 0.18423542200071097, 0.1838822589998017, 0.18441876900033094, 0.1866893539990997, 0.18951805400138255, 0.1819772639992152, 0.19158437300029618, 0.188779453999814, 0.18768468599955668, 0.18421374799982004, 0.18595438900047156, 0.1846313400001236, 0.18678395499955514, 0.18687468900134263, 0.1834458649991575, 0.18320685400067305, 0.18750780499976827, 0.18513658499978192, 0.18174731299950508, 0.1842930980001256, 0.1857807579999644, 0.1869192859994655, 0.184300136001184, 0.18208155399952375, 0.18666667000070447, 0.19106947999898694, 0.1974139930007368, 0.1869783069996629, 0.18357778199970198, 0.18288463300086732, 0.18810505399960675, 0.19006075200013584, 0.18587937700067414, 0.18344320399955905, 0.18199691100016935, 0.18283202199927473, 0.18365375400026096, 0.18279589300072985, 0.187385720999373, 0.189438326000527, 0.18933914599983837, 0.18240947400045116, 0.18216202499934298, 0.18345731999943382, 0.18378817700067884, 0.18179820600016683, 0.1828278209995915, 0.18287790400063386, 0.18193184600022505, 0.18385219300034805, 0.18447898399972473, 0.1863125999989279, 0.18403785200098355, 0.18540839799970854, 0.18878071700055443, 0.18764772699978494, 0.18425324499912676, 0.1928358479999588, 0.1843802229996072, 0.18107973400037736, 0.18395689399949333, 0.19027526600075362, 0.1891980500004138, 0.19077427400043234, 0.18840408599862712, 0.1874986920010997, 0.1853735639997467, 0.19172540699946694, 0.18644910900002287, 0.18348850999973365, 0.1850337230007426, 0.19235738899988064, 0.18577240400009032, 0.183821758000704, 0.18686398199861287, 0.1833928760006529, 0.18169225900055608, 0.18979924699851836, 0.18295492000106606, 0.18585994299974118, 0.18330168800093816, 0.18039228899942827, 0.18130509599905054, 0.18346270199981518, 0.18303647900029318, 0.18052574099965568, 0.1831804720004584, 0.18162842700075998, 0.1847504029992706, 0.18442649499957042, 0.18202005600142002, 0.18305116899864515, 0.1835709830011183, 0.18196068399993237, 0.187996397000461, 0.18445413200061012, 0.18359850299930258, 0.18621390400039672, 0.1866460940000252, 0.18199868699957733, 0.18186046600021655, 0.18249573499997496, 0.1881029110008967, 0.1823284779984533, 0.18566616400130442, 0.18175464999876567, 0.18115494300036517, 0.18642186900069646, 0.1820606269993732, 0.18346077000023797, 0.18994163299976208, 0.1883441230002063, 0.18644122099976812, 0.18344220200015116, 0.18249972500052536, 0.18717579599979217, 0.1883499559990014, 0.1858753920005256, 0.1827780570001778, 0.18352681800024584, 0.18779403299959085, 0.18575910800063866, 0.1858820969991939, 0.18398229100057506, 0.1842471879990626, 0.18686541900024167, 0.1845035040005314, 0.1838927250009874, 0.1860903459983092, 0.1831709090001823, 0.18748749299993506, 0.18520084400006454, 0.18448136200095178, 0.18609874799949466, 0.1861860200006049, 0.1903099649989599, 0.1852251990003424, 0.18435351200059813, 0.1831681679996109, 0.1851751680005691, 0.1879262069996912, 0.1871015279994026, 0.18279212299967185, 0.1848961430005147, 0.1866900099994382, 0.1916306530001748, 0.18838765700093063, 0.2026128500001505, 0.1916135449991998, 0.18759343700003228, 0.18908056700092857, 0.18509926899969287, 0.1927035650005564, 0.1908736929999577, 0.1923748319986771, 0.18940621000001556, 0.19100273399999423, 0.2009102050014917, 0.18680690999826766, 0.19059554300110904, 0.19154507399980503, 0.1858067459997983, 0.18800921100046253, 0.1887733669991576, 0.18271522000031837, 0.18082692099960695, 0.18283014799999364, 0.18136018100085494, 0.18541305600047053, 0.18459010599872272, 0.190127157000461, 0.18638174099942262, 0.1864154459999554, 0.19089165200057323, 0.1877716100007092, 0.18840783199993894, 0.189639439999155, 0.19061604800117493, 0.18712672099900374, 0.18353863899938005, 0.18249644200113835, 0.18678413399902638, 0.18122098700041533, 0.18264936600098736, 0.18263767999997071, 0.18637931599914737, 0.1841186839992588, 0.18113052400076413, 0.18213678800020716, 0.18821064800067688, 0.18658668999887595, 0.1853700720002962, 0.18652723999912268, 0.1889616960015701, 0.188174260998494, 0.18404070600081468, 0.18866037299994787, 0.20281290400089347, 0.19016664699847752, 0.1879447090013855, 0.18184067199945275, 0.18429426299917395, 0.18420100600087608, 0.19032238999898254, 0.18421357800070837, 0.18416296699979284, 0.18330468800013477, 0.1851314760006062, 0.18412784300016938, 0.1841009039999335, 0.1869996559998981, 0.18397111299964308, 0.18843828399985796, 0.19013477900080034, 0.19314452299840923, 0.18659212000056868, 0.1856008529994142, 0.18537399400156573, 0.18298162799874262, 0.18275810300110606, 0.19379685700005211, 0.18819158600126684, 0.1851912120000634, 0.18408502700003737, 0.18277585999931034, 0.18570422300035716, 0.19214689599903068, 0.18728436499986856, 0.18692430700139084, 0.18742785299946263, 0.18814579600075376, 0.18873984799938626, 0.18794319100015855, 0.18382526599998528, 0.18720787299935182, 0.18659418999959598, 0.18274500200095645, 0.18768322100004298, 0.1873243289992388, 0.1890476860007766, 0.18659607500012498, 0.18346181699962472, 0.18602144199940085, 0.19155949900050473, 0.1886905930005014, 0.19367550799870514, 0.1906862710002315, 0.18804346100114344, 0.18755468000017572, 0.18421367199880478, 0.1848449780009105, 0.18278532099975564, 0.18376765999892086, 0.1854364610007906, 0.19221285599996918, 0.18778405400007614, 0.1883607750005467, 0.19425444300031813, 0.19115789699935704, 0.18893311899955734, 0.19080352200035122, 0.18585772599908523, 0.18532202599999437, 0.19143090400029905, 0.19031846899997618, 0.19484653000108665, 0.19257719999950496, 0.18893913799911388, 0.18664264300059585, 0.19108367000080761, 0.1984676819993183, 0.19262186199921416, 0.18658534000132931, 0.18914934699932928, 0.18261620500015852, 0.18781557900001644, 0.19172807600079977, 0.18625272399913229, 0.18647009400046954, 0.1876346419994661, 0.19183231500028342, 0.1946162860003824, 0.19103996000012557, 0.1933681959999376, 0.18817074000071443, 0.18676761599999736, 0.1869912429992837, 0.19158187399989401, 0.18777442499958852, 0.1893200850008725, 0.18730827900071745, 0.18473020099918358, 0.18654187900028774, 0.18319747500027006, 0.18495188999986567, 0.1839124649995938, 0.18440752199967392, 0.18645484100125032, 0.18434979099947668, 0.18567597400033264, 0.18395878999945126, 0.20252909699956945, 0.1886116990008304, 0.18412950899983116, 0.183423866999874, 0.18533793799906562, 0.1871947440013173, 0.18437967199861305, 0.18754398400051286, 0.1844256040003529, 0.19394371700036572, 0.18235447400002158, 0.18652554499931284, 0.18547575199954736, 0.1876326670007984, 0.18733329200040316, 0.18418798499988043, 0.18444716799967864, 0.18682006000017282, 0.18439626100007445, 0.1869537719994696, 0.1857993659996282, 0.18408860100134916, 0.18290657399847987, 0.18482342900097137, 0.18729083099970012, 0.18791317700015497, 0.19039741400047205, 0.18265625899948645, 0.18508776899943769, 0.18776128600075026, 0.18658335299915052, 0.18638874700081942, 0.18774282699996547, 0.18368308099888964, 0.1880490430012287, 0.19140858999890042, 0.18980554600057076, 0.18688938100058294, 0.19359212400013348, 0.1877189869992435, 0.19035294599962072, 0.1890135100002226, 0.18998163300057058, 0.19055255300008866, 0.18953942999905848]",tokens/s,5.337211870402115,, +float32-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cpu,,42,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: DeciLMForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cpu,,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: DeciLMForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +bfloat16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cpu,,42,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3832, in from_pretrained + model = cls(config, *model_args, **model_kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ + self.model = DeciCoderModel(config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ + self.self_attn = DeciCoderAttention(config=config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ + self._init_rope() + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1729, in __getattr__ + raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") +AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +bfloat16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cpu,,42,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3826, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1565, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1731, in _check_and_enable_sdpa + raise ValueError( +ValueError: DeciLMForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,14661.173248,,,,,s,1,222.14383632499994,222.14383632499994,0.0,222.14383632499994,222.14383632499994,222.14383632499994,222.14383632499994,[222.14383632499994],,kWh,0.0025942008033326275,0.0014139390462813667,0,0.004008139849613995,,MB,14734.311424,,,,,s,10,19.63256148900291,1.963256148900291,0.009498977431844254,1.9605485795009372,1.970647456199913,1.9792422546001944,1.9861180933204197,"[1.9602265240009729, 1.9550665850001678, 1.967731377999371, 1.9687375009998505, 1.9878370530004759, 1.958238846000313, 1.9604428180009563, 1.961757117000161, 1.9518693259997235, 1.960654341000918]",tokens/s,130.39561859688928,kWh,2.3112094451040927e-05,1.2596797496166137e-05,0.0,3.570889194720706e-05,tokens/kWh,7169082.714145175,MB,14735.650816,,,,,s,10,115.45062000399776,11.545062000399776,0.05011399321867743,11.541439079000156,11.602749168499576,11.62398151624966,11.640967394449726,"[11.598030868999558, 11.531252349999704, 11.539254089000679, 11.564043585000036, 11.543624068999634, 11.477908058999674, 11.56681689799916, 11.468493997999758, 11.645213863999743, 11.51598222299981]",tokens/s,5.456878447064075,kWh,0.00013596995709944625,7.410893842811202e-05,0.0,0.00021007889552755827,tokens/kWh,299887.33443115244,,s,630,115.44592550599918,0.1832475008031733,0.0025214390008296648,0.18285939950055763,0.18647546009924554,0.18739881520059498,0.19185741017094188,"[0.1838822509998863, 0.18213555199872644, 0.1815652780005621, 0.1834873019997758, 0.1823343569994904, 0.18058121700050833, 0.18308668899953773, 0.18036198600020725, 0.18552665599963802, 0.1846950770013791, 0.1819138099999691, 0.18378460299936705, 0.18461014300009992, 0.1846305620001658, 0.18205814599969017, 0.18260979799924826, 0.18133993000083137, 0.18487683400053356, 0.18200927100042463, 0.18297377099952428, 0.18449635700017097, 0.18473344299854944, 0.18255091100036225, 0.18536223600131052, 0.18264845799967588, 0.18302046200005861, 0.18097242799922242, 0.1812167469997803, 0.1822521770009189, 0.18071665699972073, 0.18068664299971715, 0.1840508690002025, 0.1832697019999614, 0.18604760299967893, 0.18426255800113722, 0.18711395999889646, 0.184384535999925, 0.182534237001164, 0.1826730739994673, 0.18497998999919218, 0.18306448300063494, 0.1879116050004086, 0.18600703199990676, 0.18211683099980291, 0.1827495759989688, 0.18765605699991283, 0.1869535720015847, 0.1851067419993342, 0.1825420450004458, 0.18582557499939867, 0.18365953899956367, 0.18440513300083694, 0.18682259199886175, 0.18425696400117886, 0.18488195700047072, 0.18616408899833914, 0.18662220200167212, 0.18506739899930835, 0.1816311439997662, 0.1852836069992918, 0.1852276690005965, 0.19120643600035692, 0.19598217400016438, 0.1860123360002035, 0.18298671799857402, 0.18280224000045564, 0.1835303389998444, 0.1821116669998446, 0.18206354600079067, 0.18051639199984493, 0.18402206400060095, 0.1829887609983416, 0.18420858800163842, 0.18240471099852584, 0.18591832800120756, 0.1827499649989477, 0.18626179400052933, 0.18337422699914896, 0.18598720699992555, 0.18518445500012604, 0.18183509800110187, 0.18315785099912318, 0.18468020000000251, 0.18617597100092098, 0.18499461799910932, 0.1853203360005864, 0.18417877799947746, 0.1821443589997216, 0.18145889500010526, 0.18285881100018742, 0.1829015800012712, 0.18565626899908239, 0.18473657499998808, 0.1851523460009048, 0.18043078999835416, 0.1824406280011317, 0.18102812200049812, 0.17963315199995122, 0.17910515499897883, 0.18224172100053693, 0.1834837160004099, 0.18214203299976361, 0.1815537359998416, 0.18200231399896438, 0.18193523100126185, 0.18697428599989507, 0.18401591299880238, 0.18270012700122606, 0.18215219199919375, 0.18276739300017653, 0.18034400999931677, 0.1821948100005102, 0.18223271100032434, 0.18361926100078563, 0.18053553999925498, 0.18270572300025378, 0.18401272600021912, 0.18109748499955458, 0.18326038099985453, 0.18309100899932673, 0.18483579600069788, 0.18538274899947282, 0.18405132200132357, 0.18099052399884386, 0.18004803000076208, 0.17943824000030872, 0.19545081500109518, 0.18364358699909644, 0.1864307510004437, 0.18405891600014002, 0.18095973200070148, 0.18275489799998468, 0.18484584099860513, 0.18301412300024822, 0.18183977700027754, 0.18223263999971095, 0.1848772020002798, 0.1823281169999973, 0.18366944099943794, 0.18193874100143148, 0.183851974999925, 0.18415624800036312, 0.18341863899877353, 0.18189263800013578, 0.18306029700033832, 0.1825886820006417, 0.1807435729988356, 0.17906387699986226, 0.18016883299969777, 0.18176828500145348, 0.18227005100015958, 0.1839026429988735, 0.18040567100069893, 0.18195715799993195, 0.18170723000002909, 0.18445197799883317, 0.18150961700121115, 0.18538486199940962, 0.18203582600108348, 0.18270755899902724, 0.1822100749996025, 0.1823994260012114, 0.18162633099927916, 0.18412317500042263, 0.18147120799949334, 0.18085420500028704, 0.18158300100003544, 0.18312871399939468, 0.18154322800000955, 0.18384119700021984, 0.185296866000499, 0.1849395609988278, 0.18412697300118452, 0.1834680270003446, 0.18168708299890568, 0.18440553800064663, 0.18264552199980244, 0.18044836599983682, 0.1816725860007864, 0.1852548509996268, 0.1841360779999377, 0.1872923869996157, 0.18617127899960906, 0.19191791200137231, 0.18074474599961832, 0.18049876899931405, 0.18193434500062722, 0.18249254599868436, 0.18178350900052465, 0.18601582100018277, 0.18054394899991166, 0.18090536199997587, 0.1811831329996494, 0.1872501910002029, 0.18390671199995268, 0.18400139400000626, 0.18680667400076345, 0.18265331099973992, 0.18524133000028087, 0.18549065399929532, 0.1868064919999597, 0.18640045400024974, 0.18517158800023026, 0.18182766400059336, 0.18371080099859682, 0.18398489600076573, 0.1826387839992094, 0.18790669700138096, 0.18398959299884154, 0.18246998799986613, 0.1832847120003862, 0.18239315100072417, 0.18554223099999945, 0.18606840599932184, 0.18540952399962407, 0.1846758760002558, 0.1821427760005463, 0.18356023600063054, 0.18580339899926912, 0.18147696400046698, 0.18087752099927457, 0.18134683000062068, 0.18228295900007652, 0.1841168499995547, 0.18318632699993032, 0.18020962999980839, 0.18660266900042188, 0.17980777899902023, 0.18133986000066216, 0.18302425000001676, 0.18330888200034678, 0.18204831900038698, 0.18190794699876278, 0.18274078300055407, 0.1817581949999294, 0.18668150799931027, 0.18047018100151035, 0.18132281199905265, 0.1808949250007572, 0.1837715429992386, 0.18315134099975694, 0.18567070199969749, 0.18265713100117864, 0.18362261299989768, 0.18190175099880435, 0.18235392800124828, 0.18353344599927368, 0.18260314100007236, 0.1846885739996651, 0.18748022600084369, 0.18365876300049422, 0.18532484999923327, 0.18200485299894353, 0.18276955200053635, 0.18355140300081985, 0.18457701199986332, 0.18165651399976923, 0.1815401429994381, 0.17928000999927463, 0.18363144000068132, 0.18369492199963133, 0.18383010999968974, 0.18662181700165092, 0.18169065399888495, 0.18304112300029374, 0.18145906800054945, 0.17987571499907062, 0.18119155000022147, 0.1843088280002121, 0.18455802499920537, 0.18245430200113333, 0.18123001099957037, 0.18108768399906694, 0.17939333400136093, 0.18062433100021735, 0.1800713699994958, 0.18239456199989945, 0.17991072699987853, 0.179785794000054, 0.18560854099996504, 0.18077140000059444, 0.18119107900020026, 0.18094654999913473, 0.1823027690006711, 0.18224477199873945, 0.1871084280010109, 0.18429965599898424, 0.18323802099985187, 0.18423224700018181, 0.18135955700017803, 0.18233845299982931, 0.18102837500009628, 0.18711694400008128, 0.18877172400061681, 0.1834619589990325, 0.18232647000149882, 0.18428790100006154, 0.18742097099857347, 0.18373730100029206, 0.1838799619999918, 0.18646380300015153, 0.18470001199966646, 0.18442203100130428, 0.18348824999884528, 0.18228512000132469, 0.18097696199947677, 0.18495906299904163, 0.18410900399976526, 0.18300972200086107, 0.18738271999973222, 0.18696236399955524, 0.18591709000065748, 0.1840824259998044, 0.18756075000055716, 0.1829307099997095, 0.1843468759998359, 0.18269379000048502, 0.17999286699887307, 0.18104153900094389, 0.18058197800019116, 0.18073480199927872, 0.17920052599947667, 0.18249680400003854, 0.18443944200043916, 0.18200497699945117, 0.17943454300075246, 0.1804290719992423, 0.17896413800008304, 0.18026421499962453, 0.18090141900029266, 0.1811042669996823, 0.18076118499993754, 0.18202288000065892, 0.18357008299972222, 0.1836352110003645, 0.18133902500085242, 0.18387936299950525, 0.18478567599959206, 0.18200370500017016, 0.18285851899963745, 0.1848504250010592, 0.18143727399910858, 0.179959274999419, 0.18241080100051477, 0.17928265900081897, 0.18012577899935422, 0.17927463100022578, 0.18347359899962612, 0.1803801959995326, 0.18216824699993595, 0.18285086900141323, 0.18183057499845745, 0.18236217600133386, 0.1807092899998679, 0.17998679300035292, 0.18096080199939024, 0.18202196799938974, 0.18287844600126846, 0.18196263299978455, 0.18025752199901035, 0.18310106300123152, 0.1798271459992975, 0.18061655500059715, 0.18232990199976484, 0.1855181779992563, 0.18335544100045809, 0.18113401499977044, 0.18190012199920602, 0.184542267001234, 0.18412611299936543, 0.18255745900023612, 0.184934651999356, 0.18547915400085913, 0.18411057800039998, 0.1883504429988534, 0.18655228300121962, 0.1808464189998631, 0.18528505399990536, 0.18562780499996734, 0.18414183599998069, 0.1869968049995805, 0.1858574910002062, 0.18449634699936723, 0.18160512600115908, 0.1820041350001702, 0.18114132799928484, 0.1827131099998951, 0.17945435000001453, 0.18234904600103619, 0.17953450200002408, 0.18003736099853995, 0.18381101500017394, 0.1838216600008309, 0.18545439099943906, 0.19170928499988804, 0.18562516100064386, 0.1809084359992994, 0.1827575320003234, 0.18425738100086164, 0.18943632699847512, 0.18675251100103196, 0.18654492099994968, 0.18433058200025698, 0.18202010200002405, 0.1836436909998156, 0.18345302499983518, 0.1817302630006452, 0.18368490299872064, 0.18272778800019296, 0.17902322400004778, 0.18339991799985, 0.18109067499972298, 0.18195537200153922, 0.18142347599859932, 0.17876701800014416, 0.18225375300062296, 0.1829527819991199, 0.18214151400025003, 0.18049674600115395, 0.1812268759986182, 0.18316795999999158, 0.1854248470008315, 0.18358427100065455, 0.1830593639988365, 0.18299737100096536, 0.18482900699927995, 0.18417681800019636, 0.18285982300039905, 0.18275162499958242, 0.1847723529990617, 0.184347334999984, 0.18200939600137644, 0.18090812399896095, 0.19042064200039022, 0.19017268999959924, 0.18741198400130088, 0.186809495999114, 0.18289887399987492, 0.1828589760007162, 0.1835962400000426, 0.18393917100002, 0.18166840499907266, 0.1798280720013281, 0.1805094499995903, 0.17974957399928826, 0.17931552700065367, 0.17885273700085236, 0.1792824889998883, 0.1784113889989385, 0.17992503400091664, 0.18286328799877083, 0.18142108600113716, 0.1793002449994674, 0.1826696330008417, 0.1825222229999781, 0.1821052969989978, 0.18249750399991171, 0.18099923400041007, 0.18300604700016265, 0.18454514000040945, 0.18672724999851198, 0.18222770200009109, 0.18114135600080772, 0.18297000600068714, 0.17962028999863833, 0.18199583600107871, 0.17938500799937174, 0.18337800100016466, 0.18253738100065675, 0.18191627399937715, 0.18682242300019425, 0.17950164399917412, 0.18185822700070275, 0.18188576499960618, 0.18190448700079287, 0.18007447899981344, 0.18351142599931336, 0.18517217400039954, 0.1819164959997579, 0.18558740800108353, 0.18722419400000945, 0.18466607099981047, 0.18158829499952844, 0.18314356100017903, 0.18220633099917904, 0.18030084900055954, 0.18051776100037387, 0.18309971199960273, 0.1811018829994282, 0.18168737500127463, 0.18109679400004097, 0.1795413699983328, 0.1803239910004777, 0.1803668739994464, 0.18388216300081694, 0.18084861500028637, 0.18094440099957865, 0.18255846200008818, 0.18351387199982128, 0.18078697100099816, 0.183685288999186, 0.18383549899954232, 0.18391828700077895, 0.1876035099994624, 0.18367313300041133, 0.1826176279992069, 0.18194254700028978, 0.18637841999952798, 0.18244281000079354, 0.18339948699940578, 0.18154356200102484, 0.1808479149985942, 0.18095142800120811, 0.18061448200023733, 0.18040554499930295, 0.1817822419998265, 0.1812276969994855, 0.180693993001114, 0.18243860799884715, 0.17978862000018125, 0.18374958700042043, 0.18217718400046579, 0.18187091199979477, 0.186264250000022, 0.18858714899943152, 0.18286462300056883, 0.1870032059996447, 0.18728587700024946, 0.18288797499917564, 0.1835470060013904, 0.18283991399948718, 0.18401320500015572, 0.18287756400059152, 0.18397810399983427, 0.1866609039989271, 0.18788631000097666, 0.186560696000015, 0.1928484429990931, 0.18806244900042657, 0.19200483100030397, 0.1877977969998028, 0.18647501799932797, 0.18724652800119657, 0.19511262099877058, 0.18929049100006523, 0.18712073800088547, 0.18627640299928316, 0.19034869400093157, 0.18902819899994938, 0.1864794389985036, 0.1853620490001049, 0.188346476001243, 0.18327685199983534, 0.18558836199918005, 0.18242429699967033, 0.18596969800091756, 0.1841645709992008, 0.18248189000041748, 0.1834844120003254, 0.18619885699990846, 0.18408221199933905, 0.1855655220006156, 0.18050772800052073, 0.1851455599990004, 0.18262656899969443, 0.18456066900034784, 0.1851274460004788, 0.18487850099882053, 0.18532231000062893, 0.18099285000062082, 0.18228142199950526, 0.185595323000598, 0.18333861499922932, 0.18405815000005532, 0.18224383800043142, 0.18498230499972124, 0.18346549499983666, 0.18526790299983986, 0.1815218219999224, 0.18166405100055272, 0.18053069099914865, 0.1795528220009146, 0.18504276900057448, 0.1819400409985974, 0.18079099799979303, 0.17991719000019657, 0.17894978000003903, 0.18214215300031356, 0.1796281810002256, 0.18199075799930142, 0.18374201799997536, 0.18400091200055613, 0.18233151899949007, 0.18566850300157967, 0.1835727699999552, 0.18541974499930802, 0.18298315100037144, 0.17957234299865377, 0.18108590000156255, 0.18359883499942953, 0.18146096800046507, 0.1812436609998258, 0.18252264299917442, 0.18229318399971817, 0.18115770800068276, 0.18500006800059055, 0.18032129099992744, 0.18364533099884284, 0.1791218470007152, 0.1800108839997847, 0.18071778699959395, 0.18060336300004565, 0.1811295160005102, 0.18274841200036462, 0.18131448699932662, 0.18305592199976672, 0.18301817199971993, 0.18296530400039046, 0.18162709300122515, 0.18091134099995543, 0.1827695189986116, 0.18563674400138552, 0.18386825699963083, 0.18174821499997051, 0.18154076800055918, 0.18282106299920997, 0.18241257299996505, 0.189688460999605, 0.19353127900103573, 0.18855955499930133]",tokens/s,5.457100345800093,, +bfloat16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,deci,Deci/DeciLM-7B,Deci/DeciLM-7B,cpu,,42,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,14661.533696,,,,,s,1,183.44631407899942,183.44631407899942,0.0,183.44631407899942,183.44631407899942,183.44631407899942,183.44631407899942,[183.44631407899942],,kWh,0.0021375349317368057,0.0011650383873756396,0,0.0033025733191124453,,MB,14726.987776,,,,,s,10,23.701266373002,2.3701266373002,0.005067095713392068,2.3697625570002856,2.377714048099733,2.378238028550186,2.378657212910548,"[2.371560010999019, 2.3653352399996947, 2.369584884001597, 2.3728165259999514, 2.3775976079996326, 2.3787620090006385, 2.3626628300007724, 2.3689357319999544, 2.3699402299989742, 2.364071303001765]",tokens/s,108.0111062300065,kWh,2.797509288465348e-05,1.5247342793945105e-05,0.0,4.3222435678598586e-05,tokens/kWh,5922849.926913244,MB,14729.46176,,,,,s,10,89.42062511400036,8.942062511400035,0.021255108695778513,8.948121590500705,8.963389674499377,8.968437817249377,8.972476331449379,"[8.924389998001061, 8.945242551000774, 8.89487017600004, 8.952014792999762, 8.962267864999376, 8.951285637000183, 8.924503827998706, 8.951000630000635, 8.941563676000442, 8.97348595999938]",tokens/s,7.045354460414777,kWh,0.00010563345049187344,5.757432821676149e-05,0.0,0.0001632077787086349,tokens/kWh,386011.0130686242,,s,630,89.41657629800102,0.1419310734888905,0.0019725382971478005,0.14157313800023985,0.1446112905994596,0.14576741074988603,0.14721633804048906,"[0.1455389370003104, 0.14384510300078546, 0.14164809499925468, 0.1408305510012724, 0.13918188799834752, 0.1421006740001758, 0.1417317080013163, 0.14076728799955163, 0.14119919400036451, 0.14226254200002586, 0.1453529399987019, 0.1396725160011556, 0.1409550999997009, 0.14505733199985116, 0.1429090769997856, 0.1433321069998783, 0.1417323599998781, 0.14130622900120215, 0.14428146799946262, 0.14100140200025635, 0.14000661600039166, 0.14069528199979686, 0.14303866199952608, 0.13973471999997855, 0.1405607249998866, 0.13920949100065627, 0.1394636939985503, 0.14109811000162154, 0.13858000299842388, 0.13852682800097682, 0.1404360340002313, 0.14728750500034948, 0.1431398459990305, 0.14047008700072183, 0.1401858839999477, 0.14253263499995228, 0.14414210399991134, 0.1404264519987919, 0.1406215550014167, 0.13971623299948988, 0.1400124190004135, 0.14045427000019117, 0.14077303199883318, 0.13986726000075578, 0.1408722090000083, 0.140540451000561, 0.14088858899958723, 0.14282796199950099, 0.14154865199998312, 0.14687963300093543, 0.1419475489983597, 0.14227730800121208, 0.14143014699948253, 0.14103022000017518, 0.14445697299925087, 0.14304820000143081, 0.14088944399918546, 0.14142685200022242, 0.14122345500072697, 0.14259272599883843, 0.1414745160000166, 0.14120374100093613, 0.14174173199899087, 0.14699303100132965, 0.14320340899939765, 0.14541968699995778, 0.14070059500045318, 0.14229373699890857, 0.13956868599962036, 0.14112498500071524, 0.14072828800090065, 0.1394735159992706, 0.14440609399935056, 0.1407422360007331, 0.14010362999943027, 0.13920048900035908, 0.14008806200035906, 0.1429337179997674, 0.14005789300063043, 0.13969553999959317, 0.14034024399916234, 0.14072998899973754, 0.14107901600073092, 0.14049796200015408, 0.1401366579993919, 0.1462044279996917, 0.14654264400087413, 0.14703215699955763, 0.14521031600088463, 0.14273303200025111, 0.14318717599962838, 0.14346123799987254, 0.14135175899900787, 0.14189597699987644, 0.1463671820001764, 0.1442561210005806, 0.13987089700094657, 0.14157892599905608, 0.14139001100011228, 0.14084739399913815, 0.14092487700145284, 0.14091755899971758, 0.1438625509999838, 0.1405868969995936, 0.14206176699917705, 0.14179877400056284, 0.14330663600048865, 0.13999154399971303, 0.1437127370008966, 0.14003313799912576, 0.14363909299936495, 0.14164915100081998, 0.14157224399968982, 0.14132284600054845, 0.13996828099880076, 0.14035770700138528, 0.14027610499942966, 0.1445717810001952, 0.14156205600011162, 0.14219153100020776, 0.14343595999889658, 0.14334142500047165, 0.1407349080000131, 0.13990763000037987, 0.1398249419999047, 0.14179315299952577, 0.1450172509994445, 0.1405832090003969, 0.13878860200020426, 0.14046939999934693, 0.14101420599945413, 0.14061347599999863, 0.14045746100055112, 0.13970285299910756, 0.13979569200091646, 0.13961038399975223, 0.1405919349999749, 0.14268761099992844, 0.1430594920002477, 0.1460045829990122, 0.1451065440014645, 0.1408981059994403, 0.14048742200066044, 0.1415547090000473, 0.13910795499941742, 0.13972778300012578, 0.13962118099880172, 0.13957105400004366, 0.1411985970007663, 0.14205421100086824, 0.1403745239986165, 0.1423043909999251, 0.14241610000135552, 0.14195575099984126, 0.1429421339998953, 0.14047506899987638, 0.13968345600005705, 0.14104098200004955, 0.13926081200042972, 0.14187853999828803, 0.14047814400146308, 0.14040110699897923, 0.142460201999711, 0.14390842000102566, 0.14482036600020365, 0.14138584699867351, 0.14044514400120534, 0.14125377500022296, 0.14067180999882112, 0.1401666270012356, 0.14139961399996537, 0.14305269499891438, 0.14131107200046245, 0.14188313900012872, 0.1437982749994262, 0.1405024460000277, 0.13942494799994165, 0.1404779510012304, 0.14086553899869614, 0.140252920000421, 0.1402145539996127, 0.140032303001135, 0.1405291920000309, 0.1410351959984837, 0.14139829500163614, 0.14039874599984614, 0.14004284600014216, 0.14193343699844263, 0.1398735820002912, 0.14457031500023731, 0.14459716400051548, 0.14034056600030453, 0.14197881299878645, 0.1470520069997292, 0.14333324700055528, 0.1430628660000366, 0.1397421490000852, 0.1404678620001505, 0.14151536999997916, 0.14044221400035894, 0.14049354099915945, 0.14031288800106267, 0.13916171099845087, 0.1396123530012119, 0.13962116000038804, 0.14254453600005945, 0.14287530399997195, 0.1394684930000949, 0.13988314899870602, 0.13978788600070402, 0.14199371299946506, 0.14493731299990031, 0.14218381700084137, 0.14260212699991826, 0.1402577859989833, 0.1406793200003449, 0.1444082440011698, 0.14208195799983514, 0.1411832300000242, 0.14193602900013502, 0.1399804699995002, 0.14035935399988375, 0.14132777899976645, 0.1397192090007593, 0.14064952799890307, 0.14097290100107784, 0.1430269229986152, 0.14339187299992773, 0.1422976040012145, 0.1418365419995098, 0.1440080569991551, 0.14115729900004226, 0.14169061700158636, 0.14013508499920135, 0.14383439800076303, 0.1434122379996552, 0.14165804500044032, 0.14616526999998314, 0.14639054099825444, 0.14147958900139201, 0.14491378200000327, 0.14295223099907162, 0.1423357229996327, 0.14213338700028544, 0.14051518999986, 0.14134921700133418, 0.14384163599970634, 0.14075992299876816, 0.14583656400100153, 0.14604383500045515, 0.14245962599852646, 0.14184761900105514, 0.14883169399945473, 0.14171206200080633, 0.1442834360004781, 0.1427723340002558, 0.14290886899834732, 0.14422829500108492, 0.14330707299995993, 0.1457712020001054, 0.14485319599953073, 0.1435305970007903, 0.1416144439990603, 0.14338729399969452, 0.1430911110001034, 0.1431314020010177, 0.14512903099966934, 0.14415283599919348, 0.14235277100124222, 0.13953935700010334, 0.1435773409993999, 0.1409340999998676, 0.14255599900025118, 0.14029083400055242, 0.1414751789998263, 0.13998160699884465, 0.1404058250009257, 0.1401624609989085, 0.13980115300000762, 0.13934080699982587, 0.13992234800025471, 0.14059266399999615, 0.14360734300134936, 0.1449128899985226, 0.1427584620014386, 0.14398331099982897, 0.1413815039995825, 0.144529302000592, 0.13975479399960022, 0.14147574799972062, 0.14115471500008425, 0.14336053599981824, 0.1408184230003826, 0.1414106649990572, 0.1428592980009853, 0.14039728199895762, 0.1412972850012011, 0.1464829529995768, 0.13962362600068445, 0.14509334399917861, 0.1452050150001014, 0.1428094519997103, 0.14208083099947544, 0.14274979699985124, 0.14302355600011651, 0.14359015600166458, 0.13987330099917017, 0.14006429099936213, 0.14192819799973222, 0.140443704000063, 0.14080432700029633, 0.14018081900030666, 0.14093913699980476, 0.14027175800038094, 0.13937568400069722, 0.14642909000031068, 0.14085884999985865, 0.13938057700033823, 0.13847650499883457, 0.141311035000399, 0.14028233699900738, 0.13978924300135986, 0.1407187189997785, 0.13996744999894872, 0.1440992180014291, 0.14100821200008795, 0.14104401399890776, 0.14167685100073868, 0.14175061300011293, 0.14299217599909753, 0.14411558200117724, 0.1436284439987503, 0.14164933800020663, 0.1484245869996812, 0.14251589700143086, 0.14312733699989622, 0.142621828999836, 0.14113867199921515, 0.1437705990010727, 0.14303992799978005, 0.14059392199851573, 0.1410484150001139, 0.14005323300079908, 0.1412679829991248, 0.1384504810012004, 0.13945368999884522, 0.1393232860009448, 0.14066037100019457, 0.14117074599926127, 0.13923513900044782, 0.14177684499918541, 0.14362590000018827, 0.14063915299993823, 0.14709214200047427, 0.14179845500075317, 0.14226941099877877, 0.13992833300108032, 0.14397332499902404, 0.14297339100085082, 0.14074310999967565, 0.141444902999865, 0.14208050500019453, 0.14310359800037986, 0.14336116300000867, 0.1409184259991889, 0.14086609600053634, 0.1407025110001996, 0.14085046800028067, 0.14555136399940238, 0.14319063000039023, 0.14189446099953784, 0.14689945700047247, 0.14552799999910349, 0.1430203050003911, 0.1424570479994145, 0.14302273300017987, 0.14216614399992977, 0.14390795800136402, 0.14486640299946885, 0.1391201779988478, 0.14057304800007842, 0.1410706440001377, 0.14036395299990545, 0.13930476700079453, 0.1386097199992946, 0.139872223000566, 0.14210706800076878, 0.143430805999742, 0.14274852299968188, 0.14309909000076004, 0.13960852699892712, 0.13903893099995912, 0.14147517700075696, 0.1401295569994545, 0.14203860000088753, 0.14202460999877076, 0.1409918879999168, 0.14115924700126925, 0.14059657599864295, 0.13912164300018048, 0.14597368900103902, 0.14195571299933363, 0.1420979750000697, 0.1417501390005782, 0.14257505099885748, 0.14100335399962205, 0.1425567790010973, 0.1452185329999338, 0.14511083499928645, 0.1415450790009345, 0.14085990799867432, 0.14120508200176118, 0.13993616599873349, 0.14011435199972766, 0.140295163000701, 0.14225072999943222, 0.14180689900058496, 0.14450609999948938, 0.14170141100112232, 0.1399903860001359, 0.13995000999966578, 0.14136969100036367, 0.13899696099906578, 0.14204810299997916, 0.14246360400102276, 0.14622450899878459, 0.14392608500020287, 0.141095523000331, 0.14012587499928486, 0.14179277600123896, 0.14412944899959257, 0.14245203499922354, 0.14330136600074184, 0.14146963999883155, 0.14446613500149397, 0.14215963799870224, 0.14137836800000514, 0.1399588929998572, 0.14171827700010908, 0.1407634790011798, 0.14051726700017753, 0.14405239700136008, 0.14447294699857594, 0.14051850700161594, 0.14389838999886706, 0.1439368840001407, 0.14398842400078138, 0.14261564600019483, 0.14107530499859422, 0.1406387300012284, 0.14069987299990316, 0.1446241209996515, 0.14331380799922044, 0.14107183800115308, 0.1416116599993984, 0.1420075040005031, 0.14011957299953792, 0.14045703100055107, 0.1414228119992913, 0.1439080100008141, 0.14367555999888282, 0.14347884700146096, 0.1414042979995429, 0.1398733109999739, 0.14050545600002806, 0.14076435799870524, 0.14031718100159196, 0.1431506190001528, 0.14312349199826713, 0.14033886000106577, 0.14237658300044131, 0.14347970099879603, 0.1401431659996888, 0.14243732500108308, 0.13943062999896938, 0.13960279999992053, 0.1435087760000897, 0.14149425300092844, 0.1454479589992843, 0.1450734440004453, 0.14106495600026392, 0.1411335680004413, 0.14614325399998052, 0.14337431300009484, 0.14176536299964937, 0.1400320169996121, 0.13962156100024004, 0.1395443780002097, 0.13844836299904273, 0.14015459700021893, 0.14560895900103787, 0.1434097629989992, 0.14259628000036173, 0.14114803399934317, 0.14271666900094715, 0.14027363599961973, 0.142333072000838, 0.14230884799872, 0.14101953600038541, 0.14111956600027042, 0.1397705479994329, 0.14194625799973437, 0.14686120800070057, 0.14413965399944573, 0.14363419800065458, 0.14229728799909935, 0.14043395500084443, 0.14125320099992678, 0.14169923000008566, 0.1414528759996756, 0.14406162799969024, 0.14411228000062692, 0.14395078400048078, 0.14579574199888157, 0.14824962900092942, 0.14516877099958947, 0.14256718499927956, 0.14454284900057246, 0.14456678799979272, 0.14179960200090136, 0.14197337400037213, 0.1426206819996878, 0.14320993999899656, 0.1426941830013675, 0.14192883799842093, 0.13978892800150788, 0.13962599799924647, 0.13849785899947165, 0.13961114999983693, 0.14014972600125475, 0.14348934499867028, 0.1389543710010912, 0.14249372099948232, 0.1420538000002125, 0.1388133590007783, 0.13806819699857442, 0.14328086700152198, 0.14293191199976718, 0.14008919600018999, 0.1418533549986023, 0.14247623899973405, 0.14184796000154165, 0.14423639499909768, 0.14250166400051967, 0.1427109649994236, 0.14407601799939584, 0.14059143000122276, 0.1402632259996608, 0.13969680899936066, 0.1393729629999143, 0.13957833600034064, 0.1421743430000788, 0.14019741300035093, 0.1419947329995921, 0.14376563800033182, 0.14175749199966958, 0.14295506999951613, 0.14216542200119875, 0.1426388660001976, 0.13965137800005323, 0.1407193710001593, 0.14030731200000446, 0.14270975399995223, 0.14117300399993837, 0.14025942699845473, 0.1428354780000518, 0.1408120530013548, 0.14673888399920543, 0.14877944299951196, 0.14532098600102472, 0.14246509599979618, 0.1424748629997339, 0.14773251000042364, 0.14503817999866442, 0.14427072900070925, 0.14345824799966067, 0.14579858700017212, 0.1428689479998866, 0.142519324001114, 0.14117842799896607, 0.14133725499959837, 0.14323840300130541, 0.14356416299960983, 0.14215553699978045, 0.14129594800033374, 0.14702879399919766, 0.1456787750012154, 0.1446098649994383, 0.14163645999906294, 0.14157403200078988, 0.1444172970004729, 0.14258177599913324, 0.1451871610006492, 0.14278787399962312, 0.1472670660004951, 0.14106338799865625, 0.14129284000046027, 0.14471893600057228, 0.14273914099976537, 0.14273012100056803, 0.14179976899868052, 0.13954946600097173, 0.1389671710003313, 0.1426914490002673, 0.1392055899996194, 0.14297712999905343, 0.13959695400080818, 0.14050446299916075, 0.14238670700069633, 0.14383357699989574, 0.14576277699961793, 0.1436414510008035, 0.13982055300039065, 0.139863685999444, 0.14156714099954115, 0.14050567599952046, 0.13951711300069292, 0.14299081500030297, 0.14003698699889355, 0.14068465800119156, 0.14006856400010292, 0.13977111500025785, 0.13916388299912796, 0.14169135500014818, 0.1390741520008305, 0.14094929499879072, 0.139379022999492, 0.1401162350011873, 0.14110991800043848, 0.14031009399877803]",tokens/s,7.045673476698349,, +float32-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cpu,,42,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run + self.run_model_loading_tracking(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3832, in from_pretrained + model = cls(config, *model_args, **model_kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ + self.model = DeciCoderModel(config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ + self.self_attn = DeciCoderAttention(config=config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ + self._init_rope() + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1729, in __getattr__ + raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") +AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +bfloat16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cpu,,42,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) - File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking - context_stack.enter_context(energy_tracker.track()) - File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context - result = _cm_type.__enter__(cm) - File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ - return next(self.gen) - File ""/workspace/optimum_benchmark/trackers/energy.py"", line 174, in track - self.emission_tracker.start_task() - File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task - if self._scheduler: -AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3832, in from_pretrained + model = cls(config, *model_args, **model_kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ + self.model = DeciCoderModel(config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ + self.self_attn = DeciCoderAttention(config=config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ + self._init_rope() + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1729, in __getattr__ + raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") +AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-eager-pytorch,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,26443.382784,,,,,s,1,416.0725446910001,416.0725446910001,0.0,416.0725446910001,416.0725446910001,416.0725446910001,416.0725446910001,[416.0725446910001],,kWh,0.004881660385143748,0.002660693515757511,0,0.007542353900901259,,MB,27043.291136,,,,,s,10,35.84458376100338,3.5844583761003377,0.007329127151613124,3.581920324500061,3.5933664820999183,3.59342874005024,3.593478546410497,"[3.5934909980005614, 3.5806673650004086, 3.578081989000566, 3.570503140001165, 3.5930893710010423, 3.582125721999546, 3.5900059239993425, 3.593352646999847, 3.5817149270005757, 3.5815516780003236]",tokens/s,71.41943723127054,kWh,4.222885850826313e-05,2.3016187507371003e-05,0.0,6.524504601563414e-05,tokens/kWh,3923669.544790524,MB,27315.351552,,,,,s,10,191.71148664899556,19.171148664899555,0.045699471895773496,19.166182255499734,19.235882381699593,19.243812118349616,19.250155907669633,"[19.156944663000104, 19.202089746999263, 19.111819742998705, 19.251741854999636, 19.23412021799959, 19.175419847999365, 19.133905310000046, 19.11193912899944, 19.18669035899984, 19.146815776999574]",tokens/s,3.286188068393975,kWh,0.0002262354940516688,0.00012330715505750746,0.0,0.00034954264910917624,tokens/kWh,180235.51678331124,,s,630,191.70293175499864,0.30429036786507724,0.0018862060580412715,0.30394128000079945,0.30653846689892816,0.3081794959993204,0.3102356020798288,"[0.3071339079997415, 0.3039268569991691, 0.3025128429999313, 0.3030618900011177, 0.3028014229985274, 0.3019031950007047, 0.30257120600072085, 0.30231144499884977, 0.3047235340000043, 0.3035496350003086, 0.3037798580007802, 0.3035140769989084, 0.30473717200038664, 0.30349296900021727, 0.3031198520002363, 0.30223485699934827, 0.3026049550007883, 0.30127563899986853, 0.30157310299910023, 0.3016757030000008, 0.30238696800006437, 0.30201786100042227, 0.3015998770006263, 0.3017297599999438, 0.3029425739987346, 0.30220787200050836, 0.3097000960005971, 0.3053774509990035, 0.30325844300023164, 0.30385756800023955, 0.3023264450002898, 0.31021991899979184, 0.3062353109999094, 0.304095462000987, 0.3034545939990494, 0.304184388000067, 0.30408804999933636, 0.30488553600116575, 0.3042747050003527, 0.3052109849995759, 0.30435396199936804, 0.3056618350001372, 0.3046495729995513, 0.30325742900095065, 0.30284192799990706, 0.30342087899953185, 0.30393571199965663, 0.30349379400104226, 0.30432189899875084, 0.30387068500021996, 0.3044745170009264, 0.3044783900004404, 0.304281220998746, 0.30435262200080615, 0.3031663660003687, 0.3057132529993396, 0.3031677889994171, 0.3046165600007953, 0.30329834499934805, 0.31053031100054795, 0.3069283929999074, 0.30631289099983405, 0.3083560579998448, 0.3099601960002474, 0.30365896900002554, 0.30428311600007873, 0.30325002999961725, 0.30206570999871474, 0.3032417680005892, 0.3028044910006429, 0.30219072500040056, 0.3037480479997612, 0.3034019009992335, 0.3028087679995224, 0.3039518900004623, 0.3015947940002661, 0.3033112920002168, 0.3034547150000435, 0.3023228959991684, 0.3039195050005219, 0.30988555399926554, 0.3029087830000208, 0.3025268410001445, 0.304003578999982, 0.3037717960014561, 0.3034793679998984, 0.30260862700015423, 0.3040192689986725, 0.3036694020011055, 0.30422015099975397, 0.3045588280001539, 0.303696613998909, 0.3058660469996539, 0.3066598460009118, 0.30643868100014515, 0.30505161300061445, 0.3056245999996463, 0.3037997859992174, 0.30484869400061143, 0.3047657219995017, 0.3051213190010458, 0.30393826899853593, 0.30445818299995153, 0.3056851920009649, 0.3061037080005917, 0.30417158999989624, 0.3059934759985481, 0.30597544600095716, 0.3052456740006164, 0.30446598199887376, 0.304912695000894, 0.3040380839993304, 0.30489457799922093, 0.31170258200108947, 0.3047725269989314, 0.3058464810001169, 0.3063942250009859, 0.30507593899892527, 0.30452177100050903, 0.3080648489994928, 0.3061719080014882, 0.3070284819987137, 0.3064574250001897, 0.3045297570006369, 0.30587915599971893, 0.3076493959997606, 0.309199006000199, 0.30197966399828147, 0.3033080010009144, 0.3019802779999736, 0.30310353099957865, 0.30167388300105813, 0.30080179299875454, 0.30394411900124396, 0.3090539909990184, 0.30344218900063424, 0.30300476899901696, 0.30278881400045066, 0.3026891789995716, 0.3026805610006704, 0.30275334699945233, 0.30164107699965825, 0.3024530880011298, 0.3029965120003908, 0.30252654399919265, 0.3018854060010199, 0.3022042449993023, 0.30271324800014554, 0.30203671299932466, 0.3019465670004138, 0.3048901799993473, 0.3017107870000473, 0.3030862610012264, 0.3032648779990268, 0.30316766100077075, 0.3029588609988423, 0.3036780930015084, 0.3025255249995098, 0.30376672099919233, 0.3026055690006615, 0.30276281199985533, 0.30329071300002397, 0.3020725999995193, 0.30294859199966595, 0.30085602500003006, 0.30239806400095404, 0.3033662859998003, 0.3106209540001146, 0.30315211199922487, 0.3027512620010384, 0.3037636509998265, 0.3034309550002945, 0.3029242800002976, 0.3040668369994819, 0.3041110390004178, 0.30279789799897117, 0.30367870699956256, 0.3040857230007532, 0.30382603900034155, 0.3036548359996232, 0.3055684699993435, 0.303714249001132, 0.3026670399995055, 0.30504507399928116, 0.30426826500115567, 0.30346402999930433, 0.3039348599995719, 0.3037278360006894, 0.30348297400087176, 0.30950051399850054, 0.30279291800070496, 0.3041378130001249, 0.3028259620004974, 0.3072156759990321, 0.3051922200011177, 0.3038096169984783, 0.30326598700048635, 0.30468586000097275, 0.30483968399857986, 0.30671483900005114, 0.3090727760009031, 0.3098011370002496, 0.303593364000335, 0.3031097629991564, 0.30606986400016467, 0.30373424500066903, 0.3043320519991539, 0.3036697820007248, 0.30420030699860945, 0.30410157600090315, 0.304636742999719, 0.30415662999985216, 0.30533455800105, 0.30542951399911544, 0.3043979819995002, 0.30566189999990456, 0.3049073890015279, 0.3064194989983662, 0.30393844100035494, 0.30636387300000933, 0.30532961599965347, 0.3118024380000861, 0.3048845710000023, 0.305132166000476, 0.3066085509999539, 0.3041491029998724, 0.30488976200103934, 0.3045865229996707, 0.3049149269991176, 0.3058463010002015, 0.3056846060007956, 0.3051623409992317, 0.30586545200094406, 0.30559523800002353, 0.30479139399903943, 0.3054030399998737, 0.30520066800090717, 0.3077267219996429, 0.30685654700027953, 0.3059312030000001, 0.306103744000211, 0.30559365899898694, 0.3064166580006713, 0.30569289199956984, 0.30572045400003844, 0.30701942599989707, 0.30571903599957295, 0.3060324310008582, 0.30682470799911243, 0.3077705949999654, 0.30743252000138455, 0.3063038479995157, 0.30876606600031664, 0.30421788199964794, 0.30487888700008625, 0.304530364999664, 0.304191950999666, 0.3061068349998095, 0.305866648999654, 0.30420380100076727, 0.30358054299904325, 0.30448782200073765, 0.30396707699947, 0.30378431500139413, 0.3035885219996999, 0.30462202799935767, 0.30575387899989437, 0.3035869690011168, 0.30352741799833893, 0.3040374280008109, 0.30443981700045697, 0.30446703499910655, 0.30337889200018253, 0.30415278200052853, 0.3036101130001043, 0.31023782000011124, 0.3051217279989942, 0.30408917299973837, 0.3027903310012334, 0.3071485990003566, 0.3045454069997504, 0.3075414550003188, 0.3053073539995239, 0.30472302899943315, 0.30568266400041466, 0.3050216850006109, 0.305744938999851, 0.30641166399982467, 0.3047887089996948, 0.30528842599960626, 0.30465976300001785, 0.3076309530006256, 0.30723267599933024, 0.3056978250006068, 0.30652790000021923, 0.3061542750001536, 0.30450233400006255, 0.3052837709983578, 0.30549265000081505, 0.30438526699981594, 0.3051304060008988, 0.3046704389998922, 0.30587790199933806, 0.305211655999301, 0.30582402200161596, 0.3050026799992338, 0.3043625650007016, 0.30759193599988066, 0.31111400699955993, 0.3048560320003162, 0.3065619469998637, 0.3046590039994044, 0.305204563001098, 0.30615207699884195, 0.30527903599977435, 0.30840713799989317, 0.3018237220003357, 0.30143443499946443, 0.3043430210000224, 0.30393834999995306, 0.3034137089998694, 0.30247743900144997, 0.3020826729989494, 0.3028766560000804, 0.30363115100044524, 0.30283857399990666, 0.30211022599905846, 0.3034588140017149, 0.30337308999878587, 0.3090060699996684, 0.30367535799996404, 0.304038625999965, 0.3025829320013145, 0.3028827050002292, 0.30382159299915656, 0.30435478999970655, 0.3045443669998349, 0.30382548099987616, 0.30415216800065537, 0.3040650309994817, 0.30309074400065583, 0.3024819749989547, 0.30229242800123757, 0.3038300909993268, 0.3034871039999416, 0.30403829900023993, 0.30193810799937637, 0.30156624900155293, 0.3019941779984947, 0.3027257930007181, 0.30279468499975337, 0.3060970839997026, 0.30428133600071305, 0.3047387799997523, 0.30598383100004867, 0.30208902500089607, 0.3032517639985599, 0.3037736610003776, 0.30223546299930604, 0.3028673759999947, 0.30340071900172916, 0.30582628099909925, 0.3097822029994859, 0.30445412500012026, 0.30597398999998404, 0.3031032390008477, 0.30709261300034996, 0.30822156199974415, 0.30498238600011973, 0.3040002079997066, 0.30621250699914526, 0.30972251900129777, 0.3100124129996402, 0.30746895399897767, 0.30763929700151493, 0.30646571299985226, 0.3056797359986376, 0.3056611360007082, 0.30770400400069775, 0.3041033299996343, 0.3027220710009715, 0.3027192270001251, 0.309242623998216, 0.30391503400096553, 0.3023700590001681, 0.3050273600001674, 0.3028769149987056, 0.3022433020014432, 0.3028958409995539, 0.3026022539997939, 0.3009949709994544, 0.30217691200050467, 0.30339289699986693, 0.304207070999837, 0.30305328400027065, 0.30387588200028404, 0.3032120930001838, 0.30296895500032406, 0.3036315699991974, 0.30458899199948064, 0.30367828000089503, 0.30248664499958977, 0.3035238680004113, 0.3026806170000782, 0.3032152870000573, 0.3032942299996648, 0.3047138090005319, 0.30486556100004236, 0.30428716899950814, 0.30526818900034414, 0.304161498999747, 0.3047236759994121, 0.3037033970012999, 0.30497962399931566, 0.3039448830004403, 0.3102301719991374, 0.30448641300063173, 0.3032105610000144, 0.30426809699929436, 0.3037427060007758, 0.3030780440003582, 0.30323906499870645, 0.30319408099967404, 0.3026865540014114, 0.3025963079999201, 0.3029844579996279, 0.30289611499938474, 0.30230070400102704, 0.3057393549988774, 0.3033140530005767, 0.3044351270000334, 0.30327862599915534, 0.3050604840009328, 0.3034374239996396, 0.30353691799973603, 0.302935386000172, 0.30217411200101196, 0.30169431899958, 0.3038251389989455, 0.30246080700089806, 0.302132722999886, 0.3085780179990252, 0.3040854280006897, 0.30579221100015275, 0.30294481300006737, 0.3022818949993962, 0.303253575000781, 0.30197955899893714, 0.3026492600001802, 0.30352982899967174, 0.3031408150000061, 0.3030467120006506, 0.30404755999916233, 0.30528696200053673, 0.3032578599995759, 0.3037959650009725, 0.30333046600026137, 0.30418211599862843, 0.3043877620002604, 0.3038736730013625, 0.30336882099982176, 0.3028488939999079, 0.3036753010001121, 0.30444197799988615, 0.30302207800014, 0.30220265799835033, 0.30315436300043075, 0.30401704700125265, 0.3022796229997766, 0.30598990399994364, 0.30479422099961084, 0.30198419899898, 0.3020516590004263, 0.3029592680013593, 0.30269775799933996, 0.3023366069992335, 0.3023354670003755, 0.3021338510006899, 0.3021994509999786, 0.3026061549990118, 0.30240087200036214, 0.3032777660009742, 0.30331159599882085, 0.3037337839996326, 0.30326724800033844, 0.30248569499963196, 0.3041771360003622, 0.30376638599955186, 0.30348570100068173, 0.30307136100054777, 0.3023486469992349, 0.30209753300005104, 0.3030255430003308, 0.3012353030007944, 0.3029088969997247, 0.302289010000095, 0.3030771609992371, 0.30318832000011753, 0.30260597599954053, 0.30250898099984624, 0.30168785900059447, 0.30455251099920133, 0.308270095001717, 0.30380233399955614, 0.3065358579988242, 0.30344057900038024, 0.30162154600111535, 0.30159257499872183, 0.3038788600006228, 0.3012045070008753, 0.3024113709998346, 0.30266211399975873, 0.30216416499933985, 0.30507843799932743, 0.303988528001355, 0.3043352770000638, 0.30305960799887544, 0.30228713900032744, 0.3023118519995478, 0.3026115630000277, 0.30219082000076014, 0.3030412320003961, 0.30365486399932706, 0.3035577589998866, 0.310145363000629, 0.30449719699936395, 0.30427877900001477, 0.3023436679995939, 0.30800867000107246, 0.3079293310001958, 0.3047548819995427, 0.3044389469996531, 0.30331301000114763, 0.3059470919997693, 0.306084956999257, 0.30417437399955816, 0.30536087899963604, 0.3051035300013609, 0.30372757599980105, 0.30313271100021666, 0.3049493119997351, 0.30590195799959474, 0.30390774099942064, 0.30481824500020593, 0.3037258060012391, 0.3048097399987455, 0.3043334110006981, 0.3045616519993928, 0.3034695759997703, 0.3043970310009172, 0.3039207929996337, 0.3040317180002603, 0.3030412640000577, 0.3029629800003022, 0.3081280819988024, 0.3050551190008264, 0.3103016259992728, 0.30470996199983347, 0.30440754600022046, 0.3064451589998498, 0.3057579300002544, 0.3069333610001195, 0.3072845080005209, 0.30722127899935003, 0.30538110900124593, 0.3048966419992212, 0.3056088350003847, 0.31009881300087727, 0.3044546490000357, 0.30394592299853684, 0.3044848270001239, 0.30490589500004717, 0.30348448200129496, 0.30439824999848497, 0.3050631410005735, 0.3030040680005186, 0.30356067399952735, 0.30950715400103945, 0.30460050299916475, 0.3032100229993375, 0.30401359500137914, 0.30306150499927753, 0.3041821099996014, 0.3028111010007706, 0.30233202899944445, 0.30211140100072953, 0.3030771559988352, 0.3032807489998959, 0.30288984199978586, 0.3031511880017206, 0.30357551899942337, 0.3027297329990688, 0.3043640410014632, 0.3040234629988845, 0.302651773999969, 0.30261192500074685, 0.30164105699986976, 0.3018576579997898, 0.3009523460004857, 0.30081379399962316, 0.30222629200034135, 0.3022622459993727, 0.30430213200088474, 0.3029304659994523, 0.3037095500003488, 0.3037285529990186, 0.3040030840002146, 0.30296583000017563, 0.3032651609992172, 0.30354356700081553, 0.3092657279994455, 0.3046112350002659, 0.304800977000923, 0.30445797799984575, 0.3033229329994356, 0.3023557960004837, 0.30576182900040294, 0.30506530099955853, 0.303907702998913, 0.3042684280007961, 0.30499370799952885, 0.3061601130011695, 0.3047164689996862, 0.30355477999910363, 0.3034126210004615, 0.3035775600001216, 0.3042765699992742, 0.3038325610013999, 0.30556072799845424, 0.3043188170013309]",tokens/s,3.286334717119279,, -float16-eager-pytorch,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt2,openai-community/gpt2,openai-community/gpt2,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cpu_pytorch.py"", line 114, in benchmark_cpu_pytorch +float32-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cpu,,42,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) - File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking - context_stack.enter_context(energy_tracker.track()) - File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context - result = _cm_type.__enter__(cm) - File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ - return next(self.gen) - File ""/workspace/optimum_benchmark/trackers/energy.py"", line 174, in track - self.emission_tracker.start_task() - File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task - if self._scheduler: -AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3832, in from_pretrained + model = cls(config, *model_args, **model_kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ + self.model = DeciCoderModel(config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ + self.self_attn = DeciCoderAttention(config=config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ + self._init_rope() + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1729, in __getattr__ + raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") +AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-eager-pytorch,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,codegen,Salesforce/codegen-6B-nl,Salesforce/codegen-6B-nl,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cpu_pytorch.py"", line 114, in benchmark_cpu_pytorch +float16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cpu,,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) - File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking - context_stack.enter_context(energy_tracker.track()) - File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context - result = _cm_type.__enter__(cm) - File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ - return next(self.gen) - File ""/workspace/optimum_benchmark/trackers/energy.py"", line 174, in track - self.emission_tracker.start_task() - File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task - if self._scheduler: -AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3832, in from_pretrained + model = cls(config, *model_args, **model_kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ + self.model = DeciCoderModel(config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ + self.self_attn = DeciCoderAttention(config=config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ + self._init_rope() + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1729, in __getattr__ + raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") +AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-eager-pytorch,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,gpt_neox,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cpu_pytorch.py"", line 114, in benchmark_cpu_pytorch +float16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 67, in launch + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 103, in target + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 104, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run report = scenario.run(backend) File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 89, in run self.run_model_loading_tracking(backend) - File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 178, in run_model_loading_tracking - context_stack.enter_context(energy_tracker.track()) - File ""/usr/lib/python3.10/contextlib.py"", line 492, in enter_context - result = _cm_type.__enter__(cm) - File ""/usr/lib/python3.10/contextlib.py"", line 135, in __enter__ - return next(self.gen) - File ""/workspace/optimum_benchmark/trackers/energy.py"", line 174, in track - self.emission_tracker.start_task() - File ""/usr/local/lib/python3.10/dist-packages/codecarbon/emissions_tracker.py"", line 547, in start_task - if self._scheduler: -AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 182, in run_model_loading_tracking + backend.load() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 71, in load + self.load_transformers_model() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 151, in load_transformers_model + self.load_transformers_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 126, in load_transformers_model_with_no_weights + self.pretrained_model = self.automodel_loader.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 559, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3832, in from_pretrained + model = cls(config, *model_args, **model_kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 248, in __init__ + self.model = DeciCoderModel(config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in __init__ + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 215, in + self.layers = nn.ModuleList([DeciCoderDecoderLayer(config) for _ in range(config.num_hidden_layers)]) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 181, in __init__ + self.self_attn = DeciCoderAttention(config=config) + File ""/root/.cache/huggingface/modules/transformers_modules/Deci/DeciCoder-1b/d045c14763eab7225fe79a6bc309890fda7b1483/modeling_decicoder.py"", line 54, in __init__ + self._init_rope() + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1729, in __getattr__ + raise AttributeError(f""'{type(self).__name__}' object has no attribute '{name}'"") +AttributeError: 'DeciCoderAttention' object has no attribute '_init_rope' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa-pytorch,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cpu,,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): @@ -11864,8 +12629,67 @@ ChildProcessError: Traceback (most recent call last): if self._scheduler: AttributeError: 'EmissionsTracker' object has no attribute '_scheduler' +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float32-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cpu,,42,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 57, in launch + raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") +RuntimeError: Isolated process exited with non-zero code -9 +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 57, in launch + raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") +RuntimeError: Isolated process exited with non-zero code -9 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cpu,,42,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,24692.92032,,,,,s,1,61.0899072480006,61.0899072480006,0.0,61.0899072480006,61.0899072480006,61.0899072480006,61.0899072480006,[61.0899072480006],,kWh,0.0006868174514916746,0.0003743408938746743,0,0.0010611583453663488,,MB,25403.76064,,,,,s,10,15.104732252999383,1.5104732252999384,0.006551634276574619,1.509945861000233,1.5190867528990566,1.5213107964495975,1.5230900312900304,"[1.5088509880006313, 1.503296873999716, 1.5010647159997461, 1.5144136610015266, 1.5059537689994613, 1.5235348400001385, 1.5068245349993958, 1.5111596149999968, 1.5185925209989364, 1.5110407339998346]",tokens/s,169.48330874859795,kWh,1.7811766050834118e-05,9.707917005322481e-06,0.0,2.7519683056156597e-05,tokens/kWh,9302432.716161992,MB,25410.277376,,,,,s,10,102.26865267500034,10.226865267500035,0.05118511636110205,10.217199449999498,10.302062812399527,10.30311705619988,10.303960451240163,"[10.227535673999228, 10.260083082001074, 10.304171300000235, 10.159133708999434, 10.301828535999448, 10.172796729000765, 10.18930609400013, 10.177889008999045, 10.206863225999768, 10.269045316001211]",tokens/s,6.160245427326374,kWh,0.00012040841521492996,6.562728329229307e-05,0.0,0.00018603569850722307,tokens/kWh,338644.68220626994,,s,630,102.26462578399878,0.1623248028317441,0.003398559620761968,0.16204792899952736,0.16663295990128973,0.16846189580028295,0.17188225235919163,"[0.1626301689993852, 0.15950255400093738, 0.16004992999842216, 0.1610017910006718, 0.1565360980002879, 0.15939343699938036, 0.16052873500120768, 0.15982096900006582, 0.164525270998638, 0.16164533800110803, 0.16411256099854654, 0.16638799800057313, 0.16297618600037822, 0.1635209889991529, 0.16416188500079443, 0.17081045899976743, 0.16318354999930307, 0.16768738000064332, 0.163253798000369, 0.1646452519999002, 0.16090573799920094, 0.16289432700068573, 0.15975712299950828, 0.16454919900024834, 0.1577180620006402, 0.15991009099889197, 0.15786582100008673, 0.15936706200045592, 0.16833595500065712, 0.16229543000008562, 0.1675512419988081, 0.1669025720002537, 0.1592703240003175, 0.16816525300055218, 0.1649770099993475, 0.15681592200053274, 0.16450912399886874, 0.1702793560016289, 0.16109911499916052, 0.16152369200062822, 0.1610331079991738, 0.16161245399962354, 0.16370467700107838, 0.16457440399972256, 0.16265191200000118, 0.1709281090006698, 0.16131308999865723, 0.15949927600013325, 0.1574709079995955, 0.16465280300144514, 0.16077683299954515, 0.16269613199983723, 0.16249023099953774, 0.1636799730003986, 0.16567376400053035, 0.15737022499888553, 0.15702941400013515, 0.16159696300019277, 0.1589104739996401, 0.15833090700107277, 0.16348085800018453, 0.160264684998765, 0.15624293200016837, 0.16477806200055056, 0.16515490500023589, 0.1602607650002028, 0.16328406799948425, 0.16513199499968323, 0.15924291600094875, 0.16656256899841537, 0.16994910900029936, 0.16430734200002917, 0.1664214680004079, 0.16051705500103708, 0.1585253469984309, 0.16265623000072083, 0.15948363600000448, 0.1632890689998021, 0.16227299000092898, 0.16400901199995133, 0.16148745699865685, 0.16630831000111357, 0.16192712399970333, 0.16331657899900165, 0.1603943080008321, 0.16503505000036967, 0.15960129499944742, 0.16176761700080533, 0.1599917949988594, 0.15967161600019608, 0.1633397130008234, 0.16472559700014244, 0.16175019199909002, 0.16156422099993506, 0.16510033899976406, 0.16431901500072854, 0.16232153300006757, 0.16377365900007135, 0.16726790699976846, 0.16231978599898866, 0.1604183980016387, 0.1660814249989926, 0.16336653799953638, 0.16294298300090304, 0.16853546899983485, 0.17085604299973056, 0.16690997499972582, 0.15831903900107136, 0.16655775199978962, 0.1627570119999291, 0.1660345820000657, 0.15893212399896584, 0.1648049820014421, 0.15985755899964715, 0.1619981479998387, 0.16218144400045276, 0.1591653289997339, 0.1584782179998001, 0.15706385600060457, 0.16232165899964457, 0.15859342299881973, 0.15964900600010878, 0.1609714750011335, 0.16501107499971113, 0.1610095479991287, 0.16505951900035143, 0.17339343799903872, 0.1607359980007459, 0.16000725599951693, 0.15911951800080715, 0.15650677299890958, 0.16240334500071185, 0.16237630199975683, 0.1595175790007488, 0.16344207799920696, 0.1615708919998724, 0.1637056580002536, 0.16415018499901635, 0.1644572330005758, 0.1627314249999472, 0.1587266990009084, 0.16335063599944988, 0.16644392300077016, 0.15962734099957743, 0.15757266999935382, 0.16184235699984129, 0.16412977700019837, 0.16266920800080698, 0.16775561800022842, 0.1636639109983662, 0.16667027900075482, 0.1644547319992853, 0.16613574600160064, 0.16179879699848243, 0.17272345400124323, 0.15979187300035846, 0.16252119499949913, 0.16727668199928303, 0.16678051800045068, 0.16336388899981102, 0.16051698500086786, 0.16682343399952515, 0.16215978999935032, 0.1588885900000605, 0.1632761720011331, 0.16156329799923697, 0.15958487399984733, 0.16154737599936198, 0.15988718400149082, 0.1642441770000005, 0.16581142199902388, 0.16583580000042275, 0.16418262300067, 0.16055144599886262, 0.16342181800064282, 0.1699241139995138, 0.17361543299921323, 0.16310703700037266, 0.16320409000036307, 0.16512219099968206, 0.1641429920000519, 0.1641352740007278, 0.16558442499990633, 0.1619314569998096, 0.163810291000118, 0.1659233159989526, 0.16399076000016066, 0.16527900700020837, 0.1642984800000704, 0.16480339800000365, 0.16913533599836228, 0.16019245200004661, 0.15844752800148854, 0.15987218099871825, 0.16808043099990755, 0.15609100999972725, 0.1571150880008645, 0.1553257659998053, 0.1575897229995462, 0.15793700799986254, 0.16131496100024378, 0.16375706500002707, 0.16325070800121466, 0.16138504599985026, 0.15791393099971174, 0.16217757499907748, 0.16376245300125447, 0.15927049899983103, 0.15800424000008206, 0.16341511700011324, 0.16800019399852317, 0.16116582100039523, 0.15819157300029474, 0.15812605900100607, 0.15772536199983733, 0.15976416899866308, 0.1603804639998998, 0.1631934260003618, 0.16912476700053958, 0.162595694000629, 0.16312538599959225, 0.1584531770004105, 0.1612124289986241, 0.16146743900026195, 0.16235405400038871, 0.1703361859999859, 0.16621403499993903, 0.16027916699931666, 0.16070933300034085, 0.16402851000020746, 0.15833061299963447, 0.16085964500052796, 0.16322642600061954, 0.160410917000263, 0.15844380399903457, 0.1589304600001924, 0.16009184799986542, 0.15902241399999184, 0.15603671300050337, 0.15879567800038785, 0.15766233099930105, 0.1615001320005831, 0.16427440799998294, 0.16556469699935406, 0.1571925040007045, 0.15900709299967275, 0.16390097699877515, 0.16146873200159462, 0.16139795999879425, 0.1661013869997987, 0.16222558700064837, 0.15899462800007313, 0.16528040900084306, 0.16328517100009776, 0.1618790339998668, 0.16171241799929703, 0.1622503519993188, 0.16459061700152233, 0.166649464999864, 0.16592104299888888, 0.16404583300027298, 0.16241332400022657, 0.16913943699910305, 0.1648773560009431, 0.1614107219993457, 0.16000693600108207, 0.16166881099888997, 0.1657628630000545, 0.16299839900057123, 0.16460784300033993, 0.16325400899950182, 0.16096282299986342, 0.16284446499957994, 0.1639805909999268, 0.16524507500071195, 0.1622072530008154, 0.16513186799966206, 0.16742093799985014, 0.17037948200049868, 0.179486912998982, 0.1664864939994004, 0.16550918700158945, 0.16240736699910485, 0.16154435700082104, 0.16076063199943746, 0.16503225099950214, 0.16208759999972244, 0.15924979600094957, 0.162007294000432, 0.1600677979986358, 0.16108492100102012, 0.16645019899988256, 0.17080671599978814, 0.16366460300014296, 0.16121745499913231, 0.17050619900146557, 0.16066046199921402, 0.1604776240001229, 0.16133543300020392, 0.16057678399920405, 0.16203108599984262, 0.16299116000118374, 0.1624904529999185, 0.16557371799899556, 0.16138607799985039, 0.15920105000077456, 0.16180238999913854, 0.1599789870015229, 0.16652366099879146, 0.16253613899971242, 0.1608072640010505, 0.16126269800042792, 0.1620647719992121, 0.16250878800019564, 0.15891650699995807, 0.16765911800030153, 0.15909267000097316, 0.15914935499858984, 0.15986862200043106, 0.16101370599972142, 0.16212907299996004, 0.16509641300035582, 0.15966102200036403, 0.15707350299999234, 0.15966365600070276, 0.157693492999897, 0.1608439269984956, 0.16032773400002043, 0.1625966880001215, 0.1593718069998431, 0.16288349900059984, 0.15996148100020946, 0.15837000800092937, 0.1612796849985898, 0.1588249670003279, 0.16218097199998738, 0.16837197300083062, 0.16195404499922006, 0.1582090749998315, 0.15739102600127808, 0.15804147299968463, 0.1593278129985265, 0.1636774240014347, 0.1636203399993974, 0.1601387970003998, 0.15890199699970253, 0.1651418889996421, 0.1596557640004903, 0.1572712870001851, 0.15759943699958967, 0.16043333100060408, 0.15685227099857002, 0.1641941540001426, 0.16248916200129315, 0.16305037599886418, 0.16515107200029888, 0.16140240999993694, 0.16355392500008747, 0.16012848799982748, 0.1613396480006486, 0.16926145700017514, 0.15925872699881438, 0.16397613500157604, 0.16318935599883844, 0.16207275700071477, 0.15790036399994278, 0.16114398900026572, 0.16053850000025705, 0.16461731999879703, 0.16010986600122123, 0.1604218859993125, 0.1623315510005341, 0.16228649299955578, 0.16281336000065494, 0.16282646499894327, 0.1650561149999703, 0.16900161200101138, 0.1629452429988305, 0.16591663299914217, 0.15612180900097883, 0.15969478999977582, 0.15629332199932833, 0.1594881119999627, 0.16140995800014935, 0.16668866200052435, 0.15988126000047487, 0.16065538200018636, 0.16016585099896474, 0.1585729739999806, 0.1592018060000555, 0.16091052100091474, 0.16028729299978295, 0.1630878269988898, 0.1601625850016717, 0.16041226299967093, 0.16201224900032685, 0.16046018299857678, 0.1581962400014163, 0.15805453899884014, 0.15930152700093458, 0.15939668499959225, 0.15646855699924345, 0.156323520001024, 0.1597026849995018, 0.16235881700049504, 0.1657451859991852, 0.1635720169997512, 0.16172255499986932, 0.1628821970007266, 0.16571553699941433, 0.15873451800143812, 0.16417507299956924, 0.1644439430001512, 0.17328312099925824, 0.1594413919992803, 0.16309419700155559, 0.16558556400013913, 0.169352679999065, 0.16670567599976494, 0.16875284799971269, 0.16279260400006024, 0.16168961900075374, 0.16029177100062952, 0.1652143889987201, 0.1599298210003326, 0.15952212599950144, 0.1692923470000096, 0.16483182800038776, 0.16619720400012739, 0.16881593600010092, 0.16115438300039386, 0.15976711499934027, 0.15950259199962602, 0.1580185849998088, 0.1586405480011308, 0.1608043340002041, 0.15815243599899986, 0.15915036499973212, 0.15880508700138307, 0.1623741949988471, 0.15951258300083282, 0.16667775699897902, 0.16506359100094414, 0.16001006699843856, 0.15653321400168352, 0.1582270709986915, 0.15877267900032166, 0.16138106499965943, 0.1604767790013284, 0.1607077489989024, 0.16037833300106286, 0.15777190399967367, 0.15809341999920434, 0.15791429199998674, 0.1631522940006107, 0.16323993199875986, 0.16241392200026894, 0.16328856500149413, 0.1656328429999121, 0.1646186519992625, 0.16347245200086036, 0.1629915819994494, 0.15971194899975671, 0.16134115000022575, 0.16157305799970345, 0.16550121399995987, 0.1632531890008977, 0.16125974299939116, 0.15783463600018877, 0.15762349199940218, 0.15786050499991688, 0.1575705180002842, 0.15768836499955796, 0.1580913870002405, 0.16172838599959505, 0.15837519300112035, 0.15944660199966165, 0.15934770700005174, 0.16461137999976927, 0.16680121299941675, 0.1579320729997562, 0.16307331400093972, 0.16706018900003983, 0.1603375889990275, 0.16059814000072947, 0.16834837699934724, 0.16261450400088506, 0.16699929899914423, 0.15802192299997841, 0.16300341300120635, 0.16120001699891873, 0.16447898000114947, 0.1649718019998545, 0.15818552199925762, 0.16053191299943137, 0.15698607500053186, 0.16454072300075495, 0.16347940099876723, 0.15928252400044585, 0.16121503000067605, 0.16157751900027506, 0.16419065299851354, 0.1622295960005431, 0.1662086360011017, 0.16771845299990673, 0.1655433610012551, 0.160742254998695, 0.16413855300015712, 0.1623285890000261, 0.16299102300035884, 0.16358136599956197, 0.16033298900038062, 0.15932010100004845, 0.16075086999990162, 0.1614799460003269, 0.1660748699996475, 0.15702921299998707, 0.1653296019994741, 0.16287245900093694, 0.159827048999432, 0.15997274900109915, 0.15921228199840698, 0.1616280270009156, 0.15988163399924815, 0.1681205000004411, 0.16306226100095955, 0.16433093999876291, 0.16607689500051492, 0.15824941800019587, 0.15827128799901402, 0.17148435700073605, 0.16177339699970616, 0.1643790119996993, 0.16241194500071288, 0.16613384899937955, 0.1620129180009826, 0.16030950299864344, 0.1631483430010121, 0.1667169800002739, 0.1592628580001474, 0.15969466800015653, 0.16397077500005253, 0.1609136879997095, 0.16147835799893073, 0.16138991600018926, 0.15834687400092662, 0.15946630599864875, 0.16333205299997644, 0.16411626600165619, 0.16316385799837008, 0.1625667530006467, 0.1576716290001059, 0.15896516800057725, 0.15831158999935724, 0.16073353300089366, 0.15907068399974378, 0.15817503399921407, 0.15985047899994242, 0.15834288400037622, 0.15666771100040933, 0.16343494299871963, 0.16582222800025193, 0.1632239690006827, 0.1630350899995392, 0.16178039599981275, 0.16213036000044667, 0.16433214000062435, 0.1670347560011578, 0.16293325500009814, 0.1604515849994641, 0.16710302299907198, 0.16663112600144814, 0.16447498699926655, 0.16165763299977698, 0.1600690350005607, 0.1612692180005979, 0.15998833399862633, 0.16161039500002516, 0.1620892829996592, 0.15933293100169976, 0.1569637459997466, 0.15816116299902205, 0.1572523499999079, 0.1590293429999292, 0.16368133099967963, 0.15872637100073916, 0.16629312300028687, 0.16370774899951357, 0.15679536300012842, 0.15921608300050138, 0.15956331899906218, 0.1616767820014502, 0.15924570399874938, 0.15976996900099039, 0.16507920999902126, 0.16183138199994573, 0.16029945599984785, 0.15962790300000052, 0.15943415900073887, 0.15737526200064167, 0.16393380599947704, 0.16605551400061813, 0.1630637869984639, 0.16566241400141735, 0.1647035899986804, 0.16496682400065765, 0.16559119100020325, 0.1691369919990393, 0.16366919000029156, 0.1637628759999643, 0.1605099610005709, 0.16457151400027215, 0.16879639199942176, 0.17093923299944436, 0.17260875000101805, 0.16341357000055723, 0.16386060499826272, 0.16483360600068409, 0.16521839100096258, 0.1663395830000809, 0.1660651459988003, 0.159346321999692, 0.1631952520001505, 0.15870941400135052, 0.1720447729985608, 0.16336730100010755, 0.1585814530008065, 0.16058360900024127, 0.16977325599873438, 0.16698461499981931]",tokens/s,6.1604880003245,, +float16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cpu,,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,12559.220736,,,,,s,1,194.03764496800068,194.03764496800068,0.0,194.03764496800068,194.03764496800068,194.03764496800068,194.03764496800068,[194.03764496800068],,kWh,0.002262840143218744,0.0012333345715665918,0,0.003496174714785336,,MB,12693.405696,,,,,s,10,24.7333787030002,2.47333787030002,0.006109422274615187,2.4724028399996314,2.4826069646003814,2.4831298583002535,2.483548173260151,"[2.473771625998779, 2.4836527520001255, 2.4699904499993863, 2.4710340540004836, 2.468136024999694, 2.48249076600041, 2.4759957100013708, 2.4650670339997305, 2.47700279899982, 2.4662374870003987]",tokens/s,103.50385326406973,kWh,2.9426933187640368e-05,1.6038634770726618e-05,0.0,4.546556795836699e-05,tokens/kWh,5630634.598789578,MB,12695.642112,,,,,s,10,98.46601800399912,9.846601800399913,0.021928345961658498,9.844369463499788,9.877686007299598,9.884311628649902,9.889612125730146,"[9.822295871001188, 9.890937250000206, 9.845086801000434, 9.836660576000213, 9.876213646999531, 9.843652125999142, 9.817492066998966, 9.856913151999834, 9.828985422000187, 9.847781091999423]",tokens/s,6.398146414069603,kWh,0.00011556322541229013,6.298647478523028e-05,0.0,0.00017854970019752042,tokens/kWh,352842.93353786826,,s,630,98.46162090800135,0.1562882871555577,0.0021956531338376047,0.15586629300105415,0.15897853079859486,0.16021232734983643,0.16460944626976926,"[0.1548110399999132, 0.153887697000755, 0.15744844099936017, 0.15454314199996588, 0.164744049001456, 0.15503606599850173, 0.15439303200037102, 0.1533530430006067, 0.15588568900056998, 0.15305366499887896, 0.15346773900091648, 0.1542000809986348, 0.15571407100105716, 0.1548283319989423, 0.15736893100074667, 0.16139128999930108, 0.1596056940015842, 0.15564390399958938, 0.1543171049997909, 0.15418312399924616, 0.15440050100005465, 0.15345744200021727, 0.15420433200051775, 0.15697831400029827, 0.15471599400007108, 0.1565006259988877, 0.15482245500061254, 0.1540301349996298, 0.1550504099996033, 0.15577327299979515, 0.15544378600134223, 0.15570979199947033, 0.16012455499912903, 0.1606920970007195, 0.159863573999246, 0.15606716399997822, 0.15634885300096357, 0.1559136559990293, 0.1569703710010799, 0.15567215100054455, 0.1569631459988159, 0.15428695999980846, 0.15400837100060016, 0.15337684999940393, 0.1540219899998192, 0.1541502930012939, 0.15583066500039422, 0.15664694999941275, 0.15769516399996064, 0.15498517400010314, 0.15784227800031658, 0.15419312799895124, 0.1552053360010177, 0.15372853299959388, 0.15365679999922577, 0.1556007529998169, 0.1568932810005208, 0.15785745799985307, 0.15773653700125578, 0.15677203299856046, 0.15773360500134004, 0.15614669599926856, 0.15592172700053197, 0.1646791729999677, 0.16443873599928338, 0.15480388499963738, 0.15471141099988017, 0.15720726400104468, 0.15540216699992015, 0.15748394199908944, 0.16534170700106188, 0.16212383799938834, 0.15514573799919162, 0.15934547500000917, 0.15600336700117623, 0.15489047799928812, 0.15679242000078375, 0.15388709699982428, 0.1584819370000332, 0.15720813599909889, 0.15579421300026297, 0.15495797299990954, 0.1556634420012415, 0.1548560519986495, 0.15406526900005701, 0.1564678790000471, 0.1543437480013381, 0.1543386349985667, 0.1552657120009826, 0.15650814999935392, 0.15561413199975505, 0.15479582300031325, 0.15575284999977157, 0.15849375000107102, 0.15658719899874995, 0.15393668599972443, 0.1555975560004299, 0.1570803340000566, 0.1556172750006226, 0.1549826930004201, 0.15731382999911148, 0.15565313499973854, 0.15436174500064226, 0.1555051920004189, 0.1554757249996328, 0.15468487500038464, 0.1597432119997393, 0.15624411499993585, 0.15540108200002578, 0.15459886600001482, 0.15700379199915915, 0.167071391000718, 0.15956968299906293, 0.15639683400149806, 0.15799669599982735, 0.15786534200015012, 0.15921798499948636, 0.16392514300059702, 0.15548565899916866, 0.15672145899952739, 0.1569177290002699, 0.155923635000363, 0.1582480500001111, 0.15804378099892347, 0.15708565900058602, 0.15539839300072344, 0.16162534700015385, 0.15619900000092457, 0.15387531299893453, 0.1552775489999476, 0.15682893300072465, 0.15916932099935366, 0.1571563720008271, 0.15750704600031895, 0.1554610169987427, 0.15725152699997125, 0.15616522400159738, 0.15733394399831013, 0.15872571200088714, 0.1535780979993433, 0.15419037900028343, 0.1568449269998382, 0.15749198100093054, 0.15782388600018749, 0.16116029799923126, 0.15586052300022857, 0.16035124200061546, 0.15775923899855115, 0.1561803790009435, 0.15424752999933844, 0.15334149100090144, 0.15399011699992116, 0.1541704960000061, 0.15586436900048284, 0.15576324599896907, 0.15467802500097605, 0.1559695420000935, 0.1564499829983106, 0.15471705900017696, 0.15804299299998092, 0.15439544900073088, 0.15497220900033426, 0.15662695999890275, 0.1554810819998238, 0.15468460000010964, 0.15476233800109185, 0.15795423199961078, 0.15883636900070996, 0.1588872859992989, 0.15760784700069053, 0.15389763099847187, 0.15586821700162545, 0.15418640099960612, 0.15509431899954507, 0.15507752899975458, 0.15935508599977766, 0.15753102800044871, 0.15438424599960854, 0.15428519200031587, 0.1557493180007441, 0.15766103999885672, 0.15555665100146143, 0.15663119399869174, 0.15652959600083705, 0.15480379200016614, 0.15506159799952002, 0.1550307020006585, 0.15554761400017014, 0.15714376399955654, 0.15724184099963168, 0.15502407500025583, 0.15558928000064043, 0.15433814599964535, 0.15540506800061848, 0.15632776500024193, 0.1557528919984179, 0.15395767600057297, 0.1558998780001275, 0.15700829699926544, 0.16082828200160293, 0.15571411799828638, 0.15444953700171027, 0.15375884199966094, 0.15551192599923525, 0.15593352299947583, 0.15393972600031702, 0.15547476100073254, 0.15928369900029793, 0.15533906599921465, 0.1549985479996394, 0.15396060000057332, 0.15583621299992956, 0.1541283549995569, 0.1552994560006482, 0.15402677400015818, 0.15562284400039061, 0.15823057099987636, 0.15615651000007347, 0.1539962619990547, 0.15423541400014074, 0.15555488199970569, 0.1565138850000949, 0.15736570300032326, 0.15447763799966197, 0.1579159740012983, 0.15583987699938007, 0.15529722300016147, 0.1536550429991621, 0.15540725000028033, 0.15453716900083236, 0.15592443300010927, 0.1554647289995046, 0.15564350199929322, 0.15655763300128456, 0.1569867959988187, 0.1576130600005854, 0.15735306100032176, 0.1544743769991328, 0.15638808000039717, 0.15663470199979201, 0.15551324800071598, 0.15735587800008943, 0.15794258200003242, 0.1584673439992912, 0.15573692900034075, 0.15875956699892413, 0.15599480200035032, 0.15856395900118514, 0.16597338399878936, 0.15705665899986343, 0.15692451100039762, 0.15505965699958324, 0.16015414200046507, 0.15613377199952083, 0.1559978780005622, 0.16397262300051807, 0.15638810699965688, 0.15906635999999708, 0.16001795400006813, 0.1568432539988862, 0.15571852799985209, 0.16307634000077087, 0.15658687799987092, 0.1573567039995396, 0.15463321400056884, 0.15500237599917455, 0.15552891500010446, 0.1540922470012447, 0.15475931899891293, 0.15941303300132859, 0.15591326599860622, 0.15577106199998525, 0.15783142500004033, 0.15662169600000198, 0.1559217810008704, 0.15563263799958804, 0.15951894499994523, 0.15661930600072083, 0.15806967099888425, 0.15555011600008584, 0.15683685899966804, 0.15567950700096844, 0.15433320100055425, 0.15576866199990036, 0.15796047699950577, 0.1550364790000458, 0.15516324800046277, 0.1556933960000606, 0.1557010619999346, 0.15737505099968985, 0.1554215819996898, 0.1548847449994355, 0.1590307650003524, 0.15806868599975132, 0.15821291700012807, 0.1598500910004077, 0.15713104399947042, 0.15857010400031868, 0.15587020400016627, 0.15669713399984175, 0.15529616800085932, 0.15747078599997621, 0.15641929999947024, 0.15858084799947392, 0.15728163800122275, 0.15545157799897424, 0.15659000699997705, 0.15580218200011586, 0.1585019789999933, 0.1552708260005602, 0.15532064699982584, 0.15555258699896513, 0.15416167200055497, 0.15410888199949113, 0.15449669100053143, 0.16144851999888488, 0.1544283670009463, 0.15606865600057063, 0.15498231999845302, 0.15596183400157315, 0.16018672099926334, 0.16405509399919538, 0.15528649000043515, 0.1551347229997191, 0.15477758000088215, 0.15564942200035148, 0.15611513199837646, 0.15468184300152643, 0.15570374199887738, 0.15715860500131384, 0.15489179699943634, 0.15580263499941793, 0.15531923600065056, 0.15590614699976868, 0.15546035800070968, 0.15964853599871276, 0.15436811099971237, 0.1601255530003982, 0.15757970300001034, 0.15628385300078662, 0.1553667400003178, 0.1567633089998708, 0.1571152099986648, 0.15476479600147286, 0.15381632699973125, 0.15765969299900462, 0.15492981600073108, 0.15603725199980545, 0.15536607899957744, 0.1547505779999483, 0.15411913200114213, 0.1550948829990375, 0.1552895280001394, 0.15428650600006222, 0.15613732999918284, 0.15668246900168015, 0.1558098109999264, 0.15509364099852974, 0.15448621500036097, 0.15573818099983328, 0.15536807200078329, 0.15624388799915323, 0.15572487600002205, 0.15745014100139088, 0.1589727269983996, 0.15524247700159322, 0.15839755299930403, 0.157022641000367, 0.15584395400037465, 0.1597258019992296, 0.15730768899993564, 0.15569797600073798, 0.15527548199861485, 0.15652455600138637, 0.15601212099863915, 0.15544725000108883, 0.15582277499925112, 0.1547984049993829, 0.16185917899929336, 0.15628457499951764, 0.16424685200036038, 0.15437366600053792, 0.15326789199934865, 0.15400458899966907, 0.1562785750011244, 0.1557180210002116, 0.1543659649996698, 0.1544171429995913, 0.1557897820002836, 0.15519821399902867, 0.15642915799980983, 0.1558969090001483, 0.15433046399994055, 0.15654777600138914, 0.15442196099866123, 0.15279867500066757, 0.15270439100095246, 0.15261086899954535, 0.1544679910002742, 0.15408176599885337, 0.15751893699962238, 0.15379149100044742, 0.15486458300074446, 0.15468506399884063, 0.15452230200025951, 0.15434847200049262, 0.1545583990009618, 0.15472490699903574, 0.15649612699962745, 0.1559605680013192, 0.1576576319985179, 0.154999640000824, 0.15577477199985879, 0.15633554399937566, 0.15383994900003017, 0.15946220000114408, 0.16040483199867595, 0.1572152550015744, 0.15683833399998548, 0.15529472099842678, 0.15521157000148378, 0.15679037099835114, 0.1551940140016086, 0.15490828799920564, 0.1579661590003525, 0.1557068040001468, 0.15521318100036297, 0.15443428799881076, 0.15421143000094162, 0.15595251400009147, 0.156286862998968, 0.15802657400126918, 0.15970271099831734, 0.16029279000031238, 0.15537885800040385, 0.15487519800080918, 0.15597523599899432, 0.15634329800013802, 0.15556126800038328, 0.15497838299961586, 0.15462074500101153, 0.1617437479999353, 0.15714404700156592, 0.1550416359987139, 0.15410334800071723, 0.15733821200046805, 0.15730180399987148, 0.15521771899875603, 0.15693245800139266, 0.15812712499973713, 0.1571187720001035, 0.15343790799852286, 0.1532053190003353, 0.15498920299978636, 0.1566614280000067, 0.15615259000151127, 0.15439569299996947, 0.15465877299902786, 0.15291544499996235, 0.15596092300074815, 0.15895875099886325, 0.15739132100134157, 0.1559881739995035, 0.15437500100051693, 0.1575623619992257, 0.1580292720009311, 0.15403499599960924, 0.15335782900001504, 0.15773932699994475, 0.15623614499963878, 0.1552798470002017, 0.1651376119989436, 0.15843293400030234, 0.15513559300052293, 0.15497061899986875, 0.15593552099926455, 0.15407429200058687, 0.1560324980000587, 0.1555385280007613, 0.15669678099948214, 0.15690222599914705, 0.16095610200136434, 0.15693526799987012, 0.15461297900037607, 0.15461198699995293, 0.15597178999996686, 0.15827247000015632, 0.15885865299969737, 0.1575633979991835, 0.15903499399973953, 0.1575681930007704, 0.15630125099960424, 0.15788708400032192, 0.155302605999168, 0.15468054600023606, 0.15447705000042333, 0.15570323199972336, 0.15496579600039695, 0.15589436899972497, 0.15664690999983577, 0.1602433280004334, 0.1568854309989547, 0.15544259600028454, 0.15740945600009582, 0.1613317810006265, 0.15267292799944698, 0.15220529500038538, 0.1538247609987593, 0.15271386000131315, 0.15481638999881397, 0.15390883100008068, 0.1523369400001684, 0.15636268900016148, 0.15357203900020977, 0.15444846399987, 0.15604472300037742, 0.15438417800032767, 0.153223679999428, 0.15327957199951925, 0.15361135600142006, 0.1568613639992691, 0.15562621699973533, 0.1551994140008901, 0.15547967499878723, 0.15396129500004463, 0.15368022999973618, 0.1569911510014208, 0.15578773299966997, 0.15778578699973878, 0.16205482899931667, 0.15989708900087862, 0.1583903470000223, 0.15760514799876546, 0.15724728600071103, 0.15756807700017816, 0.15744672000073479, 0.15506639699924563, 0.15346380600021803, 0.1560805919998529, 0.1549357259991666, 0.15761962800024776, 0.15829120700072963, 0.1602332780003053, 0.15793556199969316, 0.1571751960000256, 0.15706122099982167, 0.1580815889992664, 0.15681558000142104, 0.1553207229990221, 0.15422158600085822, 0.1542929440001899, 0.1538795519991254, 0.15652953599965258, 0.1565849580001668, 0.15778235000107088, 0.15521825499854458, 0.156909610999719, 0.15868551000130537, 0.1544932830001926, 0.15903320199868176, 0.15855482900042261, 0.1573732150009164, 0.1543563710001763, 0.15400111700000707, 0.15693771499900322, 0.15601202700054273, 0.15525664699998742, 0.15662034299930383, 0.15745020800022758, 0.15746345099978498, 0.1537835199997062, 0.15412960700086842, 0.15183390299898747, 0.15389417400001548, 0.15450405900082842, 0.15330814399931114, 0.15436752999994496, 0.15669405200060282, 0.15644948999943153, 0.15593593300036446, 0.156041747000927, 0.15969826099899365, 0.15590534700095304, 0.1563796310001635, 0.15789168599985715, 0.15850342999874556, 0.15594527600114816, 0.15480090599885443, 0.1545159209999838, 0.156122784001127, 0.1559416379986942, 0.1552790929999901, 0.155688133001604, 0.15318608599955041, 0.15428097000039998, 0.15497389199845202, 0.15656029599995236, 0.1559653980002622, 0.154244531000586, 0.1572874129997217, 0.1563482630008366, 0.15349962999971467, 0.15630393699939304, 0.154109884000718, 0.1545387389996904, 0.1544203089997609, 0.15400430599947867, 0.1559338140014006, 0.15623478799898294, 0.15636778400039475, 0.1595788859995082, 0.15550987599999644, 0.15556269300032, 0.15623332900031528, 0.15960704399913084, 0.1564726750002592, 0.15784743100084597, 0.15488134899896977, 0.15472298500026227, 0.1567768149998301, 0.1564286049997463, 0.15534544700130937, 0.1558825430001889, 0.16634679599883384, 0.16346051599975908, 0.16128921700146748, 0.1592199159986194, 0.1590699959997437, 0.15796847800083924, 0.15772420000030252]",tokens/s,6.398432142292753,, +float32-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cpu,,42,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 57, in launch + raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") +RuntimeError: Isolated process exited with non-zero code -9 +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +bfloat16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cpu,,42,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,12557.4144,,,,,s,1,160.591200709001,160.591200709001,0.0,160.591200709001,160.591200709001,160.591200709001,160.591200709001,[160.591200709001],,kWh,0.0018675773117131965,0.0010179009647362854,0,0.0028854782764494817,,MB,12715.692032,,,,,s,10,20.744213773999945,2.0744213773999944,0.004109506701927075,2.074975260999963,2.0789888467992568,2.079361468399111,2.0796595656789942,"[2.0707532970009197, 2.0675079579996236, 2.0741580339999928, 2.0772983850001765, 2.0696799870001996, 2.0716281139993953, 2.07875537900145, 2.0789060419992893, 2.0757924879999337, 2.079734089998965]",tokens/s,123.40790679705646,kWh,2.4512968279445256e-05,1.3360325130810268e-05,0.0,3.7873293410255525e-05,tokens/kWh,6759380.474967593,MB,12718.895104,,,,,s,10,78.31957292699917,7.831957292699917,0.027485638328455634,7.834411050500421,7.862582055299572,7.8688247806498115,7.873818960930002,"[7.8171885249994375, 7.861194782999519, 7.83026300200072, 7.790752255999905, 7.87506750600005, 7.84693106099985, 7.783675560998745, 7.826763482000388, 7.849177652000435, 7.838559099000122]",tokens/s,8.043966232900889,kWh,9.215921759569203e-05,5.023038379179303e-05,0.0,0.0001423896013874851,tokens/kWh,442448.03964692604,,s,630,78.31565358299667,0.12431056124285186,0.0019272660815374935,0.1238618690003932,0.12708478719923733,0.128184211950429,0.1305181124500814,"[0.12679927199860686, 0.12304929200035986, 0.12351423800100747, 0.12111025399826758, 0.1270664920011768, 0.12066291799965256, 0.12274649599930854, 0.12159226699986903, 0.12525991900110967, 0.12233983799887937, 0.12322868300179834, 0.12253932799831091, 0.1234100790006778, 0.121671561999392, 0.12271248000070045, 0.12320764700052678, 0.12185416600004828, 0.12299845999950776, 0.12274887199964724, 0.12275456600036705, 0.12259602199992514, 0.12315522199969564, 0.12251253999966139, 0.12260255700130074, 0.12563534899891238, 0.1228489780005475, 0.12332383300054062, 0.12302703299974382, 0.12446295599875157, 0.1253717730014614, 0.12396106199958012, 0.12288781300048868, 0.12650791199848754, 0.12359863600067911, 0.12373835499965935, 0.12255361099960282, 0.12332847400102764, 0.12483265099945129, 0.12389881099988997, 0.12323818799995934, 0.12406055800056492, 0.12350578299992776, 0.12630831400019815, 0.1264408539991564, 0.13204636000045866, 0.12796061799963354, 0.1248261700002331, 0.12434843400114914, 0.12516793599934317, 0.12503397000000405, 0.12460782100060896, 0.12431022399869107, 0.12637608600016392, 0.1254124139995838, 0.12619821900079842, 0.1249298729999282, 0.12441472399950726, 0.1240548170007969, 0.12496170099984738, 0.1233459840004798, 0.12438250400009565, 0.12395085400021344, 0.12481150999883539, 0.12840714000049047, 0.1283874469991133, 0.12165164299949538, 0.12382984400028363, 0.12290941399987787, 0.12440523000077519, 0.1232712029996037, 0.12447758000053, 0.12382151499878091, 0.12515992700173229, 0.12112625199915783, 0.12195038400022895, 0.12392566199923749, 0.12428889399961918, 0.12454337600138388, 0.1248324699990917, 0.12330764400030603, 0.12336740300088422, 0.12400517499918351, 0.12349168800028565, 0.12386879499899806, 0.1250777700006438, 0.12361849000080838, 0.12517732199921738, 0.12980573499953607, 0.12841640999977244, 0.12385568800164037, 0.12428733699925942, 0.12343497300025774, 0.12362081699939154, 0.12775755300026503, 0.12617012000009709, 0.12313005400028487, 0.12362586999915948, 0.12368854499982262, 0.12465478000012808, 0.12636581200058572, 0.12504671500028053, 0.12321360299938533, 0.12603983500048344, 0.1269061259990849, 0.12450809500114701, 0.12420397699861496, 0.12488498800121306, 0.12512492799942265, 0.12616014200102654, 0.12826967099863396, 0.12455764200058184, 0.12320365800042055, 0.12461431799965794, 0.12538829500044812, 0.12379526399854512, 0.1258687710014783, 0.12419292599952314, 0.12269618400023319, 0.12422561799940013, 0.12557946800006903, 0.12779559000045992, 0.1281854310000199, 0.12737812799969106, 0.12461507499938307, 0.12289377800152579, 0.12373236899838957, 0.13130190499941818, 0.1288642620002065, 0.12303461299961782, 0.12382426600015606, 0.12145676300133346, 0.12337329899855831, 0.122545058000469, 0.12288598700069997, 0.12376255799972569, 0.12391948999902525, 0.1235200610008178, 0.1261242000000493, 0.12273783799901139, 0.12495708400092553, 0.1253465909994702, 0.12241492999964976, 0.12366791100066621, 0.12363257399920258, 0.12339990600048623, 0.12425084200003766, 0.1225025439998717, 0.12386724700081686, 0.12517162199947052, 0.12321965699993598, 0.12390747000063129, 0.12834610499885457, 0.1265350240009866, 0.12480770100046357, 0.12350441899980069, 0.12214673400012543, 0.12410127099974488, 0.123906712000462, 0.12280872599876602, 0.1301797300002363, 0.12436653099939576, 0.12515917500059004, 0.12684134300070582, 0.12400450200038904, 0.12567307399876881, 0.12407113399967784, 0.12370131900024717, 0.12303732700092951, 0.12278932699882716, 0.12273361400002614, 0.12304890999985219, 0.12617949500054237, 0.1261966450001637, 0.12370094899961259, 0.12435083800119173, 0.12369846799992956, 0.12416344699886395, 0.12257467400013411, 0.1234234399998968, 0.12299357800111466, 0.12720596299914178, 0.12450849900051253, 0.12274917300055677, 0.1270805179992749, 0.12325511800008826, 0.12275617200066335, 0.12397513299947605, 0.12330969600043318, 0.12230168799942476, 0.12691236399950867, 0.12379906800015306, 0.12284243300018716, 0.12227716000052169, 0.12261955499889154, 0.12199385600069945, 0.12223753500074963, 0.12424528100018506, 0.12336713399963628, 0.12099018799926853, 0.12290610499985632, 0.12163836000036099, 0.12126271300076041, 0.12294157299947983, 0.12307050699928368, 0.1232438330007426, 0.12419308599965007, 0.12461745700056781, 0.12482074599938642, 0.12353539200012165, 0.12454568199973437, 0.12213751300078002, 0.1257925879999675, 0.12306930900012958, 0.12271048199909274, 0.12375868799972523, 0.12848380600007658, 0.12584794000031252, 0.12321886400059157, 0.12373588499940524, 0.12271784400036267, 0.1234002450000844, 0.12330605799979821, 0.123962126999686, 0.12296642799992696, 0.12305316200036032, 0.1227581260009174, 0.12306417399850034, 0.1228662940011418, 0.12331273199924908, 0.12334770400048001, 0.12309091200040712, 0.12536141599957773, 0.12389082899971982, 0.12308494799981418, 0.12444720600069559, 0.12697375399875455, 0.12542211600157316, 0.12290106499858666, 0.1225111450003169, 0.12035778900099103, 0.12311806899924704, 0.12658428800023103, 0.1266778909994173, 0.1253979020002589, 0.12137848600104917, 0.12534481900001992, 0.1249759449983685, 0.12354410700027074, 0.1251563019995956, 0.12319600200135028, 0.12327290699977311, 0.12211163599931751, 0.1286500440000964, 0.12418371599960665, 0.12515583100139338, 0.12399055300011241, 0.12779060999855574, 0.12448962300004496, 0.12342048800019256, 0.1250911589995667, 0.12496692300010181, 0.12297112200030824, 0.12452030300119077, 0.12430962899998121, 0.12432441300006758, 0.12756056599937438, 0.12358956100069918, 0.12419043199952284, 0.12306379399888101, 0.1260287340010109, 0.12474444399958884, 0.1251310099996772, 0.1247126020007272, 0.12416612900051405, 0.12370911899961357, 0.1277325769988238, 0.12952417400083505, 0.12405991199921118, 0.12958384400008072, 0.1278391300002113, 0.1266899390011531, 0.12728537699877052, 0.13229815700105974, 0.1277020259985875, 0.1272201040010259, 0.1284782759994414, 0.1257243380005093, 0.12359887499951583, 0.12560675799977616, 0.12282294400029059, 0.1240102259998821, 0.12462478000088595, 0.12502275199949509, 0.1235549169996375, 0.1220798440008366, 0.12219746499977191, 0.12264215299910575, 0.12227729400001408, 0.12279942799978016, 0.12572532100057288, 0.1273632180000277, 0.12377572700097517, 0.12281177399927401, 0.12291656600064016, 0.1271264949991746, 0.12984633700034465, 0.1244729179998103, 0.12250627799949143, 0.12393678300031752, 0.12197763100084558, 0.12371172299936006, 0.1220567079999455, 0.12312487999952282, 0.12405607500113547, 0.12311148699882324, 0.12977618100012478, 0.12562473000070895, 0.12357406499904755, 0.12489864500093972, 0.12756933600030607, 0.13140580399885948, 0.12330442600068636, 0.1231174699987605, 0.12593572100013262, 0.1242244130007748, 0.12374353499944846, 0.1259242970008927, 0.12517956599913305, 0.12698450700008834, 0.12682075100019574, 0.1236047120000876, 0.12294963599924813, 0.12463277200004086, 0.1269494250009302, 0.12380630100051349, 0.12431282299985469, 0.12537566100036202, 0.12374913999883574, 0.12535178100006306, 0.12442363099944487, 0.12197695800023212, 0.12210772599974007, 0.1265588270016451, 0.12343468899962318, 0.12458603699997184, 0.12446026100042218, 0.12392195199936396, 0.12625151300017023, 0.12525728499895195, 0.12771608799994283, 0.1222150440007681, 0.12270257599993784, 0.1229028540001309, 0.12343365699962305, 0.12390685399986978, 0.1243867709999904, 0.12418263300060062, 0.12304490099995746, 0.1276747259998956, 0.12238243200044963, 0.12305772199943021, 0.12218892500095535, 0.12867609299973992, 0.12876437500017346, 0.1240436439984478, 0.12458052600049996, 0.12254781800038472, 0.1219762599994283, 0.12305866800124932, 0.12315174299874343, 0.12262667300092289, 0.12332628100011789, 0.12301729899991187, 0.12377960199955851, 0.12355440600003931, 0.12544034100028512, 0.12376778500038199, 0.12262904299859656, 0.12535575699985202, 0.12292437199903361, 0.12218763800046872, 0.12177322100069432, 0.12302119299965852, 0.12129038399871206, 0.12287696400017012, 0.12410120000095048, 0.12232167000001937, 0.12181059300019115, 0.12205427799926838, 0.12287663000097382, 0.12813499000003503, 0.12329127399971185, 0.12467137299972819, 0.12473418799891078, 0.12253731000055268, 0.12208043899954646, 0.12229661099991063, 0.12081742800000939, 0.12244083400037198, 0.12130962900118902, 0.12166742699992028, 0.12284696899951086, 0.12353334799990989, 0.12317632600024808, 0.12284574499972223, 0.12194762400031323, 0.12681236599928525, 0.1217876940008864, 0.12304871199921763, 0.12332478099961008, 0.12357730999974592, 0.12361225800123066, 0.12362794799992116, 0.12391733699951146, 0.12338858099974459, 0.1250736149995646, 0.12779323300128453, 0.12408751899965864, 0.12436877800064394, 0.12421653299861646, 0.12282391200096754, 0.12528303900035098, 0.12264522799887345, 0.12434737600051449, 0.12283714600016538, 0.12232946800031641, 0.12818033599978662, 0.1251672630005487, 0.12591900999905192, 0.1228797059993667, 0.12291983000068285, 0.1243465900006413, 0.12625833199854242, 0.1217715370003134, 0.12351171200134559, 0.1246158319991082, 0.1236307939998369, 0.12363569300032395, 0.12282523300018511, 0.1230602910000016, 0.12472399899888842, 0.12818272200092906, 0.12419339400003082, 0.12402869799916516, 0.12229217900130607, 0.123374277000039, 0.12494709999918996, 0.12278291499933403, 0.1253592260000005, 0.12344791400028043, 0.12322091900023224, 0.12378106500000285, 0.1251412750007148, 0.12490194000019983, 0.1232101400000829, 0.12712320999889926, 0.12422276700090151, 0.1279232619999675, 0.12324338699909276, 0.12436909900134197, 0.12222563099930994, 0.1217036799989728, 0.12383402300110902, 0.1234395699993911, 0.12398151500019594, 0.12317862900090404, 0.12311675799901423, 0.12257262099956279, 0.12260332700134313, 0.12305360299978929, 0.12946430400006648, 0.12391273499997624, 0.12351797799965425, 0.12502329000017198, 0.12295257699952344, 0.12345348700000613, 0.12201160800032085, 0.1217710500004614, 0.12314249199880578, 0.12546154499978002, 0.12302507700042042, 0.12450879900097789, 0.12602840899853618, 0.1260271470000589, 0.12383794100060186, 0.12591324400091253, 0.12299257299855526, 0.12200998800108209, 0.1236051089999819, 0.12476201099889295, 0.12797184600094624, 0.12909498099907069, 0.1232114909998927, 0.1244873590003408, 0.1229470720008976, 0.12368237699956808, 0.12326939999911701, 0.12361786000110442, 0.12352930399902107, 0.12425784600054612, 0.1241120470003807, 0.12599810900064767, 0.1268695740000112, 0.12445772499995655, 0.12704272599876276, 0.12239045100068324, 0.12220809199970972, 0.12279914400096459, 0.12208551999901829, 0.12069282000084058, 0.12156190899986541, 0.12547799599997234, 0.12560537100034708, 0.1264854799992463, 0.12332343799971568, 0.12291404400093597, 0.12289382299968565, 0.12316225399990799, 0.12204260900034569, 0.12272805999964476, 0.1232973120004317, 0.12317515099857701, 0.12298652100071195, 0.12411187299949233, 0.12491493800007447, 0.12416614300127549, 0.12492857899997034, 0.12306169200019212, 0.12240920399926836, 0.12403837000056228, 0.12502665799911483, 0.12238907500068308, 0.12253607499951613, 0.1213196059998154, 0.12630012900081056, 0.12542395499986014, 0.12978099999963888, 0.1250135810005304, 0.12434804999975313, 0.12372970799879113, 0.12441095100075472, 0.12324120300036157, 0.12496107399965695, 0.12533751400042092, 0.1271820879992447, 0.12789083200004825, 0.1263590460002888, 0.12768297299953701, 0.13065632500001811, 0.12868320100096753, 0.1261981770003331, 0.12474517299961008, 0.1284131209995394, 0.12612179899952025, 0.12805511100123113, 0.12713337899913313, 0.12417561499933072, 0.12442138600090402, 0.12351171299997077, 0.12290456399932737, 0.12432569700104068, 0.12486228600027971, 0.12338548799925775, 0.12236410100013018, 0.12629617099992174, 0.12385649099996954, 0.1252392759997747, 0.12863360899973486, 0.1262204680006107, 0.1235330569998041, 0.12334526400081813, 0.12473519199920702, 0.12247713299984753, 0.1230668690004677, 0.1235451299999113, 0.12278192799931276, 0.12470884400136129, 0.12332358799903886, 0.12340145900088828, 0.12537554499976977, 0.12427968399970268, 0.12268495699936466, 0.12407083900143334, 0.12408484399929876, 0.12388888200075598, 0.1255451519991766, 0.12317455300035363, 0.12307567700008804, 0.12664677999964624, 0.12262448399997083, 0.12307781799972872, 0.1243783270001586, 0.12324036599966348, 0.12294018400098139, 0.1229925709994859, 0.12343474599947513, 0.12417287199968996, 0.13084410600094998, 0.12801437999951304, 0.13158001699957822, 0.1266145770005096, 0.12454193099983968, 0.12444717899961688, 0.12167717900047137, 0.12303826700008358, 0.12367778599946178, 0.12662271100089129, 0.12720666200038977, 0.1289300190001086, 0.12458964100005687, 0.12373422099881282, 0.12420569800087833, 0.12360078900019289, 0.1221278720004193, 0.12370458399891504, 0.1239212350010348, 0.12388101799842843, 0.12702723600159516, 0.12452361999930872, 0.12499242900048557, 0.12381833899962658, 0.12410019599883526, 0.12173825400168425, 0.12344342999858782, 0.12272454500089225, 0.12301598299927718, 0.12514534700130753, 0.12389623199851485, 0.12395416500112333, 0.12337778899927798]",tokens/s,8.04436879700358,, +float16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cpu,,42,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 57, in launch + raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") +RuntimeError: Isolated process exited with non-zero code -9 +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +bfloat16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cpu,,42,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 57, in launch + raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") +RuntimeError: Isolated process exited with non-zero code -9 +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +float16-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cpu,,42,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,12558.036992,,,,,s,1,193.8872097809999,193.8872097809999,0.0,193.8872097809999,193.8872097809999,193.8872097809999,193.8872097809999,[193.8872097809999],,kWh,0.0022609050236562615,0.0012322798463658018,0,0.0034931848700220634,,MB,12691.718144,,,,,s,10,16.738576115998512,1.6738576115998511,0.010800180471687328,1.6720147619998897,1.6916467256998657,1.6929623938495753,1.6940149283693426,"[1.6678740850002214, 1.678766489998452, 1.672741901000336, 1.6736433009991742, 1.6712876229994436, 1.6655557890007913, 1.6627640940005222, 1.6942780619992845, 1.6913543549999304, 1.6603104160003568]",tokens/s,152.94012957011233,kWh,1.9505171417918316e-05,1.0630876048241694e-05,0.0,3.0136047466160008e-05,tokens/kWh,8494810.087071449,MB,12692.533248,,,,,s,10,98.76354555999933,9.876354555999933,0.028002269870754617,9.873164013999485,9.906572571100877,9.913453593551003,9.918958411511102,"[9.896818301000167, 9.920334616001128, 9.843533854998896, 9.865826811999796, 9.852913845999865, 9.902921904000323, 9.87743615799991, 9.90504345500085, 9.829824742999335, 9.86889186999906]",tokens/s,6.378871844138807,kWh,0.00011694550305638724,6.373986501091205e-05,0.0,0.00018068536806729928,tokens/kWh,348672.39485897165,,s,630,98.75912739899832,0.1567605196809497,0.0018994161394678324,0.15647142499983602,0.15934302229961758,0.16020810060072108,0.16220102290959404,"[0.15795900100056315, 0.15518204199906904, 0.15601076200073294, 0.16246675299953495, 0.15492045700011658, 0.15567315999942366, 0.1580178040003375, 0.16222637499959092, 0.15615350300140562, 0.15589747000012721, 0.15926959999887913, 0.15830722200007585, 0.156210930999805, 0.1620244929999899, 0.16006132000075013, 0.156063135000295, 0.1552632069997344, 0.15561418399920512, 0.15557974300099886, 0.15598960199895373, 0.15482555400012643, 0.15739377600039006, 0.15693739999915124, 0.16122408800038102, 0.1567611990012665, 0.15487658999882115, 0.1543411250004283, 0.1541997509993962, 0.1540051370011497, 0.15331073499874037, 0.15719891900153016, 0.15645142499852227, 0.15548379900064901, 0.15620615000079852, 0.15635108599963132, 0.15555461199983256, 0.15504510999926424, 0.1555596960006369, 0.15844665099939448, 0.15510152299975744, 0.1578743170011876, 0.15730231700035802, 0.15658428499955335, 0.15773438999895006, 0.15661465800076257, 0.15774520599916286, 0.16025232300125936, 0.1563573909988918, 0.16208515900143539, 0.15827408999939507, 0.15891673199985235, 0.15656332099933934, 0.15831986400007736, 0.1579144580009597, 0.15658595300010347, 0.1580141239992372, 0.15726282600007835, 0.1555645309999818, 0.1563775569993595, 0.15602944400052365, 0.15877720400021644, 0.15750452800057246, 0.15956727999946452, 0.1639513980007905, 0.15480943500006106, 0.1543118419995153, 0.15385829899969394, 0.15361888900042686, 0.15759961399999156, 0.15682254800049122, 0.1583336789990426, 0.15437387200108788, 0.15366502899996703, 0.1563723540002684, 0.15667197199945804, 0.15562148400022124, 0.1570068179989903, 0.15750992800167296, 0.1565422599996964, 0.1601540510000632, 0.15752190599960159, 0.15747339299923624, 0.15606783499970334, 0.15782750200014561, 0.16113093800049683, 0.15928753099979076, 0.16169982999963395, 0.15779414800090308, 0.1594840470006602, 0.16002597799888463, 0.16050326699951256, 0.15956262000145216, 0.1573714549995202, 0.1560825840006146, 0.1567035969983408, 0.15701297700070427, 0.15698752500065893, 0.15487595999911719, 0.15941183399991132, 0.15870551300031366, 0.160061324999333, 0.15859952800019528, 0.15781566800069413, 0.15673715499906393, 0.15915095700074744, 0.1578193219993409, 0.1594535670010373, 0.1600999629990838, 0.15508984200096165, 0.15472834799948032, 0.15701575499952014, 0.15631201799988048, 0.15802116400118393, 0.15908444399974542, 0.16157667399966158, 0.1595429780008999, 0.15848777899918787, 0.15569931199934217, 0.15475413100102742, 0.15500049500042223, 0.15663583199966524, 0.1554143789999216, 0.15562998600034916, 0.15605519499877119, 0.1575236450007651, 0.15684048699949926, 0.15910022299976845, 0.15786260700042476, 0.1573759639995842, 0.15662648899888154, 0.15572495300148148, 0.1552776399985305, 0.15604687300037767, 0.15563814000051934, 0.15668294500028423, 0.15619167200020456, 0.1558727679985168, 0.15548908700111497, 0.1535860109997884, 0.1544214400000783, 0.154137682999135, 0.15514183200139087, 0.15434232000006887, 0.15547144099946308, 0.15423493199887162, 0.15436759700060065, 0.15504173100089247, 0.1594013569992967, 0.15848900899982254, 0.15637891100050183, 0.1574788769994484, 0.16011710500060872, 0.1554418159994384, 0.15663020300053176, 0.15548402199965494, 0.15638541500084102, 0.1548730929989688, 0.1568467380002403, 0.15627633499934745, 0.1550722179999866, 0.1558002170004329, 0.15497850199972163, 0.1567187330001616, 0.15676738299953286, 0.1557361610011867, 0.15605534299902502, 0.1576220770002692, 0.15791451500081166, 0.1601093299996137, 0.1554078600001958, 0.15671634699901915, 0.15477998800088244, 0.15553805399940757, 0.15800811400004022, 0.1573175789999368, 0.15364887600117072, 0.15378997799962235, 0.15480690900039917, 0.15535562900004152, 0.15948849099913787, 0.15821693200086884, 0.1556868879997637, 0.1559820780003065, 0.1555149639989395, 0.1557149340005708, 0.15871503499874962, 0.157635878000292, 0.15547166699980153, 0.15609443100038334, 0.15907820699976583, 0.15469351399951847, 0.15487796800152864, 0.15591218800000206, 0.15576863399837748, 0.15531146500143223, 0.15620949599906453, 0.15397861400015245, 0.15750933600065764, 0.153685162000329, 0.15442128399990906, 0.15564294199975848, 0.15543484199952218, 0.1552020220005943, 0.15644568499919842, 0.15689235600075335, 0.16007464599897503, 0.15753498999947624, 0.15890997300084564, 0.15556177600046794, 0.1575890080002864, 0.15707023499999195, 0.15667770199979714, 0.15430769799968402, 0.15481072200054768, 0.15560006099985912, 0.15514868699938233, 0.1588204420004331, 0.15696043899879442, 0.15508886499992514, 0.15583502400113503, 0.16112631599935412, 0.16069720800078358, 0.1575199219987553, 0.1549210060002224, 0.1560464320009487, 0.15551838200008206, 0.15526442999907886, 0.1596265210009733, 0.1576803400002973, 0.15783636000014667, 0.1579145379982947, 0.15752054700169538, 0.15616721999867877, 0.15622325900039868, 0.15602217100058624, 0.15662694899947383, 0.15576873799909663, 0.15608619400154566, 0.15600316599920916, 0.1555801070007874, 0.15670609900007548, 0.15433857799871475, 0.15782555999976466, 0.15676750100101344, 0.1571127709994471, 0.15771672400114767, 0.15560009799992258, 0.15698759599945333, 0.15771201700044912, 0.15901316599956772, 0.15757285399922694, 0.15680030499970599, 0.15590374800012796, 0.15564260599967383, 0.15338205400075822, 0.15425157599929662, 0.15434668899979442, 0.15393217400014692, 0.155847428000925, 0.1531633199992939, 0.1563863859992125, 0.15542337400074757, 0.15840587499951653, 0.15816902500046126, 0.15624789200046507, 0.15888679199997569, 0.1568453159998171, 0.15612319600040792, 0.1598385059987777, 0.1574149160005618, 0.1576669120004226, 0.1580220569994708, 0.15704508300041198, 0.15985643999920285, 0.1605350770005316, 0.15951465699981782, 0.1582260060004046, 0.15502952600036224, 0.15717668300021614, 0.15712983799858193, 0.1572207110002637, 0.1553193480012851, 0.15927448499860475, 0.15712624000116193, 0.15640639299999748, 0.1572960539997439, 0.15668202199958614, 0.15522803999920143, 0.1548618520009768, 0.15496997399895918, 0.15868643700014218, 0.1572197680015961, 0.15804270299850032, 0.1584959150004579, 0.15568576000077883, 0.15617349999956787, 0.15468458400027885, 0.1555951209993509, 0.15549419099988881, 0.15500315600002068, 0.15511488999982248, 0.15486397400127316, 0.15926673699868843, 0.15421600100125943, 0.1557230199996411, 0.15569314300046244, 0.1540897609993408, 0.15688247199977923, 0.15494896699965466, 0.15457574700121768, 0.15413264899871137, 0.15531758700126375, 0.15471523299856926, 0.15579694400003063, 0.15723953599990637, 0.15765702099997725, 0.15451210900027945, 0.15493510299893387, 0.15511033600159863, 0.15454736399988178, 0.15513198400003603, 0.1561167439995188, 0.15689967300022545, 0.15491143399958673, 0.15730441800042172, 0.15767367300031765, 0.1565920239991101, 0.1636731510006939, 0.1608602599990263, 0.1565669350002281, 0.15688269400015997, 0.15827463200002967, 0.16199433300062083, 0.1569490419988142, 0.15940132600007928, 0.15745336300096824, 0.1566463079998357, 0.1546403789998294, 0.15502533499966376, 0.15684039400002803, 0.15780391800035432, 0.15837306799949147, 0.15685554000083357, 0.15659624799991434, 0.15784886300025391, 0.15665615999932925, 0.15552477699930023, 0.1553368570002931, 0.1555703160011035, 0.16188090600007854, 0.16693672200017318, 0.15785419799976808, 0.15610195099907287, 0.157562592999966, 0.1595274490009615, 0.16127779899943562, 0.1569513159993221, 0.15593587900002603, 0.15609796200078563, 0.15472544700060098, 0.15579732799960766, 0.15583440699992934, 0.15701682299913955, 0.15692557799957285, 0.15561855100168032, 0.1566281909999816, 0.15516279599978589, 0.1567651859986654, 0.15737931500007107, 0.15750530900004378, 0.15629134800110478, 0.15643243699923914, 0.15602363600009994, 0.1571736999994755, 0.15705595400140737, 0.16033232900008443, 0.1563315949988464, 0.1561845460000768, 0.15937354400011827, 0.15778783299901988, 0.15920792600081768, 0.16013769799974398, 0.1577872739999293, 0.15590711699951498, 0.15577171800032374, 0.15463665100105572, 0.1576257809992967, 0.15634908100037137, 0.15516026499972213, 0.1545309859993722, 0.1559220990002359, 0.15535357600128918, 0.15677954499915359, 0.1564978250007698, 0.15485977299977094, 0.1551610789992992, 0.1556794660009473, 0.15768069799923978, 0.15564972299944202, 0.1588260460011952, 0.15448697799911315, 0.1556021830001555, 0.15505010600099922, 0.15430589099923964, 0.15540327800044906, 0.15592521299913642, 0.1556755540004815, 0.15673268899990944, 0.15529094500016072, 0.1577260270005354, 0.15663608699833276, 0.15761738900073397, 0.15601889000026858, 0.16005675599990354, 0.15398116600044887, 0.1561606129998836, 0.15546098099912342, 0.15566687300088233, 0.15568288400027086, 0.1593546799995238, 0.15878027600047062, 0.15640728499965917, 0.15889942499961762, 0.1569064619998244, 0.1582187639996846, 0.15653996100081713, 0.15903344599973934, 0.15683174799960398, 0.15974693700081843, 0.15711770899906696, 0.15708876800090366, 0.15535112199904688, 0.15648263900038728, 0.1592591419994278, 0.16067250500054797, 0.1594255889995111, 0.15680376199998136, 0.15571580300093046, 0.15579034199981834, 0.15694362800059025, 0.15547074599999178, 0.15653180300068925, 0.15526074599983986, 0.15443217899883166, 0.15563506400030747, 0.15790016200116952, 0.15785092699843517, 0.15665029299998423, 0.1561482330016588, 0.16196572199987713, 0.15728452399889647, 0.16059705699990445, 0.15847546400073043, 0.15530235799997172, 0.1545616709991009, 0.15681026200036285, 0.15788729900123144, 0.1586712649987021, 0.15690036300111387, 0.15748614399853977, 0.1555010760002915, 0.15629329300099926, 0.15712518799955433, 0.1589652799993928, 0.15622388600058912, 0.16093831199941633, 0.1575262530004693, 0.15471337099916127, 0.15729933800139406, 0.1560662579995551, 0.1566152659997897, 0.155267456000729, 0.15601335799874505, 0.15825703400150815, 0.15798856599940336, 0.1566420899998775, 0.15456688700032828, 0.15671563899923058, 0.15875938499993936, 0.15683309900123277, 0.15536564399917552, 0.15960974900008296, 0.1576158989992109, 0.15765464600008272, 0.15498242400099116, 0.1564327839987527, 0.15675824500067392, 0.15772441599983722, 0.16015053000046464, 0.15662125199924049, 0.1589835150007275, 0.15789667299941357, 0.1574019199997565, 0.15709945900016464, 0.156212862000757, 0.1563952420001442, 0.15626652199898672, 0.16040126100051566, 0.16213895399960165, 0.156860315000813, 0.15712649200031592, 0.156256393998774, 0.1557203730008041, 0.15825268700064044, 0.1577300649987592, 0.1572830220011383, 0.1556455609988916, 0.1548227390012471, 0.1541378559995792, 0.15773717599950032, 0.1604498489996331, 0.1564550479997706, 0.1561112500003219, 0.1550077539995982, 0.15610009100055322, 0.15586472600080015, 0.15595972899973276, 0.15518695900027524, 0.1558315969996329, 0.15609394000057364, 0.153579963998709, 0.15491449600085616, 0.15778041100020346, 0.1543497849997948, 0.15608439899915538, 0.15616205100013758, 0.1572821740010113, 0.15678302499873098, 0.15798363200156018, 0.15773052299846313, 0.15636951700071222, 0.15482019699993543, 0.1554953859995294, 0.15818968799976574, 0.158802938000008, 0.1568371620014659, 0.15468814199994085, 0.15479950299959455, 0.15408816099989053, 0.1547404289995029, 0.1541700790003233, 0.15492282800005341, 0.1551151990006474, 0.15451013399979274, 0.1555311129995971, 0.15784686199913267, 0.15710869600115984, 0.15479942799902346, 0.155425187000219, 0.15591528299955826, 0.15861064300042926, 0.1573004479996598, 0.15474998700119613, 0.15377685199928237, 0.15640266199989128, 0.15708058800009894, 0.15644297700055176, 0.15528045800056134, 0.15539734699996188, 0.15655478199914796, 0.15654094300043653, 0.1545860209989769, 0.1541649230002804, 0.15506351600015478, 0.15568447400073637, 0.15644307099864818, 0.15806543800135842, 0.15420808799899532, 0.15854180099995574, 0.15529804499965394, 0.1555270630015002, 0.15798119600003702, 0.15646021099928475, 0.15826940800070588, 0.15469428500000504, 0.15471366599922476, 0.15560845599975437, 0.1549919860008231, 0.15500604899898462, 0.15460981599972, 0.1543147440006578, 0.15474354499929177, 0.15592305600148393, 0.15665453699875798, 0.1577566330015543, 0.1558380339993164, 0.15673678299936, 0.15919085099994845, 0.15621072300018568, 0.15812194799946155, 0.15675970100164704, 0.15688032999969437, 0.16048637299900292, 0.1595788779995928, 0.15678189700156508, 0.15486218499972892, 0.15778367999882903, 0.15785009100000025, 0.1562819570008287, 0.1571875129993714, 0.16476018000139447, 0.16377128599924617, 0.15600799399908283, 0.16101810400141403, 0.1588249039996299, 0.15583367100043688, 0.1574137339994195, 0.159341726999628, 0.1550659140011703, 0.15330401499886648, 0.15434032600023784, 0.1577323740002612, 0.15361272799964354, 0.153891656000269, 0.15565507999963302, 0.15555648900044616, 0.15489478100062115, 0.15515020599923446, 0.15455536000081338, 0.1567941549983516, 0.15549837500111607, 0.15631236799890758, 0.15558753900040756, 0.157630269000947, 0.15821730799871148, 0.155682140000863, 0.15815060299973993, 0.15731759600021178, 0.15522917099951883, 0.1554537510000955]",tokens/s,6.379157214043893,, +float32-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cpu,,42,,,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,24690.429952,,,,,s,1,63.583938114001285,63.583938114001285,0.0,63.583938114001285,63.583938114001285,63.583938114001285,63.583938114001285,[63.583938114001285],,kWh,0.0007224275369451511,0.00039374964215991274,0,0.0011161771791050638,,MB,25400.631296,,,,,s,10,15.017791419002606,1.5017791419002606,0.004572127243590556,1.4999615260003338,1.5081687393008907,1.5089704786512812,1.5096118701315937,"[1.5053218369994283, 1.5079905750008038, 1.5046895559989935, 1.4955332270001236, 1.4982708170009573, 1.500137867000376, 1.4989029230000597, 1.5097722180016717, 1.4973872139999003, 1.4997851850002917]",tokens/s,170.46447966781125,kWh,1.7700390187221478e-05,9.647228440849081e-06,0.0,2.7347618628070558e-05,tokens/kWh,9360961.31372962,MB,25403.035648,,,,,s,10,100.39095832100065,10.039095832100065,0.0452384640858615,10.033959183499974,10.094620076999309,10.103024668500348,10.109748341701179,"[10.056399371000225, 10.0259592099992, 10.082903530001204, 10.111429260001387, 10.028357924999, 9.991818063999744, 10.092752389999077, 10.001336266999715, 10.039560442000948, 9.960441862000152]",tokens/s,6.275465545269241,kWh,0.00011864792251409755,6.466773663223826e-05,0.0,0.00018331565914633582,tokens/kWh,343669.4949759248,,s,630,100.38703922599962,0.1593445067079359,0.00319573688375865,0.15884037949945196,0.16378042539981835,0.1656503564004197,0.16893867226935985,"[0.16394315400066262, 0.15955996599950595, 0.159505915000409, 0.15879419500015501, 0.1612904140001774, 0.16139141399980872, 0.15591793699968548, 0.15912588399987726, 0.1585221460009052, 0.15647293099937087, 0.15741081300075166, 0.15712456699839095, 0.15533080300156143, 0.16315009999925678, 0.1572862819994043, 0.15980783499981044, 0.15590837800118607, 0.15609263499936787, 0.1573608670005342, 0.15811314799975662, 0.15459580599963374, 0.15765573700082314, 0.15507199799867522, 0.15543713600163755, 0.1567320269987249, 0.16015537100065558, 0.15355373099919234, 0.15612580900051398, 0.1581423999996332, 0.15806769000118948, 0.1557832620001136, 0.1558481859992753, 0.1633138239994878, 0.17061631900105567, 0.15968082200015488, 0.15671740399920964, 0.1574088319994189, 0.1595265380001365, 0.15767900700120663, 0.15939525099929597, 0.1593988309996348, 0.1649805320012092, 0.16053653999915696, 0.1612194659992383, 0.1597686690001865, 0.1586297169997124, 0.16052009900158737, 0.16280451599959633, 0.15882327500003157, 0.16114091300005384, 0.16320794499915792, 0.16671795500042208, 0.16026617399984389, 0.15976979100014432, 0.16112065400011488, 0.16085010599999805, 0.1627280350003275, 0.1601276869987487, 0.1663190620001842, 0.1616988969999511, 0.16288257999985944, 0.16552896600114764, 0.16332582399991225, 0.16267713500019454, 0.161238486000002, 0.15775690999907965, 0.15770669700032158, 0.1643967390009493, 0.1594601929991768, 0.15572106699983124, 0.15709274700020615, 0.15619420099937997, 0.1585043190007127, 0.15794897100022354, 0.15913671499947668, 0.15584168500026863, 0.15639001199997438, 0.15665125799932866, 0.15810327800136292, 0.15889247299855924, 0.1593256760006625, 0.15783661599925836, 0.15632026300045254, 0.16101241599972127, 0.16093098500095948, 0.15822043799926178, 0.15671976500016171, 0.1565931189998082, 0.15927840800031845, 0.16000945000087086, 0.15685167199990246, 0.15946801099926233, 0.1597877439999138, 0.1605401470005745, 0.15816622899910726, 0.15905518399995344, 0.15610187800120912, 0.15447827899879485, 0.15421010600039153, 0.16664092800056096, 0.1666736730003322, 0.1552887899997586, 0.15784763299961924, 0.16049157899942657, 0.16087649300061457, 0.16018147000067984, 0.1587506879986904, 0.15665972900023917, 0.15693691700107593, 0.15468793099898903, 0.15894947500055423, 0.16424442199968325, 0.16003956900021876, 0.16066469900033553, 0.15692860099989048, 0.16361803799918562, 0.16002927500085207, 0.1645622979995096, 0.16108062400053313, 0.16127662699909706, 0.16139830599968263, 0.16137103800065233, 0.1604890399994474, 0.15840542899968568, 0.1587666410014208, 0.16008884300026693, 0.1647226880013477, 0.16242022899859876, 0.16096062900032848, 0.15948120200118865, 0.16294895399914822, 0.15969711700017797, 0.1671780779997789, 0.16370077800092986, 0.16143658900000446, 0.15983522099850234, 0.15955092800140847, 0.1626963529997738, 0.155839211998682, 0.15624275300069712, 0.16015418300048623, 0.16153148999910627, 0.15873224599999958, 0.15648360200066236, 0.1560569659995963, 0.15736058500078798, 0.1555011469990859, 0.1556382950002444, 0.16072146599981352, 0.15836623900031555, 0.15809157299918297, 0.15757414800100378, 0.16361723199952394, 0.15845111999988148, 0.15625131500019052, 0.15462835899961647, 0.15613302400015527, 0.15643159499995818, 0.15724931300064782, 0.1570263409994368, 0.15834507900035533, 0.15848484200068924, 0.15630494200013345, 0.15679290399930323, 0.16241761799938104, 0.1611784680007986, 0.1669943830002012, 0.1668584599992755, 0.1668079270002636, 0.16473027000029106, 0.16921774999900663, 0.16765572100121062, 0.16236383999967074, 0.1589029599999776, 0.16242305800005852, 0.158830272999694, 0.16137202599929878, 0.15541500100152916, 0.15903956399961316, 0.15584157800003595, 0.1594597959992825, 0.15600473599988618, 0.15718436000133806, 0.1610854409991589, 0.16736994399980176, 0.16257949099963298, 0.15836986300018907, 0.15801062700120383, 0.1597758089992567, 0.1616598630007502, 0.1596179320004012, 0.15741967699977977, 0.1565235660000326, 0.1558889679999993, 0.1578256060001877, 0.15825536000011198, 0.1551393879999523, 0.15474234300017997, 0.15607041199837113, 0.1586879480000789, 0.15714451800158713, 0.1580354029993032, 0.15660807200038107, 0.16118036599982588, 0.15608623499974783, 0.16554522499973245, 0.15925303000040003, 0.16294586299954972, 0.16039476500009187, 0.15675303299940424, 0.15570398999989266, 0.15597775500100397, 0.1601491530000203, 0.16155669999898237, 0.15887579600166646, 0.16043655999965267, 0.15797668999948655, 0.15959069800010184, 0.1605532100002165, 0.16070177599976887, 0.15694384400012495, 0.16401173399935942, 0.17044223400080227, 0.16138441099974443, 0.1602852050000365, 0.16225886999927752, 0.16071282099983364, 0.1662693380003475, 0.16202143500049715, 0.16034459299953596, 0.1566997949994402, 0.15742043099999137, 0.15808958900015568, 0.16398323700013862, 0.15938249500140955, 0.15780717599955096, 0.1620455479987868, 0.16573637300098198, 0.166302583000288, 0.16424225899936573, 0.1632707949993346, 0.16225008200126467, 0.16981193900028302, 0.17075948599995172, 0.16599114499877032, 0.16160037700137764, 0.15851787399878958, 0.1611370070004341, 0.16247869499966328, 0.15978678899955412, 0.15753840700017463, 0.16422804500143684, 0.16468776799956686, 0.16242675299872644, 0.1606107220013655, 0.15791308299958473, 0.15625873600038176, 0.15889511800014589, 0.161029533999681, 0.15818714799934241, 0.1576092899995274, 0.1543839250007295, 0.16260744700048235, 0.16138339799908863, 0.1550337890002993, 0.15686310499950196, 0.1583632010006113, 0.15488369800004875, 0.1568598039993958, 0.15931814800023858, 0.15663476299960166, 0.1589800199999445, 0.15960482700029388, 0.15809820000140462, 0.16778331799832813, 0.16345918500155676, 0.15842356299981475, 0.15794832199935627, 0.15789779700025974, 0.16049189300065336, 0.16295591299967782, 0.16032852500029549, 0.16211543699864706, 0.16959167500135663, 0.16246203299851913, 0.1647530200007168, 0.15817607399912958, 0.1602958880012011, 0.1575732050005172, 0.16137436899953173, 0.16287755699886475, 0.16416225300054066, 0.15997674199934409, 0.15625905100023374, 0.15801337900120416, 0.16276656199988793, 0.15901108399884833, 0.15855736100093054, 0.15954935899935663, 0.15624705600021116, 0.15853543299999728, 0.16118956399986928, 0.15517641800033743, 0.15600337899923034, 0.15684508699996513, 0.15616568200130132, 0.15475219899963122, 0.15602694699919084, 0.1561992900005862, 0.15688097599922912, 0.15641174600023078, 0.15807040199979383, 0.1573562330013374, 0.15885048599920992, 0.15581158399982087, 0.16402650300005917, 0.16103960699911113, 0.16240498800107162, 0.15925603799951205, 0.1579026330000488, 0.15768171600029746, 0.15752186499958043, 0.15570596900033706, 0.1582379730007233, 0.15781917899948894, 0.15496336499927565, 0.15605550500004028, 0.16219200200066552, 0.1637491820001742, 0.15786415799993847, 0.159825226999601, 0.1611239529993327, 0.16224383700136968, 0.1566659919990343, 0.15847435999967274, 0.15880415300125605, 0.15698632199928397, 0.15693110400025034, 0.15798576599991065, 0.1562355949990888, 0.15675042800103256, 0.15480297099929885, 0.15711704600107623, 0.1572046089986543, 0.1595334970006661, 0.16387508099978731, 0.16290367700094066, 0.15802287899896328, 0.15670142900125938, 0.16096897199895466, 0.15690333400016243, 0.15875645099913527, 0.1575905360004981, 0.15909382900099445, 0.1563791640001, 0.15573426199989626, 0.15702965600030439, 0.1563252249998186, 0.1555213960000401, 0.1540845250001439, 0.1550098319985409, 0.15458344900071097, 0.1593381889997545, 0.1593478850008978, 0.15679913199892326, 0.16232863100049144, 0.15612885799964715, 0.15892451699983212, 0.15995886600103404, 0.16222062499946333, 0.15834936799910793, 0.1574715370006743, 0.15727674799927627, 0.157024927000748, 0.16736256799958937, 0.15832101900014095, 0.16492582499995478, 0.15904534699984652, 0.16629593899961037, 0.1652714609990653, 0.15789047900034348, 0.1630478299994138, 0.16637180200086732, 0.15917490900028497, 0.1601302589988336, 0.15834271400126454, 0.1598973849995673, 0.15919107800073107, 0.15608856399921933, 0.1560753379999369, 0.1643698919997405, 0.15892523900038213, 0.15654159800033085, 0.15504227099881973, 0.1556477240010281, 0.1588598809994437, 0.16138897299970267, 0.15765176300010353, 0.1580667779999203, 0.16113852099988435, 0.1561578010005178, 0.16043500200066774, 0.15938030799952685, 0.1569582789998094, 0.1551409710009466, 0.15767678699921817, 0.16177967700059526, 0.15675325699885434, 0.16036293000070145, 0.15911534699989716, 0.16110357199977443, 0.1682554130002245, 0.16295543399974122, 0.15923397299957287, 0.1600561530012783, 0.16158219699900656, 0.15890677199968195, 0.16019480300019495, 0.15859166300106153, 0.16167608499927155, 0.16377183999975387, 0.1581206890004978, 0.159404337999149, 0.1628524190000462, 0.15927870199993777, 0.15896687700114853, 0.1609822049995273, 0.16539590899992618, 0.15813766099927307, 0.1646184710007219, 0.1573874519999663, 0.1583251020001626, 0.15865206899979967, 0.15734501000042656, 0.1598303600003419, 0.16048378899904492, 0.16109345400036545, 0.16007407600045553, 0.16259364899997308, 0.16722502899938263, 0.16610293600024306, 0.16149889499865822, 0.15767004299959808, 0.15750180600116437, 0.15668933400047536, 0.1570525989991438, 0.15623077800046303, 0.15596025299964822, 0.1618972100004612, 0.15587565299938433, 0.1600545780001994, 0.16715951500009396, 0.16284186700067949, 0.16389708899987454, 0.15634625699931348, 0.15694745099972351, 0.16133274899948447, 0.1568649370001367, 0.1586237900010019, 0.1583584099989821, 0.15668298800119373, 0.1573619749997306, 0.1595889599993825, 0.1594301559998712, 0.16084283900090668, 0.1577608509996935, 0.15473747400028515, 0.16087584100023378, 0.15645863899953838, 0.15822165200006566, 0.15574522799943225, 0.15451094600030046, 0.1553050999991683, 0.1557265980009106, 0.15701645700028166, 0.15714668100008566, 0.1573255249986687, 0.15666942300049413, 0.15632402099981846, 0.15716276400053175, 0.1560196280006494, 0.16269054799886362, 0.15560677200119244, 0.1565644769998471, 0.16116945999965537, 0.1614564989995415, 0.15777568600060476, 0.1560734449994925, 0.15599501399992732, 0.1581703980009479, 0.16985746099999233, 0.16098195099948498, 0.15979493700069725, 0.16044333299942082, 0.15878206699926523, 0.1568922990009014, 0.1582645769994997, 0.1649946870002168, 0.16090277600051195, 0.1557172630000423, 0.16320022999934736, 0.16078830799960997, 0.15979673900073976, 0.16131337099977827, 0.16155825100031507, 0.16031810200001928, 0.1575094439995155, 0.15917304899994633, 0.16209785499995633, 0.1666756680006074, 0.16385769400039862, 0.15713257899915334, 0.1594803440002579, 0.1573552589998144, 0.15589746099976765, 0.15867746000003535, 0.1615345720001642, 0.15905228200062993, 0.1575143299996853, 0.16453937600090285, 0.1573840109995217, 0.1559578029991826, 0.15530562700041628, 0.15941840099912952, 0.16066756400141458, 0.15726259300026868, 0.1571175499993842, 0.159529227999883, 0.15965680199951748, 0.1583869460009737, 0.15911830099867075, 0.16133913600060623, 0.15647762099979445, 0.1598175449998962, 0.16637308200006373, 0.1605977030012582, 0.15901175099861575, 0.1563600099998439, 0.15466956500131346, 0.1562594219994935, 0.1598853350005811, 0.16029039899876807, 0.15493180900011794, 0.15755811500093841, 0.1558189229999698, 0.15813924800022505, 0.15772918999937247, 0.1558493440006714, 0.1567315799984499, 0.1592574140013312, 0.1669647629987594, 0.16532328900029825, 0.15765367899984994, 0.1594557389998954, 0.163705265000317, 0.16422050900109753, 0.15719091699975252, 0.15821497499928228, 0.15755928900034633, 0.1590650750003988, 0.15893653300008737, 0.1582513920002384, 0.15993411999988894, 0.15959777199896052, 0.16201138200085552, 0.15923446099986904, 0.162583919998724, 0.16166556999996828, 0.1617115820008621, 0.1636974579996604, 0.1591354930005764, 0.15578859400011424, 0.15598323799895297, 0.15412720600033936, 0.15680634300042584, 0.15710928099906596, 0.15511343400066835, 0.15427948199976527, 0.15770607000013115, 0.154352839999774, 0.1563325499992061, 0.15674137200039695, 0.16058875599992462, 0.16280190700126695, 0.1589056349985185, 0.16090521000114677, 0.15517061299942725, 0.15701378400081012, 0.15978624099989247, 0.15868943999885232, 0.1567464990002918, 0.15625222500057134, 0.15749904699987383, 0.153820043999076, 0.15921522400094545, 0.16356029499911529, 0.15651115200125787, 0.15650190299857059, 0.1548378550014604, 0.15591878299892414, 0.15430013200057147, 0.15528405499935616, 0.1559528350007895, 0.1622698480005056, 0.16534640600002604, 0.1607035269989865, 0.15489278700079012, 0.1579185039990989, 0.15866073700090055, 0.16296640499967907, 0.1639980079999077, 0.16056132499943487, 0.16091861099994276, 0.16012514700014435, 0.15732748600021296, 0.1575460669992026, 0.1557996980009193, 0.15670992500054126, 0.15627564500027802, 0.1563916289996996, 0.15641719799896237, 0.1609796640004788, 0.15656502199999522, 0.15825719799977378, 0.15933852700072748, 0.15800551000029373, 0.1588067639986548, 0.1582174249997479, 0.15640816500126675, 0.1578180929991504]",tokens/s,6.27571053850579,, +bfloat16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-34B,01-ai/Yi-34B,cpu,,42,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): + File ""/workspace/llm_perf/common/benchmark_runner.py"", line 111, in execute_and_log_benchmark + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 57, in launch + raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") +RuntimeError: Isolated process exited with non-zero code -9 +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +bfloat16-sdpa,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,llama,01-ai/Yi-6B,01-ai/Yi-6B,cpu,,42,,,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,,MB,12557.336576,,,,,s,1,160.62633338100022,160.62633338100022,0.0,160.62633338100022,160.62633338100022,160.62633338100022,160.62633338100022,[160.62633338100022],,kWh,0.0018660494217909687,0.0010170680129037454,0,0.0028831174346947144,,MB,12694.07744,,,,,s,10,20.679765909000707,2.067976590900071,0.012530350348158956,2.064916370500214,2.0832773158992497,2.090563481449499,2.0963924138896983,"[2.0556161110016546, 2.0570776450003905, 2.0648791450003046, 2.056279305001226, 2.0625295429999824, 2.081658167999194, 2.097849646999748, 2.065633540998533, 2.064953596000123, 2.073289207999551]",tokens/s,123.79250380613736,kWh,2.4435538898889413e-05,1.33181122855769e-05,0.0,3.775365118446632e-05,tokens/kWh,6780801.113756404,MB,12695.53152,,,,,s,10,76.14153990500017,7.614153990500017,0.03506812326714269,7.619351832500797,7.649846638900454,7.65383846695031,7.657031929390196,"[7.642027145999236, 7.648959566000485, 7.59521109599882, 7.548814944999322, 7.562598181000794, 7.657830295000167, 7.614479887000925, 7.624223778000669, 7.644094593999398, 7.603300417000355]",tokens/s,8.274064338415466,kWh,9.008497942215174e-05,4.909984729598886e-05,0.0,0.00013918482671814063,tokens/kWh,452635.5457378955,,s,630,76.1376211099996,0.12085336684126921,0.001771900464142561,0.1205072904995177,0.1231182550003723,0.12411805325018577,0.12653798199098673,"[0.1251336169989372, 0.12180716900002153, 0.12246660299933865, 0.12156948200026818, 0.12075947000084852, 0.12197730700063403, 0.123119460999078, 0.12044800900002883, 0.12094718300068052, 0.12045296899850655, 0.11939885400170169, 0.12283164199834573, 0.1216480560015043, 0.12382758199964883, 0.1217605599995295, 0.12134747200070706, 0.11883795999892754, 0.12012769600005413, 0.12068140300107189, 0.1209350720000657, 0.12038128499989398, 0.11924509699929331, 0.12042812800063984, 0.12279652699908183, 0.12319226700128638, 0.12032389299929491, 0.1184734889993706, 0.11984119699991425, 0.11809784199977003, 0.11842994000107865, 0.11885552800049481, 0.11799271699965175, 0.12054201299906708, 0.12153414900058124, 0.12665703799939365, 0.12308185100118862, 0.12268830899847671, 0.12124853200111829, 0.12208829399969545, 0.12357320599949162, 0.12444587900063198, 0.12159557399900223, 0.12053248700067343, 0.11985025400099403, 0.12297732299884956, 0.12204438300068432, 0.11938949300019885, 0.1199669769994216, 0.12141385600079957, 0.11997403999885137, 0.1215852929999528, 0.1203454900005454, 0.12153920600030688, 0.11999414799902297, 0.12178601900086505, 0.12539332899905276, 0.12416384800053493, 0.12165309400006663, 0.12051176599925384, 0.12092028900042351, 0.122147080999639, 0.1212553969999135, 0.11861394900006417, 0.1247536080009013, 0.1214900459999626, 0.12400468899977568, 0.11933612400025595, 0.11837444599950686, 0.12154759499935608, 0.12089811400073813, 0.11904075100028422, 0.1206544529995881, 0.11808438200023375, 0.11877996899966092, 0.11911482499999693, 0.11882682499890507, 0.11836843000128283, 0.11923214899979939, 0.11929121799948916, 0.12000263499976427, 0.11868002899973362, 0.1204396250013815, 0.12094343199896684, 0.1190905430012208, 0.1184052270000393, 0.12103741799910495, 0.1202435430004698, 0.12072176999936346, 0.12034206300086225, 0.126249385999472, 0.12407070799963549, 0.12014586199984478, 0.12216790700040292, 0.11995897000088007, 0.12075890499909292, 0.12149864000093658, 0.12254793099964445, 0.12106664499879116, 0.12078946600013296, 0.1214709430005314, 0.12399151700083166, 0.12034998799936147, 0.12062244799926702, 0.12161517000095046, 0.12304995599879476, 0.12311094500000763, 0.12382310400062124, 0.12295473700032744, 0.12197574299898406, 0.12236846600171702, 0.12271895799858612, 0.12105203800092568, 0.11990860900004918, 0.12027329800002917, 0.12299829999938083, 0.12126822900063416, 0.12252868699943065, 0.12210251199940103, 0.12317159700069169, 0.12289916699955938, 0.12290822599970852, 0.12656972800141375, 0.1303522130001511, 0.12118419299986272, 0.12109181599953445, 0.12125042399929953, 0.12575898699833488, 0.12134083500131965, 0.12330815499990422, 0.1195062759998109, 0.12047666400030721, 0.12047586500011676, 0.11990435799998522, 0.11988082300013048, 0.12134346999846457, 0.12025301100038632, 0.12323361100061447, 0.12140277400067134, 0.11936337599945546, 0.11941420800030755, 0.11978676699982316, 0.12047606499982066, 0.12075953599924105, 0.1205404410011397, 0.12050672999976086, 0.12097494800036657, 0.1190388669983804, 0.1188339790005557, 0.11978379799984396, 0.12200188499991782, 0.12247915199986892, 0.11878465400150162, 0.12008818999856885, 0.1186987630007934, 0.12071531499896082, 0.1212179100002686, 0.12194260100113752, 0.12000037799953134, 0.11909258800005773, 0.12012303200026508, 0.12012881699956779, 0.12034452499938197, 0.12123636400065152, 0.1214951399997517, 0.12030149900056131, 0.12338705500042124, 0.11952407599892467, 0.12029645900111063, 0.12061862099835707, 0.11889451700153586, 0.11888936200011813, 0.11987470999883953, 0.1192837190010323, 0.11967531600021175, 0.12091247299940733, 0.12009477599895035, 0.1211076489998959, 0.12064907300009509, 0.11895764600012626, 0.11921891000019968, 0.12034563800079923, 0.12025614300000598, 0.12061724699924525, 0.12028700799965009, 0.12034587700145494, 0.11997855699883075, 0.12364549999983865, 0.12042997600110539, 0.1225071469998511, 0.12345032199846173, 0.11992350300170074, 0.12038370699883671, 0.12136434199965151, 0.12013786799980153, 0.1199292780001997, 0.12015025900109322, 0.12213682000037807, 0.12223259699931077, 0.11897078299989516, 0.11929176000012376, 0.12113880200013227, 0.12066409100043529, 0.11964756099951046, 0.11976119300015853, 0.12013622999984364, 0.12023299899919948, 0.11989201200049138, 0.11975546500070777, 0.11791744999936782, 0.11895244299921615, 0.1189465650004422, 0.11855002400079684, 0.11893942999995488, 0.12009933700028341, 0.12012558199967316, 0.12041774299905228, 0.12000536200139322, 0.1198752609998337, 0.11893551999855845, 0.12017499600005976, 0.1174757460012188, 0.12094912199972896, 0.12013500300054147, 0.1190308429995639, 0.11895772199932253, 0.12029754700051853, 0.11950187399997958, 0.11977256499994837, 0.12116198600051575, 0.11874451699986821, 0.1202995869989536, 0.1224240080009622, 0.11838059100045939, 0.11847676999968826, 0.11918704799973057, 0.11920228999952087, 0.11738252100076352, 0.11969662899900868, 0.11967469100090966, 0.11956139100038854, 0.11847922099877906, 0.12366627800111019, 0.12024856700008968, 0.11995451899929321, 0.11822164099976362, 0.11982822500067414, 0.119730210999478, 0.11924323399944114, 0.1191520299998956, 0.11963006600126391, 0.1188878159991873, 0.11893320399940421, 0.12419263400079217, 0.11857760999919265, 0.11868924400005199, 0.11893477000012354, 0.11888918499971624, 0.11885231099950033, 0.11866389800161414, 0.11905639199903817, 0.11984157400002005, 0.11993854600041232, 0.1185609009990003, 0.11985314699995797, 0.1188989300007961, 0.12023186800070107, 0.1198789489990304, 0.11999691399978474, 0.12053600900071615, 0.11782271700030833, 0.11994241299908026, 0.11944662200039602, 0.11919495099937194, 0.12030345800121722, 0.1199412489986571, 0.1200684210016334, 0.1205627370000002, 0.1209196569998312, 0.11957930499920622, 0.11928339100086305, 0.11972728199907579, 0.11946637400069449, 0.11962678099916957, 0.11968789700040361, 0.12017170000035549, 0.12093598499996006, 0.12221972400038794, 0.12121623299935891, 0.12480403900008241, 0.11968066199915484, 0.11935064200042689, 0.12094861100013077, 0.1203577159994893, 0.12019910600065487, 0.11912654399930034, 0.11867102500036708, 0.12096656200083089, 0.12077302599936957, 0.1230889369999204, 0.11940896200030693, 0.12012686399975792, 0.1224996849996387, 0.12139295700035291, 0.12194652399921324, 0.12047854400043434, 0.11928620800063072, 0.12007073799941281, 0.11961186100052146, 0.1195060480004031, 0.12008180500015442, 0.11901872299858951, 0.11925038100162055, 0.11862554999970598, 0.1187019640001381, 0.1205248459991708, 0.12384638200092013, 0.12026216000049317, 0.11911564499860106, 0.11835924400111253, 0.11908655799925327, 0.12038058300095145, 0.11930953599949135, 0.12176596499921288, 0.12216387999978906, 0.12025076700047066, 0.12468937099947652, 0.1207693090000248, 0.12047633700058213, 0.12306224100029794, 0.12233027599904744, 0.1246509380016505, 0.124110157999894, 0.12089498099885532, 0.12182642399966426, 0.12144772999999986, 0.12119378600073105, 0.12001334500018856, 0.1220215659996029, 0.12093306399947323, 0.11915424600010738, 0.12078475500129571, 0.12089011500029301, 0.12181639699883817, 0.12263686500045878, 0.1200093419993209, 0.12076296200029901, 0.12079921499935153, 0.12353704199995263, 0.12187649500083353, 0.12423025399948529, 0.11987450300148339, 0.12092611499974737, 0.12138713399872358, 0.1206450900008349, 0.12294980799924815, 0.12076088400135632, 0.1215450179988693, 0.12135088000104588, 0.12222720199861214, 0.12213303099997574, 0.11999345600088418, 0.12079213200013328, 0.12200069499886013, 0.1264602589999413, 0.12287412300065625, 0.12127580200103694, 0.12292398499994306, 0.12172564799948304, 0.12056923899945104, 0.12128273999951489, 0.1231867290007358, 0.12483643099949404, 0.12175746000139043, 0.12070219899942458, 0.11954009600049176, 0.12085533399840642, 0.12335918600001605, 0.12004661400169425, 0.12566335200062895, 0.11992379699950106, 0.11876398300046276, 0.11962232299993047, 0.11908127400056401, 0.11832283799958532, 0.11911271799908718, 0.11992399900009332, 0.12012256300113222, 0.11812419399939245, 0.11976807999963057, 0.11875111600056698, 0.11890483099887206, 0.12280139300128212, 0.12017341600039799, 0.11947381499885523, 0.1219672580009501, 0.12080325799979619, 0.12031896999906166, 0.11983528699965973, 0.12042969600042852, 0.12028177100000903, 0.11971244100095646, 0.11803114899885259, 0.11948167800073861, 0.12176998699942487, 0.12024816099983582, 0.12100554600147007, 0.12085069699969608, 0.12057134099995892, 0.11928951699883328, 0.12382623099983903, 0.1307168970015482, 0.1248424059995159, 0.12109710299955623, 0.1212537789997441, 0.12562769100077276, 0.12171605000003183, 0.11969403500006592, 0.11894088599910901, 0.12036312199961685, 0.12430887200025609, 0.12046962299973529, 0.11964504200113879, 0.11947457600035705, 0.11901620999924489, 0.12068077699950663, 0.11945054500029073, 0.12124997399951098, 0.12247635000130686, 0.12672141899929557, 0.12083853500007535, 0.12175031399965519, 0.12065210699984164, 0.12077957400106243, 0.12034097600007954, 0.12049119899893412, 0.12167684700034442, 0.11946625400014454, 0.12045089800085407, 0.12054505099877133, 0.12080046600021888, 0.12159939900084282, 0.12412451300042449, 0.11998414899971976, 0.1196146399997815, 0.11930922000101418, 0.12258898800064344, 0.1223407399993448, 0.11975932799941802, 0.12090481400082354, 0.12005625199890346, 0.11924902900136658, 0.11860365299980913, 0.11923545999889029, 0.12101181400066707, 0.1216791770002601, 0.123773078999875, 0.12220066400004725, 0.12332924600013939, 0.12284179000016593, 0.12165102099970682, 0.12011415999950259, 0.12051365700062888, 0.11842704300033802, 0.12052883700016537, 0.12249512199923629, 0.1230797080006596, 0.12033308999889414, 0.1215944909999962, 0.119898020000619, 0.12092862399913429, 0.12103451199982374, 0.12080213900117087, 0.12335276399971917, 0.12105136600075639, 0.1196696589995554, 0.11938469800043094, 0.12014067199925194, 0.1194958169999154, 0.1198796240005322, 0.11954885599880072, 0.12007280299985723, 0.12120646900075371, 0.11948378899978707, 0.11863167400042585, 0.11825291899913282, 0.12243966200003342, 0.12080467800115002, 0.12196459499864432, 0.1213838770017901, 0.12121472099897801, 0.12146955599928333, 0.11987917800070136, 0.11924499099950481, 0.12263454400090268, 0.12579768899922783, 0.11979330600115645, 0.12304032599968195, 0.12302596399968024, 0.12340083500021137, 0.12294098199890868, 0.11963276700043934, 0.12009222300002875, 0.12144073200033745, 0.12142396300077962, 0.12518793500021275, 0.1207882899998367, 0.12099835700064432, 0.12111784999979136, 0.12704969399965194, 0.12019187999976566, 0.11865199999920151, 0.11872035300075368, 0.12022398299995984, 0.11935670400089293, 0.11934961199949612, 0.1191048740001861, 0.11841419499978656, 0.12102527699971688, 0.12196309399951133, 0.12395833000118728, 0.12079836500015517, 0.12136613699840382, 0.1206565340016823, 0.11934667799869203, 0.11958636000053957, 0.11947411099936289, 0.11865852000119048, 0.11858637600016664, 0.12300505999883171, 0.12192993700045918, 0.11998002099971927, 0.12080434600102308, 0.11885779199837998, 0.1204654870016384, 0.12347350699928938, 0.12033252899891522, 0.121176649001427, 0.12038413600021158, 0.12132072499844071, 0.12224314800005232, 0.12102034700001241, 0.12135282500094036, 0.12174263199995039, 0.12037456299913174, 0.1256208280010469, 0.1266844769997988, 0.12311812100051611, 0.12397330000021611, 0.12091306899856136, 0.1201288440006465, 0.12060155799917993, 0.12227935800001433, 0.12125104900042061, 0.12061224799981574, 0.11987201200099662, 0.12160787199900369, 0.11944812000001548, 0.12075581700082694, 0.12168855100026121, 0.12549251899872615, 0.1231371810008568, 0.12221377300011227, 0.12308060599934834, 0.1223549610003829, 0.12156838500050071, 0.11957395199897292, 0.12468366299981426, 0.12560875800045324, 0.12254044800101838, 0.12039999799890211, 0.12126674400133197, 0.12026276799952029, 0.12259400199945958, 0.12024273200040625, 0.11916080599985435, 0.12159320200044021, 0.12325635599881934, 0.11987760100055311, 0.12013302900049894, 0.12085170299906167, 0.1206933930006926, 0.1199357389996294, 0.12047660200005339, 0.1214139530002285, 0.12169837500005087, 0.12157120099982421, 0.12038357299934432, 0.12222018200009188, 0.12213181100014481, 0.12031943600049999, 0.11887824299992644, 0.12077750299977197, 0.12089378299970122, 0.12091825199968298, 0.1205859670008067, 0.1197430379997968, 0.11932992299989564, 0.1206294590010657, 0.12063620399931096, 0.12068379999982426, 0.1198699159995158, 0.12058362200150441, 0.12191187899952638, 0.12020435100021132, 0.11957999599871982, 0.11813040600100067, 0.12170119199981855, 0.12229334999938146, 0.11963032000130625, 0.12021331299911253, 0.11831096400055685, 0.12166754600002605, 0.1203495629997633, 0.12300233200039656, 0.12050785099927452, 0.1204240179995395, 0.12332531799984281, 0.12001733600118314, 0.12040772199907224, 0.11857455100107472, 0.11821071299891628, 0.11986574399998062, 0.12073068999961833, 0.12043593300040811, 0.12004464600067877, 0.12230334299965762, 0.11897525799940922, 0.12170995700034837, 0.11954773000070418, 0.11868315199899371]",tokens/s,8.274490203598683,, float32-eager,pytorch,2.4.1+cpu,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,qwen,Qwen/Qwen-7B,Qwen/Qwen-7B,cpu,,42,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,False,,False,spawn, Intel(R) Xeon(R) Platinum 8488C,32,66326.188032,Linux,x86_64,Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,0.4.0,,4.44.2,,0.34.2,,,,1.22.0,,,,,,"Traceback (most recent call last): File ""/workspace/llm_perf/common/benchmark_runner.py"", line 106, in execute_and_log_benchmark benchmark_report = Benchmark.launch(benchmark_config)