|
{ |
|
"config": { |
|
"name": "pytorch_generate", |
|
"backend": { |
|
"name": "pytorch", |
|
"version": "2.3.0+cu121", |
|
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", |
|
"task": "text-generation", |
|
"library": "transformers", |
|
"model": "google/gemma-2b", |
|
"processor": "google/gemma-2b", |
|
"device": "cuda", |
|
"device_ids": "0", |
|
"seed": 42, |
|
"inter_op_num_threads": null, |
|
"intra_op_num_threads": null, |
|
"model_kwargs": {}, |
|
"processor_kwargs": {}, |
|
"hub_kwargs": {}, |
|
"no_weights": true, |
|
"device_map": null, |
|
"torch_dtype": "float16", |
|
"eval_mode": true, |
|
"to_bettertransformer": false, |
|
"low_cpu_mem_usage": null, |
|
"attn_implementation": null, |
|
"cache_implementation": null, |
|
"autocast_enabled": false, |
|
"autocast_dtype": null, |
|
"torch_compile": false, |
|
"torch_compile_target": "forward", |
|
"torch_compile_config": { |
|
"backend": "inductor", |
|
"mode": "reduce-overhead", |
|
"fullgraph": true |
|
}, |
|
"quantization_scheme": null, |
|
"quantization_config": {}, |
|
"deepspeed_inference": false, |
|
"deepspeed_inference_config": {}, |
|
"peft_type": null, |
|
"peft_config": {} |
|
}, |
|
"scenario": { |
|
"name": "inference", |
|
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", |
|
"iterations": 2, |
|
"duration": 0, |
|
"warmup_runs": 10, |
|
"input_shapes": { |
|
"batch_size": 1, |
|
"num_choices": 2, |
|
"sequence_length": 7 |
|
}, |
|
"new_tokens": null, |
|
"latency": true, |
|
"memory": true, |
|
"energy": false, |
|
"forward_kwargs": {}, |
|
"generate_kwargs": { |
|
"max_new_tokens": 128, |
|
"min_new_tokens": 128, |
|
"do_sample": false |
|
}, |
|
"call_kwargs": {} |
|
}, |
|
"launcher": { |
|
"name": "process", |
|
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", |
|
"device_isolation": true, |
|
"device_isolation_action": "warn", |
|
"start_method": "spawn" |
|
}, |
|
"environment": { |
|
"cpu": " AMD EPYC 7R32", |
|
"cpu_count": 16, |
|
"cpu_ram_mb": 66697.29792, |
|
"system": "Linux", |
|
"machine": "x86_64", |
|
"platform": "Linux-5.10.218-208.862.amzn2.x86_64-x86_64-with-glibc2.29", |
|
"processor": "x86_64", |
|
"python_version": "3.8.10", |
|
"gpu": [ |
|
"NVIDIA A10G" |
|
], |
|
"gpu_count": 1, |
|
"gpu_vram_mb": 24146608128, |
|
"optimum_benchmark_version": "0.2.1", |
|
"optimum_benchmark_commit": null, |
|
"transformers_version": "4.42.0.dev0", |
|
"transformers_commit": "1de7dc7403b3b89ec421d43a8c9ee245211a61f6", |
|
"accelerate_version": "0.32.0.dev0", |
|
"accelerate_commit": null, |
|
"diffusers_version": null, |
|
"diffusers_commit": null, |
|
"optimum_version": "1.21.0.dev0", |
|
"optimum_commit": null, |
|
"timm_version": "0.9.16", |
|
"timm_commit": null, |
|
"peft_version": "0.11.2.dev0", |
|
"peft_commit": null |
|
} |
|
}, |
|
"report": { |
|
"prefill": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 1730.535424, |
|
"max_global_vram": 6793.199616, |
|
"max_process_vram": 0.0, |
|
"max_reserved": 6146.752512, |
|
"max_allocated": 5034.576896 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"count": 2, |
|
"total": 0.03975424003601074, |
|
"mean": 0.01987712001800537, |
|
"stdev": 0.0001449594497680666, |
|
"p50": 0.01987712001800537, |
|
"p90": 0.019993087577819823, |
|
"p95": 0.020007583522796628, |
|
"p99": 0.020019180278778075, |
|
"values": [ |
|
0.020022079467773436, |
|
0.019732160568237303 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 352.16369341530174 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"decode": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 1730.617344, |
|
"max_global_vram": 6795.296768, |
|
"max_process_vram": 0.0, |
|
"max_reserved": 6148.849664, |
|
"max_allocated": 5034.577408 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"count": 2, |
|
"total": 4.788974609375, |
|
"mean": 2.3944873046875, |
|
"stdev": 0.0072429199218748685, |
|
"p50": 2.3944873046875, |
|
"p90": 2.400281640625, |
|
"p95": 2.401005932617187, |
|
"p99": 2.4015853662109374, |
|
"values": [ |
|
2.401730224609375, |
|
2.387244384765625 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 53.03849377333597 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"per_token": { |
|
"memory": null, |
|
"latency": { |
|
"unit": "s", |
|
"count": 253, |
|
"total": 4.789055486679078, |
|
"mean": 0.018929073069877775, |
|
"stdev": 0.0013167771977965525, |
|
"p50": 0.01884774398803711, |
|
"p90": 0.019217817306518555, |
|
"p95": 0.019372851181030274, |
|
"p99": 0.02100580284118652, |
|
"values": [ |
|
0.02079641532897949, |
|
0.018611200332641603, |
|
0.018375680923461913, |
|
0.018353151321411132, |
|
0.018342912673950194, |
|
0.018343936920166014, |
|
0.018356224060058594, |
|
0.018390016555786134, |
|
0.018348031997680665, |
|
0.018544639587402344, |
|
0.018551807403564453, |
|
0.01845452880859375, |
|
0.018354175567626953, |
|
0.01841971206665039, |
|
0.01840127944946289, |
|
0.01841049575805664, |
|
0.01855897521972656, |
|
0.01841868782043457, |
|
0.018280448913574218, |
|
0.01825279998779297, |
|
0.01836031913757324, |
|
0.01842278480529785, |
|
0.018415615081787108, |
|
0.018381824493408205, |
|
0.01839820861816406, |
|
0.018415615081787108, |
|
0.01842790412902832, |
|
0.01838489532470703, |
|
0.01844428825378418, |
|
0.018420736312866212, |
|
0.018380800247192384, |
|
0.018372608184814454, |
|
0.018363391876220703, |
|
0.01839923286437988, |
|
0.01866854476928711, |
|
0.0190382080078125, |
|
0.01900441551208496, |
|
0.019005439758300782, |
|
0.018971647262573242, |
|
0.018670591354370117, |
|
0.01882828712463379, |
|
0.018967552185058592, |
|
0.01904537582397461, |
|
0.019129344940185547, |
|
0.019141632080078123, |
|
0.018903039932250978, |
|
0.01900032043457031, |
|
0.018807807922363282, |
|
0.018580480575561522, |
|
0.018749439239501953, |
|
0.01884774398803711, |
|
0.018980863571166993, |
|
0.019308544158935546, |
|
0.0193832950592041, |
|
0.020512767791748047, |
|
0.02000383949279785, |
|
0.02027827262878418, |
|
0.01946112060546875, |
|
0.018569215774536133, |
|
0.018976768493652343, |
|
0.019105791091918945, |
|
0.019104768753051758, |
|
0.019573759078979493, |
|
0.019150848388671874, |
|
0.019331071853637697, |
|
0.019215360641479492, |
|
0.019154943466186524, |
|
0.019105791091918945, |
|
0.019133440017700197, |
|
0.018930688858032226, |
|
0.019144704818725586, |
|
0.019090431213378906, |
|
0.01927884864807129, |
|
0.019360767364501954, |
|
0.019122175216674805, |
|
0.01929523277282715, |
|
0.019107839584350587, |
|
0.019129344940185547, |
|
0.019186687469482423, |
|
0.019111936569213867, |
|
0.018946048736572265, |
|
0.018503679275512695, |
|
0.018525184631347655, |
|
0.018536447525024414, |
|
0.018531328201293946, |
|
0.01863987159729004, |
|
0.019104768753051758, |
|
0.01904025650024414, |
|
0.019072000503540038, |
|
0.018917375564575196, |
|
0.018526208877563476, |
|
0.019143680572509765, |
|
0.019104768753051758, |
|
0.019099647521972657, |
|
0.019170303344726563, |
|
0.019078144073486326, |
|
0.019152896881103516, |
|
0.019117055892944337, |
|
0.019171327590942384, |
|
0.019130367279052735, |
|
0.01921843147277832, |
|
0.01921023941040039, |
|
0.019147775650024415, |
|
0.01904844856262207, |
|
0.019100671768188478, |
|
0.019155967712402345, |
|
0.019325952529907226, |
|
0.019133440017700197, |
|
0.018563072204589845, |
|
0.018486272811889647, |
|
0.018852863311767578, |
|
0.019110912322998046, |
|
0.019195903778076173, |
|
0.019365888595581054, |
|
0.01924300765991211, |
|
0.019116031646728517, |
|
0.019165184020996092, |
|
0.019161088943481445, |
|
0.019087360382080077, |
|
0.01906892776489258, |
|
0.018592767715454102, |
|
0.019191808700561523, |
|
0.019112960815429687, |
|
0.019129344940185547, |
|
0.01859686470031738, |
|
0.018536447525024414, |
|
0.03857100677490234, |
|
0.01882828712463379, |
|
0.019087423324584962, |
|
0.018466751098632814, |
|
0.018811904907226562, |
|
0.018766847610473633, |
|
0.018877439498901367, |
|
0.018408447265625, |
|
0.018762752532958983, |
|
0.01899519920349121, |
|
0.018977792739868164, |
|
0.019107839584350587, |
|
0.019021823883056642, |
|
0.01900032043457031, |
|
0.018980863571166993, |
|
0.01845145606994629, |
|
0.01843609619140625, |
|
0.018495487213134765, |
|
0.018431999206542968, |
|
0.01865727996826172, |
|
0.018408447265625, |
|
0.018365440368652345, |
|
0.01843609619140625, |
|
0.018449407577514648, |
|
0.019135488510131835, |
|
0.018447359085083007, |
|
0.018491392135620118, |
|
0.018479103088378905, |
|
0.01906380844116211, |
|
0.0190248966217041, |
|
0.01938739204406738, |
|
0.01926655960083008, |
|
0.019162111282348633, |
|
0.019219455718994142, |
|
0.019594240188598632, |
|
0.01902694320678711, |
|
0.01904640007019043, |
|
0.019334144592285156, |
|
0.019056640625, |
|
0.019389440536499023, |
|
0.018969600677490234, |
|
0.018962432861328125, |
|
0.018964479446411133, |
|
0.018898944854736328, |
|
0.018923519134521484, |
|
0.018907136917114258, |
|
0.018974720001220705, |
|
0.018469888687133788, |
|
0.018412544250488282, |
|
0.019042303085327148, |
|
0.01861529541015625, |
|
0.018504703521728515, |
|
0.018954240798950195, |
|
0.018982912063598634, |
|
0.019074047088623047, |
|
0.018984960556030273, |
|
0.01842790412902832, |
|
0.018720767974853517, |
|
0.019171327590942384, |
|
0.01920921516418457, |
|
0.019170303344726563, |
|
0.019086336135864256, |
|
0.019143680572509765, |
|
0.019204095840454103, |
|
0.019139583587646485, |
|
0.01920204734802246, |
|
0.01908940887451172, |
|
0.019159040451049804, |
|
0.019186687469482423, |
|
0.02123263931274414, |
|
0.021794815063476563, |
|
0.019360767364501954, |
|
0.019105791091918945, |
|
0.019072000503540038, |
|
0.01907302474975586, |
|
0.018535423278808593, |
|
0.018584575653076172, |
|
0.01861324882507324, |
|
0.01857535934448242, |
|
0.018531328201293946, |
|
0.018514944076538087, |
|
0.01856716728210449, |
|
0.018538496017456055, |
|
0.018493440628051756, |
|
0.018537471771240235, |
|
0.01901670455932617, |
|
0.01862451171875, |
|
0.018557952880859374, |
|
0.01863987159729004, |
|
0.018521087646484375, |
|
0.018588672637939452, |
|
0.018557952880859374, |
|
0.018569215774536133, |
|
0.018505727767944336, |
|
0.018534400939941405, |
|
0.018551807403564453, |
|
0.018544639587402344, |
|
0.01856716728210449, |
|
0.018532352447509767, |
|
0.018519039154052733, |
|
0.018578432083129884, |
|
0.018497535705566406, |
|
0.018534400939941405, |
|
0.018543615341186523, |
|
0.01843609619140625, |
|
0.018526208877563476, |
|
0.018556928634643553, |
|
0.01847500801086426, |
|
0.018485248565673826, |
|
0.018499584197998048, |
|
0.018478080749511717, |
|
0.01848422431945801, |
|
0.018537471771240235, |
|
0.0184770565032959, |
|
0.018504703521728515, |
|
0.018485248565673826, |
|
0.0184586238861084, |
|
0.018510847091674804, |
|
0.018506752014160157, |
|
0.01848422431945801, |
|
0.018485248565673826, |
|
0.018503679275512695, |
|
0.018529279708862305, |
|
0.018465791702270508, |
|
0.018494464874267577, |
|
0.018521087646484375, |
|
0.018508800506591795 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 52.82878862099807 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
} |
|
} |
|
} |