sharpenb's picture
Upload folder using huggingface_hub (#2)
c7aac50 verified
raw
history blame
1.69 kB
{
"base_current_gpu_type": "NVIDIA A100-PCIE-40GB",
"base_current_gpu_total_memory": 40339.3125,
"base_perplexity": 7.6169915199279785,
"base_token_generation_latency_sync": 36.8033504486084,
"base_token_generation_latency_async": 37.4163955450058,
"base_token_generation_throughput_sync": 0.027171439225251618,
"base_token_generation_throughput_async": 0.02672625156523064,
"base_token_generation_CO2_emissions": null,
"base_token_generation_energy_consumption": null,
"base_inference_latency_sync": 118.06535568237305,
"base_inference_latency_async": 38.49177360534668,
"base_inference_throughput_sync": 0.008469885126084436,
"base_inference_throughput_async": 0.025979577097509882,
"base_inference_CO2_emissions": null,
"base_inference_energy_consumption": null,
"smashed_current_gpu_type": "NVIDIA A100-PCIE-40GB",
"smashed_current_gpu_total_memory": 40339.3125,
"smashed_perplexity": 8.069676399230957,
"smashed_token_generation_latency_sync": 44.376717758178714,
"smashed_token_generation_latency_async": 45.14040872454643,
"smashed_token_generation_throughput_sync": 0.02253433896236497,
"smashed_token_generation_throughput_async": 0.02215310025441175,
"smashed_token_generation_CO2_emissions": null,
"smashed_token_generation_energy_consumption": null,
"smashed_inference_latency_sync": 161.3275146484375,
"smashed_inference_latency_async": 69.88525390625,
"smashed_inference_throughput_sync": 0.0061985706665052456,
"smashed_inference_throughput_async": 0.014309170305676857,
"smashed_inference_CO2_emissions": null,
"smashed_inference_energy_consumption": null
}