hf-transformers-bot's picture
Upload folder using huggingface_hub
b0afb70 verified
raw
history blame
1.73 kB
[
{
"model": "google/gemma-2b",
"commit": "1f9387d33d40562add4be03176e22ddf9ed0d4ab",
"config": "backend.cache_implementation=null,backend.torch_compile=False",
"metrics": {
"prefill.latency.mean": 0.020195904731750487,
"prefill.throughput.value": 346.6049227789793,
"decode.latency.mean": 2.4827767333984374,
"decode.throughput.value": 51.15240460069954,
"per_token.latency.mean": 0.01962751399104303,
"per_token.throughput.value": 50.948887386163506
}
},
{
"model": "google/gemma-2b",
"commit": "1f9387d33d40562add4be03176e22ddf9ed0d4ab",
"config": "backend.cache_implementation=static,backend.torch_compile=False",
"metrics": {
"prefill.latency.mean": 0.022427599906921385,
"prefill.throughput.value": 312.11543049863883,
"decode.latency.mean": 2.6374805908203127,
"decode.throughput.value": 48.15201311509947,
"per_token.latency.mean": 0.02085368712052055,
"per_token.throughput.value": 47.95315064528684
}
},
{
"model": "google/gemma-2b",
"commit": "1f9387d33d40562add4be03176e22ddf9ed0d4ab",
"config": "backend.cache_implementation=static,backend.torch_compile=True",
"metrics": {
"prefill.latency.mean": 0.014046095848083496,
"prefill.throughput.value": 498.35912240020116,
"decode.latency.mean": 1.5763082275390625,
"decode.throughput.value": 80.5679991902807,
"per_token.latency.mean": 0.012464994140293287,
"per_token.throughput.value": 80.22466667412898
}
}
]