SuperNova-Medius-FP8-Dynamic

This is a FP8-quantized version of arcee-ai/SuperNova-Medius using the llmcompressor library.

For more information about the quantization method, please visit FP8 documentation used for quantization.

Downloads last month
59
Safetensors
Model size
14.8B params
Tensor type
BF16
·
F8_E4M3
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for nfunctor/SuperNova-Medius-FP8-Dynamic

Base model

Qwen/Qwen2.5-14B
Quantized
(17)
this model