EXL2 Quantizations of calme-3.2-instruct-78b

Using exllamav2 release 0.2.6 for quantization.

Original model: https://huggingface.co/MaziyarPanahi/calme-3.2-instruct-78b

Bits 6.5,lm_head 8.0

"quantization_config": {
    "quant_method": "exl2",
    "version": "0.2.6",
    "bits": 6.5,
    "head_bits": 8,
    "calibration": {
        "rows": 115,
        "length": 2048,
        "dataset": "(default)"
    }
Downloads last month
17
Inference API
Unable to determine this model's library. Check the docs .

Model tree for Zenabius/calme-3.2-instruct-78b-exl2

Quantized
(5)
this model