Using turboderp's ExLlamaV2 v0.2.8 for quantization.
2.2bpw
3.0bpw
4.0bpw
5.0bpw
6.0bpw
7.0bpw
8.0bpw
Calibration Dataset
TFMC/imatrix-dataset-for-japanese-llm
qwen2.5-bakeneko-32b-exl2
- Model creator: rinna
- Original model: qwen2.5-bakeneko-32b
License
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
HF Inference deployability: The model has no library tag.