Quantized GLM-4 9b q8_0

Quantization based on ChatGLM.CPP convert.py script.

Download: chatglm4-ggml-int8.bin

Download in shell CLI

https://huggingface.co/npc0/chatglm-4-9b-int8/resolve/main/chatglm4-ggml-int8.bin

协议

GLM-4 模型的权重的使用则需要遵循 LICENSE

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including npc0/chatglm-4-9b-int8