GLM-4-9b Quantized with AutoFP8

使用 m-a-p/COIG-CQIA 的 COIG_pc 集作为校准量化的 glm-4-9b 基座

通常来讲你不会这样用基座模型。

Downloads last month
14
Safetensors
Model size
9.4B params
Tensor type
BF16
·
F8_E4M3
·
Inference Examples
Inference API (serverless) does not yet support model repos that contain custom code.