Edit model card

NOTE: You will need a recent build of llama.cpp to run these quants (i.e. at least commit 494c870).

GGUF importance matrix (imatrix) quants for https://huggingface.co/ibm/labradorite-13b

Layers Context Template
40
4096
<|system|>
{sys_prompt}
<|user|>
{inputs}
<|assistant|>
{response}<|endoftext|>
Downloads last month
8
GGUF
Model size
13B params
Architecture
llama

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Examples
Inference API (serverless) does not yet support gguf models for this pipeline type.

Model tree for dranger003/labradorite-13b-iMat.GGUF

Quantized
(3)
this model