Edit model card

Model Card for Cerebrum-1.0-8x7b-GGUF

Quantized from https://huggingface.co/AetherResearch/Cerebrum-1.0-8x7b using llama.cpp commit 46acb3676718b983157058aecf729a2064fc7d34

Actual quants are currently uploading with my shitty german broadband speed of ~ 40Mbit/s, stay tuned.

Downloads last month
77
GGUF
Model size
46.7B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.