ArtusDev/ToastyPigeon_Gemma-3-Confetti-12B-GGUF

โœจ Overview

This repository contains GGUF format model files converted from ToastyPigeon/Gemma-3-Confetti-12B.

The conversion was performed by ArtusDev using llama.cpp, specifically utilizing the imatrix quantization option for potentially improved performance.

๐Ÿ“„ Original Model Details

For comprehensive information about the model's architecture, training data, intended use, and evaluation, please refer to the original model card.

๐Ÿ’ฌ Instruct Format

This model uses the standard Gemma 2/3 instruct format. While primarily trained on this format, it might recognize an optional system role based on its fine-tuning:

<start_of_turn>system
{optional system prompt here}<end_of_turn>
<start_of_turn>user
{User messages. You can also place the system prompt here.}<end_of_turn>
<start_of_turn>model
{Model's response}<end_of_turn>

Note: Always ensure the format strictly adheres to the required tokens and structure for optimal model performance.

GGUF Quantizations (imatrix) by ArtusDev
Downloads last month
348
GGUF
Model size
11.8B params
Architecture
gemma3

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ArtusDev/ToastyPigeon_Gemma-3-Confetti-12B-GGUF

Quantized
(3)
this model