QuantFactory/Dolus-14b-Mini-GGUF

This is quantized version of Cognitive-Machines-Labs/Dolus-14b-Mini created using llama.cpp

Original Model Card

Dolus-14b-Mini

logo

Ursidae-12b-Mini
The little sister to Ursidae-300b,Dolus 14b Mini has been developed with a focus on complex multi-step chain of thought problem solving while still being deployable on edge systems. A model focused complex multi-step chain of thought problem solving while still being deployable on edge systems. Now better at reasoning!

Main Goals:

Dolus was designed to address specific issues found in other chat models:

  • Overcome limitations in logical reasoning found in other chat models.
  • Efficiently solve complex, multi-step problems.
  • Provide better decision-making assistance by enhancing the model's ability to reason and think critically.
  • Removing restrictions and allowing the model to gain a true understanding of reality, greatly increasing overall results.

By focusing on these specific goals, the Ursidae-12b-Mini aims to provide a more sophisticated AI system that excels at critical thinking and problem-solving tasks requiring advanced logical reasoning skills. Its compact design makes it an efficient choice for applications where high cognitive abilities are necessary without occupying excessive computing resources.

Recommended Settings:

Defaults:

min_p: 0.074
top_k: 40
repetition_penalty: 1.12
temp: 1.18
context: 8192

Creative:

min_p: 0.062
top_k: 40
repetition_penalty: 1.11
temp: 1.24
context: 8192

Benchmarks:

PENDING FULL EVAL

Downloads last month
38
GGUF
Model size
11.5B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.