Adept-14B-AWQ

Introduction

Adept-14B is a 14-billion-parameter generative AI model, leveraging Qwen2.5 14B and employing 4-bit AWQ quantization for efficiency. It is designed to provide organizations and developers with cutting-edge generative AI capabilities in a compact form, enabling high-quality instruction-following, complex reasoning, and tasks tailored for business applications.

From Qwen2.5:

  • Significantly more knowledge and has greatly improved capabilities in coding and mathematics, thanks to our specialized expert models in these domains.
  • Significant improvements in instruction following, generating long texts (over 8K tokens), understanding structured data (e.g, tables), and generating structured outputs especially JSON. More resilient to the diversity of system prompts, enhancing role-play implementation and condition-setting for chatbots.
  • Long-context Support up to 128K tokens and can generate up to 8K tokens.
  • Multilingual support for over 29 languages, including Chinese, English, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, Arabic, and more.

has the following features:

  • Type: Causal Language Models
  • Training Stage: Pretraining & Post-training
  • Architecture: transformers with RoPE, SwiGLU, RMSNorm, and Attention QKV bias
  • Number of Parameters: 14.7B
  • Number of Paramaters (Non-Embedding): 13.1B
  • Number of Layers: 48
  • Number of Attention Heads (GQA): 40 for Q and 8 for KV
  • Context Length: Full 131,072 tokens and generation 8192 tokens
Downloads last month
84
Safetensors
Model size
3.33B params
Tensor type
I32
·
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for sayed0am/Adept-14B-AWQ

Base model

Qwen/Qwen2.5-14B
Quantized
(81)
this model