Adept-14B-AWQ / README.md
sayed0am's picture
Update README.md
bdb1bf4 verified
---
license: apache-2.0
language:
- en
- ar
- zh
- fr
- ru
- pt
- es
pipeline_tag: text-generation
base_model: Qwen/Qwen2.5-14B
tags:
- chat
library_name: transformers
---
# Adept-14B-AWQ
## Introduction
Adept-14B is a 14-billion-parameter generative AI model, leveraging Qwen2.5 14B and employing 4-bit AWQ quantization for efficiency. It is designed to provide organizations and developers with cutting-edge generative AI capabilities in a compact form, enabling high-quality instruction-following, complex reasoning, and tasks tailored for business applications.
**From Qwen2.5:**
- Significantly **more knowledge** and has greatly improved capabilities in **coding** and **mathematics**, thanks to our specialized expert models in these domains.
- Significant improvements in **instruction following**, **generating long texts** (over 8K tokens), **understanding structured data** (e.g, tables), and **generating structured outputs** especially JSON. **More resilient to the diversity of system prompts**, enhancing role-play implementation and condition-setting for chatbots.
- **Long-context Support** up to 128K tokens and can generate up to 8K tokens.
- **Multilingual support** for over 29 languages, including Chinese, English, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, Arabic, and more.
has the following features:
- Type: Causal Language Models
- Training Stage: Pretraining & Post-training
- Architecture: transformers with RoPE, SwiGLU, RMSNorm, and Attention QKV bias
- Number of Parameters: 14.7B
- Number of Paramaters (Non-Embedding): 13.1B
- Number of Layers: 48
- Number of Attention Heads (GQA): 40 for Q and 8 for KV
- Context Length: Full 131,072 tokens and generation 8192 tokens