|
|
|
--- |
|
|
|
license: llama2 |
|
|
|
--- |
|
|
|
[![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory) |
|
|
|
|
|
# QuantFactory/LLaMA-Pro-8B-Instruct-GGUF |
|
This is quantized version of [TencentARC/LLaMA-Pro-8B-Instruct](https://huggingface.co/TencentARC/LLaMA-Pro-8B-Instruct) created using llama.cpp |
|
|
|
# Original Model Card |
|
|
|
|
|
# LLaMA-PRO-Instruct Model Card |
|
|
|
## Model Description |
|
LLaMA-PRO-Instruct is a transformative expansion of the LLaMA2-7B model, now boasting 8.3 billion parameters. It uniquely specializes in programming, coding, and mathematical reasoning, maintaining versatility in general language tasks. |
|
|
|
## Development and Training |
|
This model, developed by Tencent ARC team, extends LLaMA2-7B using innovative block expansion techniques. It's meticulously trained on a diverse blend of coding and mathematical data, encompassing over 80 billion tokens. |
|
|
|
## Intended Use |
|
LLaMA-PRO-Instruct is ideal for complex NLP challenges, excelling in programming, mathematical reasoning, and general language processing, suitable for both specialized and broad applications. |
|
|
|
## Performance |
|
It surpasses its predecessors in the LLaMA series, especially in code domains, demonstrating exceptional competence as a comprehensive language model. |
|
|
|
## Limitations |
|
Despite advancements, it may encounter difficulties in highly niche or nuanced tasks. |
|
|
|
## Ethical Considerations |
|
Users are advised to consider inherent biases and responsibly manage its application across various fields. |
|
|