Triangle104 commited on
Commit
f28a91a
·
verified ·
1 Parent(s): 98c8bbc

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +0 -70
README.md CHANGED
@@ -19,76 +19,6 @@ language:
19
  This model was converted to GGUF format from [`Spestly/Athena-1-3B`](https://huggingface.co/Spestly/Athena-1-3B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/Spestly/Athena-1-3B) for more details on the model.
21
 
22
- ---
23
- Model details:
24
- -
25
- Athena-1 3B is a fine-tuned, instruction-following large language model derived from Qwen/Qwen2.5-3B-Instruct. It is designed to provide efficient, high-quality text generation while maintaining a compact size. Athena 3B is optimized for lightweight applications, conversational AI, and structured data tasks, making it ideal for real-world use cases where performance and resource efficiency are critical.
26
-
27
- Key Features
28
- -
29
- ⚡ Lightweight and Efficient
30
- -
31
- Compact Size: At just 3.09 billion parameters, Athena-1 3B offers excellent performance with reduced computational requirements.
32
- Instruction Following: Fine-tuned for precise and reliable adherence to user prompts.
33
- Coding and Mathematics: Proficient in solving coding challenges and handling mathematical tasks.
34
-
35
- 📖 Long-Context Understanding
36
- -
37
- Context Length: Supports up to 32,768 tokens, enabling the processing of moderately lengthy documents or conversations.
38
- Token Generation: Can generate up to 8K tokens of output.
39
-
40
- 🌍 Multilingual Support
41
- -
42
- Supports 29+ languages, including:
43
- English, Chinese, French, Spanish, Portuguese, German, Italian, Russian
44
- Japanese, Korean, Vietnamese, Thai, Arabic, and more.
45
-
46
- 📊 Structured Data & Outputs
47
- -
48
- Structured Data Interpretation: Processes structured formats like tables and JSON.
49
- Structured Output Generation: Generates well-formatted outputs, including JSON and other structured formats.
50
-
51
- Details
52
- -
53
- Base Model: Qwen/Qwen2.5-3B-Instruct
54
- Architecture: Transformers with RoPE, SwiGLU, RMSNorm, Attention QKV bias, and tied word embeddings.
55
- Parameters: 3.09B total (2.77B non-embedding).
56
- Layers: 36
57
- Attention Heads: 16 for Q, 2 for KV.
58
- Context Length: Up to 32,768 tokens.
59
-
60
- Applications
61
- -
62
- Athena 3B is designed for a variety of real-world applications:
63
-
64
- Conversational AI: Build fast, responsive, and lightweight chatbots.
65
- Code Generation: Generate, debug, or explain code snippets.
66
- Mathematical Problem Solving: Assist with calculations and reasoning.
67
- Document Processing: Summarize and analyze moderately large documents.
68
- Multilingual Applications: Support for global use cases with diverse language requirements.
69
- Structured Data: Process and generate structured data, such as tables and JSON.
70
-
71
- Quickstart
72
- -
73
- Here’s how you can use Athena 3B for quick text generation:
74
-
75
- ---
76
- # Use a pipeline as a high-level helper
77
- from transformers import pipeline
78
-
79
- messages = [
80
- {"role": "user", "content": "Who are you?"},
81
- ]
82
- pipe = pipeline("text-generation", model="Spestly/Athena-1-3B")
83
- pipe(messages)
84
-
85
- # Load model directly
86
- from transformers import AutoTokenizer, AutoModelForCausalLM
87
-
88
- tokenizer = AutoTokenizer.from_pretrained("Spestly/Athena-1-3B")
89
- model = AutoModelForCausalLM.from_pretrained("Spestly/Athena-1-3B")
90
-
91
- ---
92
  ## Use with llama.cpp
93
  Install llama.cpp through brew (works on Mac and Linux)
94
 
 
19
  This model was converted to GGUF format from [`Spestly/Athena-1-3B`](https://huggingface.co/Spestly/Athena-1-3B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/Spestly/Athena-1-3B) for more details on the model.
21
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  ## Use with llama.cpp
23
  Install llama.cpp through brew (works on Mac and Linux)
24