Triangle104 commited on
Commit
98c8bbc
·
verified ·
1 Parent(s): 0db8e96

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -0
README.md CHANGED
@@ -19,6 +19,76 @@ language:
19
  This model was converted to GGUF format from [`Spestly/Athena-1-3B`](https://huggingface.co/Spestly/Athena-1-3B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/Spestly/Athena-1-3B) for more details on the model.
21
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  ## Use with llama.cpp
23
  Install llama.cpp through brew (works on Mac and Linux)
24
 
 
19
  This model was converted to GGUF format from [`Spestly/Athena-1-3B`](https://huggingface.co/Spestly/Athena-1-3B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/Spestly/Athena-1-3B) for more details on the model.
21
 
22
+ ---
23
+ Model details:
24
+ -
25
+ Athena-1 3B is a fine-tuned, instruction-following large language model derived from Qwen/Qwen2.5-3B-Instruct. It is designed to provide efficient, high-quality text generation while maintaining a compact size. Athena 3B is optimized for lightweight applications, conversational AI, and structured data tasks, making it ideal for real-world use cases where performance and resource efficiency are critical.
26
+
27
+ Key Features
28
+ -
29
+ ⚡ Lightweight and Efficient
30
+ -
31
+ Compact Size: At just 3.09 billion parameters, Athena-1 3B offers excellent performance with reduced computational requirements.
32
+ Instruction Following: Fine-tuned for precise and reliable adherence to user prompts.
33
+ Coding and Mathematics: Proficient in solving coding challenges and handling mathematical tasks.
34
+
35
+ 📖 Long-Context Understanding
36
+ -
37
+ Context Length: Supports up to 32,768 tokens, enabling the processing of moderately lengthy documents or conversations.
38
+ Token Generation: Can generate up to 8K tokens of output.
39
+
40
+ 🌍 Multilingual Support
41
+ -
42
+ Supports 29+ languages, including:
43
+ English, Chinese, French, Spanish, Portuguese, German, Italian, Russian
44
+ Japanese, Korean, Vietnamese, Thai, Arabic, and more.
45
+
46
+ 📊 Structured Data & Outputs
47
+ -
48
+ Structured Data Interpretation: Processes structured formats like tables and JSON.
49
+ Structured Output Generation: Generates well-formatted outputs, including JSON and other structured formats.
50
+
51
+ Details
52
+ -
53
+ Base Model: Qwen/Qwen2.5-3B-Instruct
54
+ Architecture: Transformers with RoPE, SwiGLU, RMSNorm, Attention QKV bias, and tied word embeddings.
55
+ Parameters: 3.09B total (2.77B non-embedding).
56
+ Layers: 36
57
+ Attention Heads: 16 for Q, 2 for KV.
58
+ Context Length: Up to 32,768 tokens.
59
+
60
+ Applications
61
+ -
62
+ Athena 3B is designed for a variety of real-world applications:
63
+
64
+ Conversational AI: Build fast, responsive, and lightweight chatbots.
65
+ Code Generation: Generate, debug, or explain code snippets.
66
+ Mathematical Problem Solving: Assist with calculations and reasoning.
67
+ Document Processing: Summarize and analyze moderately large documents.
68
+ Multilingual Applications: Support for global use cases with diverse language requirements.
69
+ Structured Data: Process and generate structured data, such as tables and JSON.
70
+
71
+ Quickstart
72
+ -
73
+ Here’s how you can use Athena 3B for quick text generation:
74
+
75
+ ---
76
+ # Use a pipeline as a high-level helper
77
+ from transformers import pipeline
78
+
79
+ messages = [
80
+ {"role": "user", "content": "Who are you?"},
81
+ ]
82
+ pipe = pipeline("text-generation", model="Spestly/Athena-1-3B")
83
+ pipe(messages)
84
+
85
+ # Load model directly
86
+ from transformers import AutoTokenizer, AutoModelForCausalLM
87
+
88
+ tokenizer = AutoTokenizer.from_pretrained("Spestly/Athena-1-3B")
89
+ model = AutoModelForCausalLM.from_pretrained("Spestly/Athena-1-3B")
90
+
91
+ ---
92
  ## Use with llama.cpp
93
  Install llama.cpp through brew (works on Mac and Linux)
94