munish0838 commited on
Commit
73ca0e7
1 Parent(s): a3876a2

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +206 -0
README.md ADDED
@@ -0,0 +1,206 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ language:
5
+ - en
6
+ license: mit
7
+ library_name: transformers
8
+ tags:
9
+ - chat
10
+ - phi
11
+ - phi3
12
+ - phi3.5
13
+ - finetune
14
+ base_model: microsoft/Phi-3.5-mini-instruct
15
+ datasets:
16
+ - MaziyarPanahi/truthy-dpo-v0.1-axolotl
17
+ model_name: calme-2.1-phi3.5-4b
18
+ pipeline_tag: text-generation
19
+ inference: false
20
+ model_creator: MaziyarPanahi
21
+ quantized_by: MaziyarPanahi
22
+ model-index:
23
+ - name: calme-2.1-phi3.5-4b
24
+ results:
25
+ - task:
26
+ type: text-generation
27
+ name: Text Generation
28
+ dataset:
29
+ name: IFEval (0-Shot)
30
+ type: HuggingFaceH4/ifeval
31
+ args:
32
+ num_few_shot: 0
33
+ metrics:
34
+ - type: inst_level_strict_acc and prompt_level_strict_acc
35
+ value: 56.59
36
+ name: strict accuracy
37
+ source:
38
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.1-phi3.5-4b
39
+ name: Open LLM Leaderboard
40
+ - task:
41
+ type: text-generation
42
+ name: Text Generation
43
+ dataset:
44
+ name: BBH (3-Shot)
45
+ type: BBH
46
+ args:
47
+ num_few_shot: 3
48
+ metrics:
49
+ - type: acc_norm
50
+ value: 36.11
51
+ name: normalized accuracy
52
+ source:
53
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.1-phi3.5-4b
54
+ name: Open LLM Leaderboard
55
+ - task:
56
+ type: text-generation
57
+ name: Text Generation
58
+ dataset:
59
+ name: MATH Lvl 5 (4-Shot)
60
+ type: hendrycks/competition_math
61
+ args:
62
+ num_few_shot: 4
63
+ metrics:
64
+ - type: exact_match
65
+ value: 14.43
66
+ name: exact match
67
+ source:
68
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.1-phi3.5-4b
69
+ name: Open LLM Leaderboard
70
+ - task:
71
+ type: text-generation
72
+ name: Text Generation
73
+ dataset:
74
+ name: GPQA (0-shot)
75
+ type: Idavidrein/gpqa
76
+ args:
77
+ num_few_shot: 0
78
+ metrics:
79
+ - type: acc_norm
80
+ value: 12.53
81
+ name: acc_norm
82
+ source:
83
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.1-phi3.5-4b
84
+ name: Open LLM Leaderboard
85
+ - task:
86
+ type: text-generation
87
+ name: Text Generation
88
+ dataset:
89
+ name: MuSR (0-shot)
90
+ type: TAUR-Lab/MuSR
91
+ args:
92
+ num_few_shot: 0
93
+ metrics:
94
+ - type: acc_norm
95
+ value: 9.77
96
+ name: acc_norm
97
+ source:
98
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.1-phi3.5-4b
99
+ name: Open LLM Leaderboard
100
+ - task:
101
+ type: text-generation
102
+ name: Text Generation
103
+ dataset:
104
+ name: MMLU-PRO (5-shot)
105
+ type: TIGER-Lab/MMLU-Pro
106
+ config: main
107
+ split: test
108
+ args:
109
+ num_few_shot: 5
110
+ metrics:
111
+ - type: acc
112
+ value: 32.61
113
+ name: accuracy
114
+ source:
115
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.1-phi3.5-4b
116
+ name: Open LLM Leaderboard
117
+
118
+ ---
119
+
120
+ ![](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)
121
+
122
+ # QuantFactory/calme-2.1-phi3.5-4b-GGUF
123
+ This is quantized version of [MaziyarPanahi/calme-2.1-phi3.5-4b](https://huggingface.co/MaziyarPanahi/calme-2.1-phi3.5-4b) created using llama.cpp
124
+
125
+ # Original Model Card
126
+
127
+
128
+ <img src="./calme-2.webp" alt="Calme-2 Models" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
129
+
130
+ # MaziyarPanahi/calme-2.1-phi3.5-4b
131
+
132
+ This model is a fine-tuned version of the `microsoft/Phi-3.5-mini-instruct`, pushing the boundaries of natural language understanding and generation even further. My goal was to create a versatile and robust model that excels across a wide range of benchmarks and real-world applications.
133
+
134
+ ## Use Cases
135
+
136
+ This model is suitable for a wide range of applications, including but not limited to:
137
+
138
+ - Advanced question-answering systems
139
+ - Intelligent chatbots and virtual assistants
140
+ - Content generation and summarization
141
+ - Code generation and analysis
142
+ - Complex problem-solving and decision support
143
+
144
+ # ⚡ Quantized GGUF
145
+
146
+ Here are the quants: [calme-2.1-phi3.5-4b-GGUF](https://huggingface.co/MaziyarPanahi/calme-2.1-phi3.5-4b-GGUF)
147
+
148
+
149
+ # 🏆 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
150
+
151
+ Coming soon!
152
+
153
+ # Prompt Template
154
+
155
+ This model uses `ChatML` prompt template:
156
+
157
+ ```
158
+ <|system|>
159
+ You are a helpful assistant.<|end|>
160
+ <|user|>
161
+ How to explain Internet for a medieval knight?<|end|>
162
+ <|assistant|>
163
+ ````
164
+
165
+ # How to use
166
+
167
+
168
+ ```python
169
+
170
+ # Use a pipeline as a high-level helper
171
+
172
+ from transformers import pipeline
173
+
174
+ messages = [
175
+ {"role": "user", "content": "Who are you?"},
176
+ ]
177
+ pipe = pipeline("text-generation", model="MaziyarPanahi/calme-2.1-phi3.5-4b")
178
+ pipe(messages)
179
+
180
+
181
+ # Load model directly
182
+
183
+ from transformers import AutoTokenizer, AutoModelForCausalLM
184
+
185
+ tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/calme-2.1-phi3.5-4b")
186
+ model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/calme-2.1-phi3.5-4b")
187
+ ```
188
+
189
+
190
+ # Ethical Considerations
191
+
192
+ As with any large language model, users should be aware of potential biases and limitations. We recommend implementing appropriate safeguards and human oversight when deploying this model in production environments.
193
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
194
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_MaziyarPanahi__calme-2.1-phi3.5-4b)
195
+
196
+ | Metric |Value|
197
+ |-------------------|----:|
198
+ |Avg. |27.01|
199
+ |IFEval (0-Shot) |56.59|
200
+ |BBH (3-Shot) |36.11|
201
+ |MATH Lvl 5 (4-Shot)|14.43|
202
+ |GPQA (0-shot) |12.53|
203
+ |MuSR (0-shot) | 9.77|
204
+ |MMLU-PRO (5-shot) |32.61|
205
+
206
+