itlwas commited on
Commit
9210118
1 Parent(s): d3f64c2

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +301 -0
README.md ADDED
@@ -0,0 +1,301 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: other
5
+ tags:
6
+ - axolotl
7
+ - generated_from_trainer
8
+ - instruct
9
+ - finetune
10
+ - chatml
11
+ - gpt4
12
+ - synthetic data
13
+ - science
14
+ - physics
15
+ - chemistry
16
+ - biology
17
+ - math
18
+ - llama
19
+ - llama3
20
+ - llama-cpp
21
+ - gguf-my-repo
22
+ base_model: Weyaxi/Einstein-v6.1-Llama3-8B
23
+ datasets:
24
+ - allenai/ai2_arc
25
+ - camel-ai/physics
26
+ - camel-ai/chemistry
27
+ - camel-ai/biology
28
+ - camel-ai/math
29
+ - metaeval/reclor
30
+ - openbookqa
31
+ - mandyyyyii/scibench
32
+ - derek-thomas/ScienceQA
33
+ - TIGER-Lab/ScienceEval
34
+ - jondurbin/airoboros-3.2
35
+ - LDJnr/Capybara
36
+ - Cot-Alpaca-GPT4-From-OpenHermes-2.5
37
+ - STEM-AI-mtl/Electrical-engineering
38
+ - knowrohit07/saraswati-stem
39
+ - sablo/oasst2_curated
40
+ - lmsys/lmsys-chat-1m
41
+ - TIGER-Lab/MathInstruct
42
+ - bigbio/med_qa
43
+ - meta-math/MetaMathQA-40K
44
+ - openbookqa
45
+ - piqa
46
+ - metaeval/reclor
47
+ - derek-thomas/ScienceQA
48
+ - scibench
49
+ - sciq
50
+ - Open-Orca/SlimOrca
51
+ - migtissera/Synthia-v1.3
52
+ - TIGER-Lab/ScienceEval
53
+ - allenai/WildChat
54
+ - microsoft/orca-math-word-problems-200k
55
+ - openchat/openchat_sharegpt4_dataset
56
+ - teknium/GPTeacher-General-Instruct
57
+ - m-a-p/CodeFeedback-Filtered-Instruction
58
+ - totally-not-an-llm/EverythingLM-data-V3
59
+ - HuggingFaceH4/no_robots
60
+ - OpenAssistant/oasst_top1_2023-08-25
61
+ - WizardLM/WizardLM_evol_instruct_70k
62
+ model-index:
63
+ - name: Einstein-v6.1-Llama3-8B
64
+ results:
65
+ - task:
66
+ type: text-generation
67
+ name: Text Generation
68
+ dataset:
69
+ name: AI2 Reasoning Challenge (25-Shot)
70
+ type: ai2_arc
71
+ config: ARC-Challenge
72
+ split: test
73
+ args:
74
+ num_few_shot: 25
75
+ metrics:
76
+ - type: acc_norm
77
+ value: 62.46
78
+ name: normalized accuracy
79
+ source:
80
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
81
+ name: Open LLM Leaderboard
82
+ - task:
83
+ type: text-generation
84
+ name: Text Generation
85
+ dataset:
86
+ name: HellaSwag (10-Shot)
87
+ type: hellaswag
88
+ split: validation
89
+ args:
90
+ num_few_shot: 10
91
+ metrics:
92
+ - type: acc_norm
93
+ value: 82.41
94
+ name: normalized accuracy
95
+ source:
96
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
97
+ name: Open LLM Leaderboard
98
+ - task:
99
+ type: text-generation
100
+ name: Text Generation
101
+ dataset:
102
+ name: MMLU (5-Shot)
103
+ type: cais/mmlu
104
+ config: all
105
+ split: test
106
+ args:
107
+ num_few_shot: 5
108
+ metrics:
109
+ - type: acc
110
+ value: 66.19
111
+ name: accuracy
112
+ source:
113
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
114
+ name: Open LLM Leaderboard
115
+ - task:
116
+ type: text-generation
117
+ name: Text Generation
118
+ dataset:
119
+ name: TruthfulQA (0-shot)
120
+ type: truthful_qa
121
+ config: multiple_choice
122
+ split: validation
123
+ args:
124
+ num_few_shot: 0
125
+ metrics:
126
+ - type: mc2
127
+ value: 55.1
128
+ source:
129
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
130
+ name: Open LLM Leaderboard
131
+ - task:
132
+ type: text-generation
133
+ name: Text Generation
134
+ dataset:
135
+ name: Winogrande (5-shot)
136
+ type: winogrande
137
+ config: winogrande_xl
138
+ split: validation
139
+ args:
140
+ num_few_shot: 5
141
+ metrics:
142
+ - type: acc
143
+ value: 79.32
144
+ name: accuracy
145
+ source:
146
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
147
+ name: Open LLM Leaderboard
148
+ - task:
149
+ type: text-generation
150
+ name: Text Generation
151
+ dataset:
152
+ name: GSM8k (5-shot)
153
+ type: gsm8k
154
+ config: main
155
+ split: test
156
+ args:
157
+ num_few_shot: 5
158
+ metrics:
159
+ - type: acc
160
+ value: 66.11
161
+ name: accuracy
162
+ source:
163
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
164
+ name: Open LLM Leaderboard
165
+ - task:
166
+ type: text-generation
167
+ name: Text Generation
168
+ dataset:
169
+ name: IFEval (0-Shot)
170
+ type: HuggingFaceH4/ifeval
171
+ args:
172
+ num_few_shot: 0
173
+ metrics:
174
+ - type: inst_level_strict_acc and prompt_level_strict_acc
175
+ value: 45.68
176
+ name: strict accuracy
177
+ source:
178
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
179
+ name: Open LLM Leaderboard
180
+ - task:
181
+ type: text-generation
182
+ name: Text Generation
183
+ dataset:
184
+ name: BBH (3-Shot)
185
+ type: BBH
186
+ args:
187
+ num_few_shot: 3
188
+ metrics:
189
+ - type: acc_norm
190
+ value: 29.38
191
+ name: normalized accuracy
192
+ source:
193
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
194
+ name: Open LLM Leaderboard
195
+ - task:
196
+ type: text-generation
197
+ name: Text Generation
198
+ dataset:
199
+ name: MATH Lvl 5 (4-Shot)
200
+ type: hendrycks/competition_math
201
+ args:
202
+ num_few_shot: 4
203
+ metrics:
204
+ - type: exact_match
205
+ value: 5.74
206
+ name: exact match
207
+ source:
208
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
209
+ name: Open LLM Leaderboard
210
+ - task:
211
+ type: text-generation
212
+ name: Text Generation
213
+ dataset:
214
+ name: GPQA (0-shot)
215
+ type: Idavidrein/gpqa
216
+ args:
217
+ num_few_shot: 0
218
+ metrics:
219
+ - type: acc_norm
220
+ value: 4.25
221
+ name: acc_norm
222
+ source:
223
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
224
+ name: Open LLM Leaderboard
225
+ - task:
226
+ type: text-generation
227
+ name: Text Generation
228
+ dataset:
229
+ name: MuSR (0-shot)
230
+ type: TAUR-Lab/MuSR
231
+ args:
232
+ num_few_shot: 0
233
+ metrics:
234
+ - type: acc_norm
235
+ value: 11.23
236
+ name: acc_norm
237
+ source:
238
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
239
+ name: Open LLM Leaderboard
240
+ - task:
241
+ type: text-generation
242
+ name: Text Generation
243
+ dataset:
244
+ name: MMLU-PRO (5-shot)
245
+ type: TIGER-Lab/MMLU-Pro
246
+ config: main
247
+ split: test
248
+ args:
249
+ num_few_shot: 5
250
+ metrics:
251
+ - type: acc
252
+ value: 23.68
253
+ name: accuracy
254
+ source:
255
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
256
+ name: Open LLM Leaderboard
257
+ ---
258
+
259
+ # AIronMind/Einstein-v6.1-Llama3-8B-Q4_K_M-GGUF
260
+ This model was converted to GGUF format from [`Weyaxi/Einstein-v6.1-Llama3-8B`](https://huggingface.co/Weyaxi/Einstein-v6.1-Llama3-8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
261
+ Refer to the [original model card](https://huggingface.co/Weyaxi/Einstein-v6.1-Llama3-8B) for more details on the model.
262
+
263
+ ## Use with llama.cpp
264
+ Install llama.cpp through brew (works on Mac and Linux)
265
+
266
+ ```bash
267
+ brew install llama.cpp
268
+
269
+ ```
270
+ Invoke the llama.cpp server or the CLI.
271
+
272
+ ### CLI:
273
+ ```bash
274
+ llama-cli --hf-repo AIronMind/Einstein-v6.1-Llama3-8B-Q4_K_M-GGUF --hf-file einstein-v6.1-llama3-8b-q4_k_m.gguf -p "The meaning to life and the universe is"
275
+ ```
276
+
277
+ ### Server:
278
+ ```bash
279
+ llama-server --hf-repo AIronMind/Einstein-v6.1-Llama3-8B-Q4_K_M-GGUF --hf-file einstein-v6.1-llama3-8b-q4_k_m.gguf -c 2048
280
+ ```
281
+
282
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
283
+
284
+ Step 1: Clone llama.cpp from GitHub.
285
+ ```
286
+ git clone https://github.com/ggerganov/llama.cpp
287
+ ```
288
+
289
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
290
+ ```
291
+ cd llama.cpp && LLAMA_CURL=1 make
292
+ ```
293
+
294
+ Step 3: Run inference through the main binary.
295
+ ```
296
+ ./llama-cli --hf-repo AIronMind/Einstein-v6.1-Llama3-8B-Q4_K_M-GGUF --hf-file einstein-v6.1-llama3-8b-q4_k_m.gguf -p "The meaning to life and the universe is"
297
+ ```
298
+ or
299
+ ```
300
+ ./llama-server --hf-repo AIronMind/Einstein-v6.1-Llama3-8B-Q4_K_M-GGUF --hf-file einstein-v6.1-llama3-8b-q4_k_m.gguf -c 2048
301
+ ```