Triangle104 commited on
Commit
2d9d59b
·
verified ·
1 Parent(s): 61113e2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +65 -0
README.md CHANGED
@@ -113,6 +113,71 @@ model-index:
113
  This model was converted to GGUF format from [`Locutusque/Hercules-6.1-Llama-3.1-8B`](https://huggingface.co/Locutusque/Hercules-6.1-Llama-3.1-8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
114
  Refer to the [original model card](https://huggingface.co/Locutusque/Hercules-6.1-Llama-3.1-8B) for more details on the model.
115
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
116
  ## Use with llama.cpp
117
  Install llama.cpp through brew (works on Mac and Linux)
118
 
 
113
  This model was converted to GGUF format from [`Locutusque/Hercules-6.1-Llama-3.1-8B`](https://huggingface.co/Locutusque/Hercules-6.1-Llama-3.1-8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
114
  Refer to the [original model card](https://huggingface.co/Locutusque/Hercules-6.1-Llama-3.1-8B) for more details on the model.
115
 
116
+ ---
117
+ Model details:
118
+ -
119
+
120
+ Hercules-6.1-Llama-3.1-8B is a fine-tuned language model derived from Llama-3.1-8B. It is specifically designed to excel in instruction following, function calls, and conversational interactions across various scientific and technical domains. This fine-tuning has hercules-v6.1 with enhanced abilities in:
121
+
122
+ Complex Instruction Following: Understanding and accurately executing multi-step instructions, even those involving specialized terminology.
123
+ Function Calling: Seamlessly interpreting and executing function calls, providing appropriate input and output values.
124
+ Domain-Specific Knowledge: Engaging in informative and educational conversations about Biology, Chemistry, Physics, Mathematics, Medicine, Computer Science, and more.
125
+
126
+ Intended Uses & Potential Bias
127
+
128
+ Hercules-6.1-Llama-3.1-8B is well-suited to the following applications:
129
+
130
+ Specialized Chatbots: Creating knowledgeable chatbots and conversational agents in scientific and technical fields.
131
+ Instructional Assistants: Supporting users with educational and step-by-step guidance in various disciplines.
132
+ Code Generation and Execution: Facilitating code execution through function calls, aiding in software development and prototyping.
133
+
134
+ Important Note: Although Hercules-v6.1 is carefully constructed, it's important to be aware that the underlying data sources may contain biases or reflect harmful stereotypes. Use this model with caution and consider additional measures to mitigate potential biases in its responses.
135
+ Limitations and Risks
136
+
137
+ Toxicity: The dataset contains toxic or harmful examples.
138
+ Hallucinations and Factual Errors: Like other language models, Llama-3-Hercules-6.0-8B may generate incorrect or misleading information, especially in specialized domains where it lacks sufficient expertise.
139
+ Potential for Misuse: The ability to engage in technical conversations and execute function calls could be misused for malicious purposes.
140
+
141
+ Evaluations
142
+ Tasks Version Filter n-shot Metric Value Stderr
143
+ agieval_nous 0.0 none acc ↑ 0.4427 ± 0.0094
144
+ - agieval_aqua_rat 1.0 none 0 acc ↑ 0.2913 ± 0.0286
145
+ none 0 acc_norm ↑ 0.2480 ± 0.0272
146
+ - agieval_logiqa_en 1.0 none 0 acc ↑ 0.3825 ± 0.0191
147
+ none 0 acc_norm ↑ 0.3794 ± 0.0190
148
+ - agieval_lsat_ar 1.0 none 0 acc ↑ 0.2087 ± 0.0269
149
+ none 0 acc_norm ↑ 0.2043 ± 0.0266
150
+ - agieval_lsat_lr 1.0 none 0 acc ↑ 0.4431 ± 0.0220
151
+ none 0 acc_norm ↑ 0.4000 ± 0.0217
152
+ - agieval_lsat_rc 1.0 none 0 acc ↑ 0.6097 ± 0.0298
153
+ none 0 acc_norm ↑ 0.5428 ± 0.0304
154
+ - agieval_sat_en 1.0 none 0 acc ↑ 0.7621 ± 0.0297
155
+ none 0 acc_norm ↑ 0.6942 ± 0.0322
156
+ - agieval_sat_en_without_passage 1.0 none 0 acc ↑ 0.4126 ± 0.0344
157
+ none 0 acc_norm ↑ 0.3641 ± 0.0336
158
+ - agieval_sat_math 1.0 none 0 acc ↑ 0.4318 ± 0.0335
159
+ none 0 acc_norm ↑ 0.3500 ± 0.0322
160
+ arc_challenge 1.0 none 0 acc ↑ 0.5247 ± 0.0146
161
+ none 0 acc_norm ↑ 0.5606 ± 0.0145
162
+ eq_bench 2.1 none 0 eqbench ↑ 63.2023 ± 2.6818
163
+ none 0 percent_parseable ↑ 98.8304 ± 0.8246
164
+ gsm8k 3.0 flexible-extract 5 exact_match ↑ 0.7801 ± 0.0114
165
+ strict-match 5 exact_match ↑ 0.7809 ± 0.0114
166
+ truthfulqa_mc2 2.0 none 0 acc ↑ 0.5389 ± 0.0150
167
+ Open LLM Leaderboard Evaluation Results
168
+
169
+ Detailed results can be found here
170
+ Metric Value
171
+ Avg. 22.40
172
+ IFEval (0-Shot) 60.07
173
+ BBH (3-Shot) 24.15
174
+ MATH Lvl 5 (4-Shot) 15.63
175
+ GPQA (0-shot) 1.45
176
+ MuSR (0-shot) 3.42
177
+ MMLU-PRO (5-shot) 29.65
178
+
179
+ ---
180
+
181
  ## Use with llama.cpp
182
  Install llama.cpp through brew (works on Mac and Linux)
183