weezywitasneezy
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -120,6 +120,24 @@ Oxide-F1-7B-slerp is a merge of the following models using [LazyMergekit](https:
|
|
120 |
* [weezywitasneezy/BenchmarkEngineering-7B-slerp](https://huggingface.co/weezywitasneezy/BenchmarkEngineering-7B-slerp)
|
121 |
* [weezywitasneezy/OxytocinErosEngineeringFX-7B-slerp](https://huggingface.co/weezywitasneezy/OxytocinErosEngineeringFX-7B-slerp)
|
122 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
123 |
## 🧩 Configuration
|
124 |
|
125 |
```yaml
|
@@ -165,16 +183,3 @@ pipeline = transformers.pipeline(
|
|
165 |
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
166 |
print(outputs[0]["generated_text"])
|
167 |
```
|
168 |
-
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
169 |
-
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_weezywitasneezy__Oxide-F1-7B-slerp)
|
170 |
-
|
171 |
-
| Metric |Value|
|
172 |
-
|---------------------------------|----:|
|
173 |
-
|Avg. |73.64|
|
174 |
-
|AI2 Reasoning Challenge (25-Shot)|69.71|
|
175 |
-
|HellaSwag (10-Shot) |87.82|
|
176 |
-
|MMLU (5-Shot) |64.77|
|
177 |
-
|TruthfulQA (0-shot) |70.00|
|
178 |
-
|Winogrande (5-shot) |81.85|
|
179 |
-
|GSM8k (5-shot) |67.70|
|
180 |
-
|
|
|
120 |
* [weezywitasneezy/BenchmarkEngineering-7B-slerp](https://huggingface.co/weezywitasneezy/BenchmarkEngineering-7B-slerp)
|
121 |
* [weezywitasneezy/OxytocinErosEngineeringFX-7B-slerp](https://huggingface.co/weezywitasneezy/OxytocinErosEngineeringFX-7B-slerp)
|
122 |
|
123 |
+
|
124 |
+
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
125 |
+
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_weezywitasneezy__Oxide-F1-7B-slerp)
|
126 |
+
|
127 |
+
|
128 |
+
| Metric |Value|
|
129 |
+
|---------------------------------|----:|
|
130 |
+
|Avg. |73.64|
|
131 |
+
|AI2 Reasoning Challenge (25-Shot)|69.71|
|
132 |
+
|HellaSwag (10-Shot) |87.82|
|
133 |
+
|MMLU (5-Shot) |64.77|
|
134 |
+
|TruthfulQA (0-shot) |70.00|
|
135 |
+
|Winogrande (5-shot) |81.85|
|
136 |
+
|GSM8k (5-shot) |67.70|
|
137 |
+
|
138 |
+
|
139 |
+
|
140 |
+
|
141 |
## 🧩 Configuration
|
142 |
|
143 |
```yaml
|
|
|
183 |
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
184 |
print(outputs[0]["generated_text"])
|
185 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|