Update README.md
Browse files
README.md
CHANGED
|
@@ -9,7 +9,6 @@ tags:
|
|
| 9 |
- merge
|
| 10 |
- custom-research
|
| 11 |
base_model:
|
| 12 |
-
- Qwen/Qwen2.5-14B-Instruct
|
| 13 |
- sthenno-com/miscii-14b-1225
|
| 14 |
- sthenno/tempesthenno-ppo-ckpt40
|
| 15 |
- sometimesanotion/Qwenvergence-14B-v10
|
|
@@ -31,7 +30,8 @@ model-index:
|
|
| 31 |
value: 70.04
|
| 32 |
name: strict accuracy
|
| 33 |
source:
|
| 34 |
-
url:
|
|
|
|
| 35 |
name: Open LLM Leaderboard
|
| 36 |
- task:
|
| 37 |
type: text-generation
|
|
@@ -46,7 +46,8 @@ model-index:
|
|
| 46 |
value: 49.28
|
| 47 |
name: normalized accuracy
|
| 48 |
source:
|
| 49 |
-
url:
|
|
|
|
| 50 |
name: Open LLM Leaderboard
|
| 51 |
- task:
|
| 52 |
type: text-generation
|
|
@@ -61,7 +62,8 @@ model-index:
|
|
| 61 |
value: 39.27
|
| 62 |
name: exact match
|
| 63 |
source:
|
| 64 |
-
url:
|
|
|
|
| 65 |
name: Open LLM Leaderboard
|
| 66 |
- task:
|
| 67 |
type: text-generation
|
|
@@ -76,7 +78,8 @@ model-index:
|
|
| 76 |
value: 18.68
|
| 77 |
name: acc_norm
|
| 78 |
source:
|
| 79 |
-
url:
|
|
|
|
| 80 |
name: Open LLM Leaderboard
|
| 81 |
- task:
|
| 82 |
type: text-generation
|
|
@@ -91,7 +94,8 @@ model-index:
|
|
| 91 |
value: 20.21
|
| 92 |
name: acc_norm
|
| 93 |
source:
|
| 94 |
-
url:
|
|
|
|
| 95 |
name: Open LLM Leaderboard
|
| 96 |
- task:
|
| 97 |
type: text-generation
|
|
@@ -108,7 +112,8 @@ model-index:
|
|
| 108 |
value: 48.36
|
| 109 |
name: accuracy
|
| 110 |
source:
|
| 111 |
-
url:
|
|
|
|
| 112 |
name: Open LLM Leaderboard
|
| 113 |
---
|
| 114 |
|
|
@@ -233,5 +238,4 @@ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-le
|
|
| 233 |
|MATH Lvl 5 (4-Shot)|39.27|
|
| 234 |
|GPQA (0-shot) |18.68|
|
| 235 |
|MuSR (0-shot) |20.21|
|
| 236 |
-
|MMLU-PRO (5-shot) |48.36|
|
| 237 |
-
|
|
|
|
| 9 |
- merge
|
| 10 |
- custom-research
|
| 11 |
base_model:
|
|
|
|
| 12 |
- sthenno-com/miscii-14b-1225
|
| 13 |
- sthenno/tempesthenno-ppo-ckpt40
|
| 14 |
- sometimesanotion/Qwenvergence-14B-v10
|
|
|
|
| 30 |
value: 70.04
|
| 31 |
name: strict accuracy
|
| 32 |
source:
|
| 33 |
+
url: >-
|
| 34 |
+
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno/tempesthenno-nuslerp-0124
|
| 35 |
name: Open LLM Leaderboard
|
| 36 |
- task:
|
| 37 |
type: text-generation
|
|
|
|
| 46 |
value: 49.28
|
| 47 |
name: normalized accuracy
|
| 48 |
source:
|
| 49 |
+
url: >-
|
| 50 |
+
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno/tempesthenno-nuslerp-0124
|
| 51 |
name: Open LLM Leaderboard
|
| 52 |
- task:
|
| 53 |
type: text-generation
|
|
|
|
| 62 |
value: 39.27
|
| 63 |
name: exact match
|
| 64 |
source:
|
| 65 |
+
url: >-
|
| 66 |
+
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno/tempesthenno-nuslerp-0124
|
| 67 |
name: Open LLM Leaderboard
|
| 68 |
- task:
|
| 69 |
type: text-generation
|
|
|
|
| 78 |
value: 18.68
|
| 79 |
name: acc_norm
|
| 80 |
source:
|
| 81 |
+
url: >-
|
| 82 |
+
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno/tempesthenno-nuslerp-0124
|
| 83 |
name: Open LLM Leaderboard
|
| 84 |
- task:
|
| 85 |
type: text-generation
|
|
|
|
| 94 |
value: 20.21
|
| 95 |
name: acc_norm
|
| 96 |
source:
|
| 97 |
+
url: >-
|
| 98 |
+
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno/tempesthenno-nuslerp-0124
|
| 99 |
name: Open LLM Leaderboard
|
| 100 |
- task:
|
| 101 |
type: text-generation
|
|
|
|
| 112 |
value: 48.36
|
| 113 |
name: accuracy
|
| 114 |
source:
|
| 115 |
+
url: >-
|
| 116 |
+
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno/tempesthenno-nuslerp-0124
|
| 117 |
name: Open LLM Leaderboard
|
| 118 |
---
|
| 119 |
|
|
|
|
| 238 |
|MATH Lvl 5 (4-Shot)|39.27|
|
| 239 |
|GPQA (0-shot) |18.68|
|
| 240 |
|MuSR (0-shot) |20.21|
|
| 241 |
+
|MMLU-PRO (5-shot) |48.36|
|
|
|