update results
Browse files
README.md
CHANGED
@@ -149,7 +149,7 @@ Models have been evaluated on [Open PL LLM Leaderboard](https://huggingface.co/s
|
|
149 |
| Meta-Llama-3.1-405B-Instruct-FP8,API | 405 | 69.44 |
|
150 |
| Mistral-Large-Instruct-2407 | 123 | 69.11 |
|
151 |
| Qwen2-72B-Instruct | 72 | 65.87 |
|
152 |
-
| **Bielik-11B-v2.3-Instruct** | **11** |
|
153 |
| Bielik-11B-v2.2-Instruct | 11 | 65.57 |
|
154 |
| Meta-Llama-3.1-70B-Instruct | 70 | 65.49 |
|
155 |
| Bielik-11B-v2.1-Instruct | 11 | 65.45 |
|
@@ -189,7 +189,7 @@ This section presents a focused comparison of generative Polish language task pe
|
|
189 |
|
190 |
| Model | Parameters (B) | Average g |
|
191 |
|-------------------------------|----------------|---------------|
|
192 |
-
| **Bielik-11B-v2.3-Instruct** | 11 |
|
193 |
| Bielik-11B-v2.1-Instruct | 11 | 66.58 |
|
194 |
| Bielik-11B-v2.2-Instruct | 11 | 66.11 |
|
195 |
| Bielik-11B-v2.0-Instruct | 11 | 65.58 |
|
@@ -280,12 +280,12 @@ For more information - answers to test tasks and values in each category, visit
|
|
280 |
| Qwen2-72B-Instruct | 72 | 71.23 |
|
281 |
| Meta-Llama-3-70B-Instruct | 70 | 71.21 |
|
282 |
| gpt-4o-mini-2024-07-18 | ? | 71.15 |
|
|
|
283 |
| WizardLM-2-8x22B | 141 | 69.56 |
|
284 |
| Bielik-11B-v2.2-Instruct | 11 | 69.05 |
|
285 |
| Bielik-11B-v2.0-Instruct | 11 | 68.24 |
|
286 |
| Qwen1.5-72B-Chat | 72 | 68.03 |
|
287 |
| Mixtral-8x22B-Instruct-v0.1 | 141 | 67.63 |
|
288 |
-
| **Bielik-11B-v2.3-Instruct** | **11** | **~~60.07~~** |
|
289 |
| Bielik-11B-v2.1-Instruct | 11 | 60.07 |
|
290 |
| Qwen1.5-32B-Chat | 32 | 59.63 |
|
291 |
| openchat-3.5-0106-gemma | 7 | 59.58 |
|
|
|
149 |
| Meta-Llama-3.1-405B-Instruct-FP8,API | 405 | 69.44 |
|
150 |
| Mistral-Large-Instruct-2407 | 123 | 69.11 |
|
151 |
| Qwen2-72B-Instruct | 72 | 65.87 |
|
152 |
+
| **Bielik-11B-v2.3-Instruct** | **11** | **65.71** |
|
153 |
| Bielik-11B-v2.2-Instruct | 11 | 65.57 |
|
154 |
| Meta-Llama-3.1-70B-Instruct | 70 | 65.49 |
|
155 |
| Bielik-11B-v2.1-Instruct | 11 | 65.45 |
|
|
|
189 |
|
190 |
| Model | Parameters (B) | Average g |
|
191 |
|-------------------------------|----------------|---------------|
|
192 |
+
| **Bielik-11B-v2.3-Instruct** | 11 | **67.47**
|
193 |
| Bielik-11B-v2.1-Instruct | 11 | 66.58 |
|
194 |
| Bielik-11B-v2.2-Instruct | 11 | 66.11 |
|
195 |
| Bielik-11B-v2.0-Instruct | 11 | 65.58 |
|
|
|
280 |
| Qwen2-72B-Instruct | 72 | 71.23 |
|
281 |
| Meta-Llama-3-70B-Instruct | 70 | 71.21 |
|
282 |
| gpt-4o-mini-2024-07-18 | ? | 71.15 |
|
283 |
+
| **Bielik-11B-v2.3-Instruct** | **11** | **70.86** |
|
284 |
| WizardLM-2-8x22B | 141 | 69.56 |
|
285 |
| Bielik-11B-v2.2-Instruct | 11 | 69.05 |
|
286 |
| Bielik-11B-v2.0-Instruct | 11 | 68.24 |
|
287 |
| Qwen1.5-72B-Chat | 72 | 68.03 |
|
288 |
| Mixtral-8x22B-Instruct-v0.1 | 141 | 67.63 |
|
|
|
289 |
| Bielik-11B-v2.1-Instruct | 11 | 60.07 |
|
290 |
| Qwen1.5-32B-Chat | 32 | 59.63 |
|
291 |
| openchat-3.5-0106-gemma | 7 | 59.58 |
|