djstrong commited on
Commit
6e77fb1
·
verified ·
1 Parent(s): 9e510b2

update results

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -149,7 +149,7 @@ Models have been evaluated on [Open PL LLM Leaderboard](https://huggingface.co/s
149
  | Meta-Llama-3.1-405B-Instruct-FP8,API | 405 | 69.44 |
150
  | Mistral-Large-Instruct-2407 | 123 | 69.11 |
151
  | Qwen2-72B-Instruct | 72 | 65.87 |
152
- | **Bielik-11B-v2.3-Instruct** | **11** | **~~65.57~~** |
153
  | Bielik-11B-v2.2-Instruct | 11 | 65.57 |
154
  | Meta-Llama-3.1-70B-Instruct | 70 | 65.49 |
155
  | Bielik-11B-v2.1-Instruct | 11 | 65.45 |
@@ -189,7 +189,7 @@ This section presents a focused comparison of generative Polish language task pe
189
 
190
  | Model | Parameters (B) | Average g |
191
  |-------------------------------|----------------|---------------|
192
- | **Bielik-11B-v2.3-Instruct** | 11 | **~~66.58~~**
193
  | Bielik-11B-v2.1-Instruct | 11 | 66.58 |
194
  | Bielik-11B-v2.2-Instruct | 11 | 66.11 |
195
  | Bielik-11B-v2.0-Instruct | 11 | 65.58 |
@@ -280,12 +280,12 @@ For more information - answers to test tasks and values in each category, visit
280
  | Qwen2-72B-Instruct | 72 | 71.23 |
281
  | Meta-Llama-3-70B-Instruct | 70 | 71.21 |
282
  | gpt-4o-mini-2024-07-18 | ? | 71.15 |
 
283
  | WizardLM-2-8x22B | 141 | 69.56 |
284
  | Bielik-11B-v2.2-Instruct | 11 | 69.05 |
285
  | Bielik-11B-v2.0-Instruct | 11 | 68.24 |
286
  | Qwen1.5-72B-Chat | 72 | 68.03 |
287
  | Mixtral-8x22B-Instruct-v0.1 | 141 | 67.63 |
288
- | **Bielik-11B-v2.3-Instruct** | **11** | **~~60.07~~** |
289
  | Bielik-11B-v2.1-Instruct | 11 | 60.07 |
290
  | Qwen1.5-32B-Chat | 32 | 59.63 |
291
  | openchat-3.5-0106-gemma | 7 | 59.58 |
 
149
  | Meta-Llama-3.1-405B-Instruct-FP8,API | 405 | 69.44 |
150
  | Mistral-Large-Instruct-2407 | 123 | 69.11 |
151
  | Qwen2-72B-Instruct | 72 | 65.87 |
152
+ | **Bielik-11B-v2.3-Instruct** | **11** | **65.71** |
153
  | Bielik-11B-v2.2-Instruct | 11 | 65.57 |
154
  | Meta-Llama-3.1-70B-Instruct | 70 | 65.49 |
155
  | Bielik-11B-v2.1-Instruct | 11 | 65.45 |
 
189
 
190
  | Model | Parameters (B) | Average g |
191
  |-------------------------------|----------------|---------------|
192
+ | **Bielik-11B-v2.3-Instruct** | 11 | **67.47**
193
  | Bielik-11B-v2.1-Instruct | 11 | 66.58 |
194
  | Bielik-11B-v2.2-Instruct | 11 | 66.11 |
195
  | Bielik-11B-v2.0-Instruct | 11 | 65.58 |
 
280
  | Qwen2-72B-Instruct | 72 | 71.23 |
281
  | Meta-Llama-3-70B-Instruct | 70 | 71.21 |
282
  | gpt-4o-mini-2024-07-18 | ? | 71.15 |
283
+ | **Bielik-11B-v2.3-Instruct** | **11** | **70.86** |
284
  | WizardLM-2-8x22B | 141 | 69.56 |
285
  | Bielik-11B-v2.2-Instruct | 11 | 69.05 |
286
  | Bielik-11B-v2.0-Instruct | 11 | 68.24 |
287
  | Qwen1.5-72B-Chat | 72 | 68.03 |
288
  | Mixtral-8x22B-Instruct-v0.1 | 141 | 67.63 |
 
289
  | Bielik-11B-v2.1-Instruct | 11 | 60.07 |
290
  | Qwen1.5-32B-Chat | 32 | 59.63 |
291
  | openchat-3.5-0106-gemma | 7 | 59.58 |