Update README.md
Browse files
README.md
CHANGED
@@ -59,10 +59,11 @@ We achieve the following MT-Bench scores across 6 languages:
|
|
59 |
| **German** π©πͺ | NaN | 7.26 | 6.99 | 7.68 |
|
60 |
| **French** π«π· | NaN | 7.66 | 7.29 | 7.74 |
|
61 |
| **Japanese** π―π΅ | NaN | 6.56 | 6.22 | 7.84 |
|
62 |
-
| **Russian** π·πΊ
|
63 |
| **Chinese** π¨π³ | NaN | 7.11 | 6.97 | 7.55 |
|
64 |
| **English** πΊπΈ | 7.98 | 7.73 | 7.92 | 8.26 |
|
65 |
-
|
|
|
66 |
|
67 |
We observe minimal degredation of Llama 3's English ability while achieving best-in-class multilingual abilities compared to the top rated 7B model ([Nexusflow/Starling-LM-7B-beta](https://huggingface.co/Nexusflow/Starling-LM-7B-beta)) on the [Chatbot Arena Leaderboard](https://chat.lmsys.org/?leaderboard).
|
68 |
|
|
|
59 |
| **German** π©πͺ | NaN | 7.26 | 6.99 | 7.68 |
|
60 |
| **French** π«π· | NaN | 7.66 | 7.29 | 7.74 |
|
61 |
| **Japanese** π―π΅ | NaN | 6.56 | 6.22 | 7.84 |
|
62 |
+
| **Russian** π·πΊ * | NaN | 8.19 | 8.28 | 7.94 |
|
63 |
| **Chinese** π¨π³ | NaN | 7.11 | 6.97 | 7.55 |
|
64 |
| **English** πΊπΈ | 7.98 | 7.73 | 7.92 | 8.26 |
|
65 |
+
|
66 |
+
\* (Note the Russian scores exclude code, reasoning and math problems due to not having any translated reference answers for these questions.)
|
67 |
|
68 |
We observe minimal degredation of Llama 3's English ability while achieving best-in-class multilingual abilities compared to the top rated 7B model ([Nexusflow/Starling-LM-7B-beta](https://huggingface.co/Nexusflow/Starling-LM-7B-beta)) on the [Chatbot Arena Leaderboard](https://chat.lmsys.org/?leaderboard).
|
69 |
|