Update README.md
Browse files
README.md
CHANGED
@@ -13,20 +13,18 @@ base_model:
|
|
13 |
|
14 |
# Chimera-8B
|
15 |
|
|
|
|
|
16 |
Chimera-8B is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
|
17 |
* [NousResearch/Meta-Llama-3-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3-8B-Instruct)
|
18 |
* [mlabonne/OrpoLlama-3-8B](https://huggingface.co/mlabonne/OrpoLlama-3-8B)
|
19 |
* [Locutusque/Llama-3-Orca-1.0-8B](https://huggingface.co/Locutusque/Llama-3-Orca-1.0-8B)
|
20 |
* [abacusai/Llama-3-Smaug-8B](https://huggingface.co/abacusai/Llama-3-Smaug-8B)
|
21 |
|
22 |
-
|
23 |
-
|
24 |
## 🏆 Evaluation
|
25 |
|
26 |
### Nous
|
27 |
|
28 |
-
OrpoLlama-4-8B outperforms Llama-3-8B-Instruct on the GPT4All and TruthfulQA datasets.
|
29 |
-
|
30 |
Evaluation performed using [LLM AutoEval](https://github.com/mlabonne/llm-autoeval), see the entire leaderboard [here](https://huggingface.co/spaces/mlabonne/Yet_Another_LLM_Leaderboard).
|
31 |
|
32 |
| Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
|
|
|
13 |
|
14 |
# Chimera-8B
|
15 |
|
16 |
+
Chimera-8B outperforms Llama 3 8B Instruct on Nous' benchmark suite.
|
17 |
+
|
18 |
Chimera-8B is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
|
19 |
* [NousResearch/Meta-Llama-3-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3-8B-Instruct)
|
20 |
* [mlabonne/OrpoLlama-3-8B](https://huggingface.co/mlabonne/OrpoLlama-3-8B)
|
21 |
* [Locutusque/Llama-3-Orca-1.0-8B](https://huggingface.co/Locutusque/Llama-3-Orca-1.0-8B)
|
22 |
* [abacusai/Llama-3-Smaug-8B](https://huggingface.co/abacusai/Llama-3-Smaug-8B)
|
23 |
|
|
|
|
|
24 |
## 🏆 Evaluation
|
25 |
|
26 |
### Nous
|
27 |
|
|
|
|
|
28 |
Evaluation performed using [LLM AutoEval](https://github.com/mlabonne/llm-autoeval), see the entire leaderboard [here](https://huggingface.co/spaces/mlabonne/Yet_Another_LLM_Leaderboard).
|
29 |
|
30 |
| Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
|