Update README.md
Browse files
README.md
CHANGED
@@ -5,17 +5,20 @@ language:
|
|
5 |
- ja
|
6 |
---
|
7 |
|
8 |
-
# EvoLLM-JP-v1-7B
|
|
|
|
|
|
|
9 |
|
10 |
<!-- Provide a quick summary of what the model is/does. -->
|
11 |
-
EvoLLM-JP-v1-7B is a Japanese Math LLM by Evolutionary Model Merge.
|
12 |
|
13 |
## Model Details
|
14 |
|
15 |
### Model Description
|
16 |
|
17 |
<!-- Provide a longer summary of what this model is. -->
|
18 |
-
EvoLLM-JP-v1-7B is a Japanese Math LLM, merged the following source models in the Parameter Space (PS) by Evolutionary Model Merge.
|
19 |
|
20 |
- **Developed by:** [Sakana AI](https://sakana.ai/)
|
21 |
- **Model type:** Autoregressive Language Model
|
@@ -30,7 +33,7 @@ EvoLLM-JP-v1-7B is a Japanese Math LLM, merged the following source models in th
|
|
30 |
|
31 |
<!-- Provide the basic links for the model. -->
|
32 |
|
33 |
-
- **Repository:** [SakanaAI/
|
34 |
- **Paper:** TODO
|
35 |
- **Blog:** TODO
|
36 |
|
@@ -70,8 +73,9 @@ print(generated_text)
|
|
70 |
```
|
71 |
|
72 |
## Evaluation
|
|
|
73 |
|
74 |
-
|
75 |
|
76 |
| Id. | Model | Type | Params | MGSM-JA (acc ↑ ) |
|
77 |
| :--: | :-- | :-- | --: | --: |
|
@@ -79,9 +83,13 @@ We present the results on the [MGSM-JA](https://huggingface.co/datasets/juletxar
|
|
79 |
| 2 | [WizardMath 7B V1.1](https://huggingface.co/WizardLM/WizardMath-7B-V1.1) | EN math | 7B | 18.4 |
|
80 |
| 3 | [Abel 7B 002](https://huggingface.co/GAIR/Abel-7B-002) | EN math | 7B | 30.0 |
|
81 |
| 4 | [Arithmo2 Mistral 7B](https://huggingface.co/upaya07/Arithmo2-Mistral-7B) | EN math | 7B | 24.0 |
|
82 |
-
| 5 | [
|
83 |
-
| 6 | [
|
84 |
-
| 7 | [
|
|
|
|
|
|
|
|
|
85 |
|
86 |
|
87 |
## Citation
|
|
|
5 |
- ja
|
6 |
---
|
7 |
|
8 |
+
# π EvoLLM-JP-v1-7B
|
9 |
+
|
10 |
+
π€ [Models](https://huggingface.co/SakanaAI) | π [Paper](TODO) | π [Blog](TODO) | π¦ [Twitter](https://twitter.com/SakanaAILabs)
|
11 |
+
|
12 |
|
13 |
<!-- Provide a quick summary of what the model is/does. -->
|
14 |
+
**EvoLLM-JP-v1-7B** is a Japanese Math LLM by Evolutionary Model Merge.
|
15 |
|
16 |
## Model Details
|
17 |
|
18 |
### Model Description
|
19 |
|
20 |
<!-- Provide a longer summary of what this model is. -->
|
21 |
+
**EvoLLM-JP-v1-7B** is a Japanese Math LLM, merged the following source models in the Parameter Space (PS) by Evolutionary Model Merge.
|
22 |
|
23 |
- **Developed by:** [Sakana AI](https://sakana.ai/)
|
24 |
- **Model type:** Autoregressive Language Model
|
|
|
33 |
|
34 |
<!-- Provide the basic links for the model. -->
|
35 |
|
36 |
+
- **Repository:** [SakanaAI/evolutionary-model-merge](https://github.com/SakanaAI/evolutionary-model-merge)
|
37 |
- **Paper:** TODO
|
38 |
- **Blog:** TODO
|
39 |
|
|
|
73 |
```
|
74 |
|
75 |
## Evaluation
|
76 |
+
For details on the evaluation, please refer to Section 4.1 of the paper.
|
77 |
|
78 |
+
If you want to reproduce the results, please see [our Github repository](https://github.com/SakanaAI/evolutionary-model-merge).
|
79 |
|
80 |
| Id. | Model | Type | Params | MGSM-JA (acc ↑ ) |
|
81 |
| :--: | :-- | :-- | --: | --: |
|
|
|
83 |
| 2 | [WizardMath 7B V1.1](https://huggingface.co/WizardLM/WizardMath-7B-V1.1) | EN math | 7B | 18.4 |
|
84 |
| 3 | [Abel 7B 002](https://huggingface.co/GAIR/Abel-7B-002) | EN math | 7B | 30.0 |
|
85 |
| 4 | [Arithmo2 Mistral 7B](https://huggingface.co/upaya07/Arithmo2-Mistral-7B) | EN math | 7B | 24.0 |
|
86 |
+
| 5 | [EvoLLM-JP-v1-7B](https://huggingface.co/SakanaAI/EvoLLM-JP-v1-7B) | 1+2+3 | 7B | **52.0** |
|
87 |
+
| 6 | [EvoLLM-JP-A-v1-7B](https://huggingface.co/SakanaAI/EvoLLM-JP-A-v1-7B) | 1+3+4 | 7B | **52.4** |
|
88 |
+
| 7 | [EvoLLM-JP-v1-10B](https://huggingface.co/SakanaAI/EvoLLM-JP-v1-10B) | 1 + 5 | 10B | **55.6** |
|
89 |
+
|
90 |
+
## Acknowledgement
|
91 |
+
|
92 |
+
We would like to thank the developers of the source models for their contributions and for making their work available.
|
93 |
|
94 |
|
95 |
## Citation
|