|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- meta-math/MetaMathQA |
|
language: |
|
- en |
|
--- |
|
|
|
Supervised Finetuning the phi1.5 on MetaMathQA datasets. The results are as follows: |
|
|
|
| Model | GSM8k Pass@1 | MATH Pass@1 | |
|
|---------------------|--------------|-------------| |
|
| MPT-7B | 6.8 | 3.0 | |
|
| Falcon-7B | 6.8 | 2.3 | |
|
| LLaMA-1-7B | 11.0 | 2.9 | |
|
| LLaMA-2-7B | 14.6 | 2.5 | |
|
| MPT-30B | 15.2 | 3.1 | |
|
| LLaMA-1-13B | 17.8 | 3.9 | |
|
| GPT-Neo-2.7B | 19.5 | -- | |
|
| Falcon-40B | 19.6 | 2.5 | |
|
| Baichuan-chat-13B | 23.9 | -- | |
|
| Vicuna-v1.3-13B | 27.6 | -- | |
|
| LLaMA-2-13B | 28.7 | 3.9 | |
|
| InternLM-7B | 31.2 | -- | |
|
| ChatGLM-2-6B | 32.4 | -- | |
|
| GPT-J-6B | 34.9 | -- | |
|
| LLaMA-1-33B | 35.6 | 3.9 | |
|
| LLaMA-2-34B | 42.2 | 6.24 | |
|
| RFT-7B | 50.3 | -- | |
|
| LLaMA-1-65B | 50.9 | 10.6 | |
|
| Qwen-7B | 51.6 | -- | |
|
| **Phi1.5-1.3B** | **54.3** | **15.5** | |
|
| WizardMath-7B | 54.9 | 10.7 | |
|
| LLaMA-2-70B | 56.8 | 13.5 | |
|
| WizardMath-13B | 63.9 | 14.0 | |
|
| MAmmoTH-7B (COT) | 50.5 | 10.4 | |
|
| MAmmoTH-7B (POT+COT)| 53.6 | 31.5 | |
|
| Arithmo-Mistral-7B | 74.7 | 25.3 | |
|
| MetaMath-7B | 66.5 | 19.8 | |
|
| MetaMath-13B | 72.3 | 22.4 | |
|
| **MetaMath-Mistral-7B** | **77.7** | **28.2** | |
|
|
|
|
|
It achieves remarkable performance with only 1.3B parameters !!! |
|
|
|
You can evaluate the results by [metamath](https://huggingface.co/meta-math/MetaMath-Mistral-7B) evaluation code. |