OrionStarAI/OrionStar-Yi-34B-Chat-Llama
This model is identical to OrionStarAI/OrionStar-Yi-34B with the only difference being that the tensors have been renamed to follow the LLaMA format for automatic evaluation on the HF leaderboard.
Model Introduction
OrionStar-Yi-34B-Chat from OrionStarAI is based on the open-source Yi-34B model, fine-tuned on a high-quality corpus of over 15 million sentences. OrionStar-Yi-34B-Chat aims to provide an excellent interactive experience for users in the large model community.
The Yi series models, open-sourced by the 01-ai team, have shown impressive performance on various benchmarks in Chinese, English, and general domains. OrionStar-Yi-34B-Chat further explores the potential of Yi-34B. Through extensive fine-tuning on a large and high-quality corpus, OrionStar-Yi-34B-Chat performs exceptionally well on evaluation data. We strive to make it an outstanding open-source alternative in the ChatGPT domain!
Our fine-tuned model is completely open for academic research, but please adhere to the agreement and the Yi License.
Model Evaluation Results
We use opencompass to perform 5-shot on the following general domain datasets Testing. The evaluation results of other models are taken from opencompass leaderboard.
C-Eval | MMLU | CMMLU | |
---|---|---|---|
GPT-4 | 69.9 | 83 | 71 |
ChatGPT | 52.5 | 69.1 | 53.9 |
Claude-1 | 52 | 65.7 | - |
TigerBot-70B-Chat-V2 | 57.7 | 65.9 | 59.9 |
WeMix-LLaMA2-70B | 55.2 | 71.3 | 56 |
LLaMA-2-70B-Chat | 44.3 | 63.8 | 43.3 |
Qwen-14B-Chat | 71.7 | 66.4 | 70 |
Baichuan2-13B-Chat | 56.7 | 57 | 58.4 |
OrionStar-Yi-34B-Chat | 77.71 | 78.32 | 73.52 |
Discord Link: https://discord.gg/zumjDWgdAs
- Downloads last month
- 1,252