license: other | |
language: | |
- en | |
- zh | |
library_name: transformers | |
pipeline_tag: text-generation | |
inference: false | |
tags: | |
- baichuan | |
- llama2 | |
- baichuan2 | |
This is the LLaMAfied version of [Baichuan2-7B-Base](https://huggingface.co/baichuan-inc/Baichuan2-7B-Base) model by Baichuan Inc. | |
This model is converted with https://github.com/hiyouga/LLaMA-Factory/blob/main/tests/llamafy_baichuan2.py | |
You may use this model for fine-tuning in downstream tasks, we recommend using our efficient fine-tuning toolkit. https://github.com/hiyouga/LLaMA-Factory | |
- **Developed by:** Baichuan Inc. | |
- **Language(s) (NLP):** Chinese/English | |
- **License:** [Baichuan2 License](https://huggingface.co/baichuan-inc/Baichuan2-7B-Base/resolve/main/Baichuan%202%E6%A8%A1%E5%9E%8B%E7%A4%BE%E5%8C%BA%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE.pdf) | |
Usage: | |
```python | |
from transformers import AutoModelForCausalLM, AutoTokenizer | |
tokenizer = AutoTokenizer.from_pretrained("hiyouga/Baichuan2-7B-Base-LLaMAfied", use_fast=False) | |
model = AutoModelForCausalLM.from_pretrained("hiyouga/Baichuan2-7B-Base-LLaMAfied").cuda() | |
``` | |
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) | |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_hiyouga__Baichuan2-7B-Base-LLaMAfied) | |
| Metric | Value | | |
|-----------------------|---------------------------| | |
| Avg. | 42.83 | | |
| ARC (25-shot) | 49.57 | | |
| HellaSwag (10-shot) | 73.45 | | |
| MMLU (5-shot) | 54.86 | | |
| TruthfulQA (0-shot) | 37.54 | | |
| Winogrande (5-shot) | 70.72 | | |
| GSM8K (5-shot) | 7.81 | | |
| DROP (3-shot) | 5.85 | | |