Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- neulab/PangeaInstruct
|
5 |
+
---
|
6 |
+
# Pangea-7B Model Card
|
7 |
+
|
8 |
+
[Homepage](https://neulab.github.io/Pangea/) | [Pangea-7B](https://huggingface.co/neulab/Pangea-7B) | [PangeaIns](https://huggingface.co/datasets/neulab/PangeaInstruct)
|
9 |
+
| [PangeaBench](https://huggingface.co/collections/neulab/pangea-6713c3b0d78a453906eb2ed8) | [Github](https://github.com/neulab/Pangea/tree/main)
|
10 |
+
| [Arxiv](https://arxiv.org/abs/2410.16153) | [PDF](https://arxiv.org/pdf/2410.16153)
|
11 |
+
|
12 |
+
## Model details
|
13 |
+
|
14 |
+
- **Model:** Pangea is an fully open-source Multilingual Multimodal Multicultural LLM.
|
15 |
+
- **Date:** Pangea-7B was trained in 2024.
|
16 |
+
- **Training Dataset:** [6M PangeaIns](https://huggingface.co/datasets/neulab/PangeaInstruct).
|
17 |
+
|
18 |
+
## Uses
|
19 |
+
|
20 |
+
### Direct Use
|
21 |
+
```python
|
22 |
+
from transformers import AutoProcessor, AutoModelForCausalLM
|
23 |
+
|
24 |
+
processor = AutoProcessor.from_pretrained("neulab/Pangea-7B")
|
25 |
+
model = AutoModelForCausalLM.from_pretrained("neulab/Pangea-7B")
|
26 |
+
```
|
27 |
+
|
28 |
+
## Citing the Model
|
29 |
+
|
30 |
+
**BibTeX Citation:**
|
31 |
+
|
32 |
+
```
|
33 |
+
@article{yue2024pangeafullyopenmultilingual,
|
34 |
+
title={Pangea: A Fully Open Multilingual Multimodal LLM for 39 Languages},
|
35 |
+
author={Xiang Yue and Yueqi Song and Akari Asai and Seungone Kim and Jean de Dieu Nyandwi and Simran Khanuja and Anjali Kantharuban and Lintang Sutawika and Sathyanarayanan Ramamoorthy and Graham Neubig},
|
36 |
+
year={2024},
|
37 |
+
journal={arXiv preprint arXiv:2410.16153},
|
38 |
+
url={https://arxiv.org/abs/2410.16153}
|
39 |
+
}
|
40 |
+
```
|
41 |
+
|
42 |
+
## Model Card Authors
|
43 |
+
|
44 |
+
The model is trained by the following authors:
|
45 |
+
|
46 |
+
Xiang Yue, Yueqi Song, Akari Asai, Seungone Kim, Jean de Dieu Nyandwi, Simran Khanuja, Anjali Kantharuban, Lintang Sutawika, Sathyanarayanan Ramamoorthy, Graham Neubig
|