|
--- |
|
license: apache-2.0 |
|
base_model: |
|
- lmsys/vicuna-7b-v1.5 |
|
- openai/clip-vit-large-patch14-336 |
|
pipeline_tag: image-text-to-text |
|
--- |
|
|
|
# p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay |
|
This is the official model checkpoint of [p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay](https://arxiv.org/abs/2412.04449). |
|
Please refer to [this repository](https://github.com/MCG-NJU/p-MoD) for our code. |
|
|
|
## Model Description |
|
This model is pretrained on [LCS-558K](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain) image caption data, and instruction-tuned on [779K LLaVA-NeXT instruction data](https://huggingface.co/datasets/lmms-lab/LLaVA-NeXT-Data). |
|
|
|
## Citation |
|
If you find our model helpful for your research and applications, please cite our paper: |
|
```Bibtex |
|
@article{zhang2024pmod, |
|
title={p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay}, |
|
author={Zhang, Jun and Meng, Desen and Qi, Ji and Huang, Zhenpeng and Wu, Tao and Wang, Limin}, |
|
journal={arXiv preprint arXiv:2412.04449}, |
|
year={2024} |
|
} |
|
``` |
|
|
|
## License |
|
Llama 2 is licensed under the LLAMA 2 Community License, |
|
Copyright (c) Meta Platforms, Inc. All Rights Reserved. |