Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,20 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
base_model:
|
4 |
+
- lmsys/vicuna-7b-v1.5
|
5 |
+
- openai/clip-vit-large-patch14-336
|
6 |
+
---
|
7 |
+
|
8 |
+
# p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay
|
9 |
+
This is the official model checkpoint of [p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay](https://arxiv.org/abs/2412.04449).
|
10 |
+
Please refer to [this repository](https://github.com/MCG-NJU/p-MoD) for our code.
|
11 |
+
|
12 |
+
## Model Description
|
13 |
+
This model is pretrained on [LCS-558K](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain) image caption data, and instruction-tuned on [779K LLaVA-NeXT instruction data](https://huggingface.co/datasets/lmms-lab/LLaVA-NeXT-Data).
|
14 |
+
|
15 |
+
## Citation
|
16 |
+
TBD
|
17 |
+
|
18 |
+
## License
|
19 |
+
Llama 2 is licensed under the LLAMA 2 Community License,
|
20 |
+
Copyright (c) Meta Platforms, Inc. All Rights Reserved.
|