Image-Text-to-Text
Safetensors
pmod_llava_llama
JungleGym commited on
Commit
4c09c38
1 Parent(s): a5bfe56

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -3
README.md CHANGED
@@ -1,3 +1,20 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model:
4
+ - lmsys/vicuna-7b-v1.5
5
+ - openai/clip-vit-large-patch14-336
6
+ ---
7
+
8
+ # p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay
9
+ This is the official model checkpoint of [p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay](https://arxiv.org/abs/2412.04449).
10
+ Please refer to [this repository](https://github.com/MCG-NJU/p-MoD) for our code.
11
+
12
+ ## Model Description
13
+ This model is pretrained on [LCS-558K](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain) image caption data, and instruction-tuned on [779K LLaVA-NeXT instruction data](https://huggingface.co/datasets/lmms-lab/LLaVA-NeXT-Data).
14
+
15
+ ## Citation
16
+ TBD
17
+
18
+ ## License
19
+ Llama 2 is licensed under the LLAMA 2 Community License,
20
+ Copyright (c) Meta Platforms, Inc. All Rights Reserved.