shikiw commited on
Commit
4b19068
1 Parent(s): d67de9c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -0
README.md ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama2
3
+ language:
4
+ - en
5
+ - zh
6
+ tags:
7
+ - multimodal
8
+ datasets:
9
+ - liuhaotian/LLaVA-Pretrain
10
+ base_model:
11
+ - lmsys/vicuna-7b-v1.5
12
+ pipeline_tag: image-text-to-text
13
+ library_name: transformers
14
+ ---
15
+
16
+
17
+ ## **Citation**
18
+ If you find this model useful, please cite the following paper
19
+ ```
20
+ @article{huang2024deciphering,
21
+ title={Deciphering Cross-Modal Alignment in Large Vision-Language Models with Modality Integration Rate},
22
+ author={Huang, Qidong and Dong, Xiaoyi and Zhang, Pan and Zang, Yuhang and Cao, Yuhang and Wang, Jiaqi and Lin, Dahua and Zhang, Weiming and Yu, Nenghai},
23
+ journal={arXiv preprint arXiv:2410.07167},
24
+ year={2024}
25
+ }
26
+ ```