Huanjin Yao

HuanjinYao

AI & ML interests

None yet

Recent Activity

updated a model about 1 month ago
HuanjinYao/Mulberry_llava_8b
updated a collection about 1 month ago
Mulberry
updated a model about 1 month ago
HuanjinYao/Mulberry_qwen2vl_7b
View all activity

Organizations

Tsinghua University's profile picture ZeroGPU Explorers's profile picture

HuanjinYao's activity

updated a collection about 1 month ago
updated a collection about 2 months ago
updated a collection about 2 months ago
New activity in HuanjinYao/Mulberry_llava_8b about 2 months ago

Add metadata

#1 opened about 2 months ago by
nielsr
New activity in ShareGPT4Video/ShareGPT4Video 9 months ago
reacted to merve's post with 👀 10 months ago
view post
Post
2061
Do we fully leverage ViT encoders in vision language models?

A new paper (by @HuanjinYao et al) built a dense connector that does it better! HuanjinYao/DenseConnector-v1.5-8B
HuanjinYao/denseconnector-66500e173fc8c9f05dc98dea

VLMs consist of an image encoder block, a projection layer that projects image embeddings to text embedding space and then a text decoder sequentially connected 📖
This paper explores using intermediate states of image encoder and not a single output 🤩
The authors explore three different ways of instantiating dense connector: sparse token integration, sparse channel integration and dense channel integration. (see paper on how they do it Dense Connector for MLLMs (2405.13800))

They explore all three of them integrated to LLaVA 1.5 and found out each of the new models are superior to the original LLaVA 1.5 🥹 I tried the model and it seems to work very well. As part of the release, the authors have released various ckpts based on different decoders (Vicuna 7/13B and Llama 3-8B) that you can find in the collection 🤗