[NeurIPS'24]Q-VLM: Post-training Quantization for Large Vision-Language Models

Efficient and accurate memory saving method towards W4A4 large multi-modal models. [Paper][Code]

Q-VLM: Post-training Quantization for Large Vision-Language Models
Changyuan Wang, Ziwei Wang, Xiuwei Xu, Yansong Tang, Jie Zhou, Jiwen Lu

Finetuning LLaVA Model on ScienceQA Dataset

Thanks for LLaVA (https://github.com/haotian-liu/LLaVA) for the amazing open-source model!

We combined the LLaVA-7B-v1.1 model (LLaVA-7B-v1.1) and the projector from LLaVA-7B-v1.3 (LLaVA-7B-v1.3 projector) and finetuned the model on the ScienceQA dataset. This model is used to test the effectiveness of our quantization method on the ScienceQA dataset.

Downloads last month
17
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.