mlx-community/llava-interleave-qwen-0.5b-8bit

This model was converted to MLX format from llava-hf/llava-interleave-qwen-0.5b-hf using mlx-vlm version 0.0.15. Refer to the original model card for more details on the model.

Use with mlx

pip install -U mlx-vlm
python -m mlx_vlm.generate --model mlx-community/llava-interleave-qwen-0.5b-8bit --max-tokens 100 --temp 0.0
Downloads last month
24
Safetensors
Model size
245M params
Tensor type
FP16
U32
F32
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the HF Inference API does not support mlx models with pipeline type image-text-to-text