Famous Vision Language Models and Their Architectures
-
Updated
Feb 24, 2025 - Markdown
Famous Vision Language Models and Their Architectures
A minimal codebase for finetuning large multimodal models, supporting llava-1.5/1.6, llava-interleave, llava-next-video, llava-onevision, llama-3.2-vision, qwen-vl, qwen2-vl, phi3-v etc.
Mark web pages for use with vision-language models
Qwen-VL base model for use with Autodistill.
Add a description, image, and links to the qwen-vl topic page so that developers can more easily learn about it.
To associate your repository with the qwen-vl topic, visit your repo's landing page and select "manage topics."