aimagelab / LLaVA-MORELinks
LLaVA-MORE: A Comparative Study of LLMs and Visual Backbones for Enhanced Visual Instruction Tuning
☆148Updated 3 weeks ago
Alternatives and similar repositories for LLaVA-MORE
Users that are interested in LLaVA-MORE are comparing it to the libraries listed below
Sorting:
- [CVPR2024] ViP-LLaVA: Making Large Multimodal Models Understand Arbitrary Visual Prompts☆330Updated last year
- CuMo: Scaling Multimodal LLM with Co-Upcycled Mixture-of-Experts☆153Updated last year
- Official implementation of the Law of Vision Representation in MLLMs☆163Updated 9 months ago
- ☆139Updated 11 months ago
- Matryoshka Multimodal Models☆113Updated 7 months ago
- [COLM-2024] List Items One by One: A New Data Source and Learning Paradigm for Multimodal LLMs☆144Updated last year
- [CVPR'2025] VoCo-LLaMA: This repo is the official implementation of "VoCo-LLaMA: Towards Vision Compression with Large Language Models".☆186Updated 2 months ago
- [NeurIPS 2024] MoVA: Adapting Mixture of Vision Experts to Multimodal Context