FudanDISC / Awesome-Large-Multimodal-ModelsLinks
Papers of "A Survey on Large Multi-Modal Models from the Perspective of Input-Output Space Extension"
☆10Updated 5 months ago
Alternatives and similar repositories for Awesome-Large-Multimodal-Models
Users that are interested in Awesome-Large-Multimodal-Models are comparing it to the libraries listed below
Sorting:
- ☆21Updated last year
- Official Codes for Fine-Grained Visual Prompting, NeurIPS 2023☆52Updated last year
- Implementation of "VL-Mamba: Exploring State Space Models for Multimodal Learning"☆81Updated last year
- 🔎Official code for our paper: "VL-Uncertainty: Detecting Hallucination in Large Vision-Language Model via Uncertainty Estimation".☆35Updated 2 months ago
- [ICML 2024] Memory-Space Visual Prompting for Efficient Vision-Language Fine-Tuning☆49Updated last year
- [CVPR 2025] PyTorch implementation of paper "FLAME: Frozen Large Language Models Enable Data-Efficient Language-Image Pre-training"☆28Updated last month
- [ACM MM 2024] Hierarchical Multimodal Fine-grained Modulation for Visual Grounding.☆50Updated last month
- [CVPR 2024] Offical implemention of the paper "DePT: Decoupled Prompt Tuning"☆104Updated last week
- Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path Reasoning☆22Updated 8 months ago
- [CVPR2024] The code of "UniPT: Universal Parallel Tuning for Transfer Learning with Efficient Parameter and Memory"☆68Updated 7 months ago
- ✨A curated list of papers on the uncertainty in multi-modal large language model (MLLM).☆45Updated 2 months ago
- Code for paper: Visual Signal Enhancement for Object Hallucination Mitigation in Multimodal Large language Models☆19Updated 5 months ago
- [NeurIPS 2024] MoME: Mixture of Multimodal Experts for Generalist Multimodal Large Language Models☆65Updated last month
- [NeurIPS 2024] OneRef: Unified One-tower Expression Grounding and Segmentation with Mask Referring Modeling.☆20Updated 3 months ago
- [CVPR 2025] Augmenting Multimodal LLMs with Self-Reflective Tokens for Knowledge-based Visual Question Answering☆33Updated 2 months ago
- LLaVA-MR: Large Language-and-Vision Assistant for Video Moment Retrieval☆8Updated 5 months ago
- ☆12Updated this week
- The official implementation of RAR☆88Updated last year
- GroundVLP: Harnessing Zero-shot Visual Grounding from Vision-Language Pre-training and Open-Vocabulary Object Detection (AAAI 2024)☆67Updated last year
- Code for "DAMEX: Dataset-aware Mixture-of-Experts for visual understanding of mixture-of-datasets", accepted at Neurips 2023 (Main confer…☆21Updated last year
- [CVPR 2025 Highlight] Official Pytorch codebase for paper: "Assessing and Learning Alignment of Unimodal Vision and Language Models"☆39Updated last month
- [CVPR2025] Code Release of F-LMM: Grounding Frozen Large Multimodal Models☆91Updated last week
- [CVPR 2025] Adaptive Keyframe Sampling for Long Video Understanding☆66Updated last month
- Implementation of PyramidCLIP(NeurIPS2022).☆30Updated 2 years ago
- [CVPR2025] Official implementation of the paper "Multi-Layer Visual Feature Fusion in Multimodal LLMs: Methods, Analysis, and Best Practi…☆17Updated 3 months ago
- [ECCV 2024] ControlCap: Controllable Region-level Captioning☆75Updated 7 months ago
- This repo holds the official code and data for "Unveiling Parts Beyond Objects: Towards Finer-Granularity Referring Expression Segmentati…☆70Updated last year
- Emerging Pixel Grounding in Large Multimodal Models Without Grounding Supervision☆41Updated 2 months ago
- The official repo for "Ref-AVS: Refer and Segment Objects in Audio-Visual Scenes", ECCV 2024☆41Updated 6 months ago
- [LLaVA-Video-R1]✨First Adaptation of R1 to LLaVA-Video (2025-03-18)☆28Updated 3 weeks ago