MICV-yonsei / LocalizationHeadsLinks
[CVPR 2025] Your Large Vision-Language Model Only Needs A Few Attention Heads For Visual Grounding
☆16Updated 3 months ago
Alternatives and similar repositories for LocalizationHeads
Users that are interested in LocalizationHeads are comparing it to the libraries listed below
Sorting:
- [ICCV-2025] Multi-Granular Spatio-Temporal Token Merging for Training-Free Acceleration of Video LLMs☆51Updated 5 months ago
- Question-Aware Gaussian Experts for Audio-Visual Question Answering -- Official Pytorch Implementation (CVPR'25, Highlight)☆25Updated 7 months ago
- 🌋👵🏻 Yo'LLaVA: Your Personalized Language and Vision Assistant (NeurIPS 2024)☆118Updated 9 months ago
- ☆29Updated 5 months ago
- ☆15Updated 4 months ago
- Rare-to-Frequent (R2F), ICLR'25, Spotlight☆52Updated 8 months ago
- Official Implementation of CODE☆16Updated last year
- Code and data for the paper "Emergent Visual-Semantic Hierarchies in Image-Text Representations" (ECCV 2024)☆32Updated last year
- [EMNLP 2024] Preserving Multi-Modal Capabilities of Pre-trained VLMs for Improving Vision-Linguistic Compositionality☆21Updated last year
- Official implementation of "VIRAL: Visual Representation Alignment for MLLMs".☆144Updated 3 months ago
- Code for "CAFe: Unifying Representation and Generation with Contrastive-Autoregressive Finetuning"☆30Updated 9 months ago
- Implementation of "Conditional Score Guidance for Text-Driven Image-to-Image Translation" (NeurIPS 2023).☆11Updated 2 years ago
- [ICLR 2025] SAFREE: Training-Free and Adaptive Guard for Safe Text-to-Image and Video Generation☆50Updated 11 months ago
- 🔥 [NeurIPS 2025] Official implementation of "Generate, but Verify: Reducing Visual Hallucination in Vision-Language Models with Retrospe…☆50Updated 3 months ago
- Official PyTorch implementation of Extract Free Dense Misalignment from CLIP (AAAI'25)☆25Updated 8 months ago
- Awesome Vision-Language Compositionality, a comprehensive curation of research papers in literature.☆34Updated 10 months ago
- ☆55Updated 4 months ago
- LLaVA-NeXT-Image-Llama3-Lora, Modified from https://github.com/arielnlee/LLaVA-1.6-ft☆45Updated last year
- [ICLR 2025] See What You Are Told: Visual Attention Sink in Large Multimodal Models☆82Updated 10 months ago
- [CVPR 2025 Highlight] Your Large Vision-Language Model Only Needs A Few Attention Heads For Visual Grounding☆53Updated 4 months ago
- [ECCV’24] Official repository for "BEAF: Observing Before-AFter Changes to Evaluate Hallucination in Vision-language Models"☆21Updated 9 months ago
- Code for "CLIP Behaves like a Bag-of-Words Model Cross-modally but not Uni-modally"☆18Updated 10 months ago
- Official code for the paper "Two Effects, One Trigger: On the Modality Gap, Object Bias, and Information Imbalance in Contrastive Vision-…☆21Updated 7 months ago
- LongVALE: Vision-Audio-Language-Event Benchmark Towards Time-Aware Omni-Modal Perception of Long Videos. (CVPR 2025))☆54Updated 7 months ago
- [CVPR 2024] Improving language-visual pretraining efficiency by perform cluster-based masking on images.☆30Updated last year
- Official implementation of paper VideoLLM Knows When to Speak: Enhancing Time-Sensitive Video Comprehension with Video-Text Duet Interact…☆40Updated 11 months ago
- CycleReward is a reward model trained on cycle consistency preferences to measure image-text alignment.☆52Updated 2 months ago
- Official Repository of Personalized Visual Instruct Tuning☆33Updated 10 months ago
- [AAAI 2025] ConVis: Contrastive Decoding with Hallucination Visualization for Mitigating Hallucinations in Multimodal Large Language Mode…☆26Updated last year
- ACL'24 (Oral) Tuning Large Multimodal Models for Videos using Reinforcement Learning from AI Feedback☆76Updated last year