jiaangli / VLCA
Do Vision and Language Models Share Concepts? A Vector Space Alignment Study
☆14Updated 4 months ago
Alternatives and similar repositories for VLCA:
Users that are interested in VLCA are comparing it to the libraries listed below
- DeepPerception: Advancing R1-like Cognitive Visual Perception in MLLMs for Knowledge-Intensive Visual Grounding☆48Updated 2 weeks ago
- ☆40Updated 5 months ago
- ☆10Updated 5 months ago
- [EMNLP 2024] Preserving Multi-Modal Capabilities of Pre-trained VLMs for Improving Vision-Linguistic Compositionality☆16Updated 6 months ago
- [ICLR 2025] MLLM can see? Dynamic Correction Decoding for Hallucination Mitigation☆51Updated 4 months ago
- ☆17Updated 8 months ago
- An Enhanced CLIP Framework for Learning with Synthetic Captions☆28Updated last week
- COLA: Evaluate how well your vision-language model can Compose Objects Localized with Attributes!☆24Updated 4 months ago
- Official Repository of Personalized Visual Instruct Tuning☆28Updated last month
- If CLIP Could Talk: Understanding Vision-Language Model Representations Through Their Preferred Concept Descriptions☆16Updated last year
- PyTorch implementation of StableMask (ICML'24)☆12Updated 9 months ago
- [EMNLP 2024] Official code for "Beyond Embeddings: The Promise of Visual Table in Multi-Modal Models"☆17Updated 5 months ago
- We introduce new approach, Token Reduction using CLIP Metric (TRIM), aimed at improving the efficiency of MLLMs without sacrificing their…☆12Updated 4 months ago
- This is the implementation of CounterCurate, the data curation pipeline of both physical and semantic counterfactual image-caption pairs.☆18Updated 9 months ago
- ☆18Updated 9 months ago
- ☆11Updated 6 months ago
- CLIP-MoE: Mixture of Experts for CLIP☆31Updated 6 months ago
- Official repository of "Interactive Text-to-Image Retrieval with Large Language Models: A Plug-and-Play Approach" (ACL 2024 Oral)☆21Updated 3 weeks ago
- Emerging Pixel Grounding in Large Multimodal Models Without Grounding Supervision☆38Updated 3 weeks ago
- (ICLR2025 Spotlight) DEEM: Official implementation of Diffusion models serve as the eyes of large language models for image perception.☆27Updated last month
- Official Pytorch implementation of "Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations" (ICLR '25)☆66Updated last month
- ☆34Updated 9 months ago
- Code for "Are “Hierarchical” Visual Representations Hierarchical?" in NeurIPS Workshop for Symmetry and Geometry in Neural Representation…☆20Updated last year
- Code for paper: Unified Text-to-Image Generation and Retrieval☆14Updated 9 months ago
- The released data for paper "Measuring and Improving Chain-of-Thought Reasoning in Vision-Language Models".☆32Updated last year
- Official pytorch implementation of "Don't Miss the Forest for the Trees: Attentional Vision Calibration for Large Vision Language Models"☆14Updated 8 months ago
- [ICLR 2025] CREMA: Generalizable and Efficient Video-Language Reasoning via Multimodal Modular Fusion☆41Updated 2 months ago
- Official implementation of Scaling Laws in Patchification: An Image Is Worth 50,176 Tokens And More☆17Updated last month
- Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path Reasoning☆20Updated 7 months ago
- Implementation and dataset for paper "Can MLLMs Perform Text-to-Image In-Context Learning?"☆36Updated 3 weeks ago