ShyFoo / NemesisLinks
Official implementation of Nemesis: Normalizing the Soft-prompt Vectors of Vision-Language Models (ICLR 2024 Spotlight)
☆13Updated 6 months ago
Alternatives and similar repositories for Nemesis
Users that are interested in Nemesis are comparing it to the libraries listed below
Sorting:
- Autoregressive Semantic Visual Reconstruction Helps VLMs Understand Better☆31Updated 3 weeks ago
- ☆17Updated 7 months ago
- ☆21Updated 8 months ago
- Official PyTorch implementation of RACRO (https://www.arxiv.org/abs/2506.04559)☆15Updated last week
- Fast-Slow Thinking for Large Vision-Language Model Reasoning☆16Updated 2 months ago
- The official repo of our work "Pensieve: Retrospect-then-Compare mitigates Visual Hallucination"☆16Updated last year
- [ECCV 2024] STEVE in Minecraft is for See and Think: Embodied Agent in Virtual Environment☆39Updated last year
- [CVPRW 2024] TrafficVLM: A Controllable Visual Language Model for Traffic Video Captioning. Official code for the 3rd place solution of t…☆39Updated 5 months ago
- Repo for paper "T2Vid: Translating Long Text into Multi-Image is the Catalyst for Video-LLMs"☆49Updated 4 months ago
- ☆19Updated 2 months ago
- Github repository for "Why Is Spatial Reasoning Hard for VLMs? An Attention Mechanism Perspective on Focus Areas" (ICML 2025)☆35Updated 2 months ago
- Official Repository of Personalized Visual Instruct Tuning☆30Updated 4 months ago
- (NeurIPS 2024) What Makes CLIP More Robust to Long-Tailed Pre-Training Data? A Controlled Study for Transferable Insights☆27Updated 8 months ago
- [NeurIPS-2024] The offical Implementation of "Instruction-Guided Visual Masking"☆35Updated 7 months ago
- ☆37Updated last month
- [NeurIPS2023] Official implementation of the paper "Large Language Models are Visual Reasoning Coordinators"☆105Updated last year
- ☆19Updated last month
- WorldSense: Evaluating Real-world Omnimodal Understanding for Multimodal LLMs☆26Updated 2 months ago
- [ICLR 2025] Official implementation and benchmark evaluation repository of <PhysBench: Benchmarking and Enhancing Vision-Language Models …☆64Updated last month
- ☆19Updated 2 months ago
- This repo contains evaluation code for the paper "AV-Odyssey: Can Your Multimodal LLMs Really Understand Audio-Visual Information?"☆26Updated 6 months ago
- iLLaVA: An Image is Worth Fewer Than 1/3 Input Tokens in Large Multimodal Models☆19Updated 5 months ago
- Scaffold Prompting to promote LMMs☆43Updated 6 months ago
- LEO: A powerful Hybrid Multimodal LLM☆18Updated 5 months ago
- A novel alignment framework that leverages image retrieval to mitigate hallucinations in Vision Language Models.☆44Updated 2 months ago
- Code and data for paper "Exploring Hallucination of Large Multimodal Models in Video Understanding: Benchmark, Analysis and Mitigation".☆16Updated last month
- [ICLR 2025] CREMA: Generalizable and Efficient Video-Language Reasoning via Multimodal Modular Fusion☆48Updated last week
- ☆12Updated 6 months ago
- ☆16Updated 6 months ago
- official repo for paper "[CLS] Token Tells Everything Needed for Training-free Efficient MLLMs"☆22Updated 2 months ago