KejiaZhang-Robust / VAP
Poison as Cure: Visual Noise for Mitigating Object Hallucinations in LVMs
☆18Updated last month
Alternatives and similar repositories for VAP:
Users that are interested in VAP are comparing it to the libraries listed below
- Official pytorch implementation of "RITUAL: Random Image Transformations as a Universal Anti-hallucination Lever in Large Vision Language…☆10Updated 4 months ago
- ☆11Updated 6 months ago
- ☆47Updated 5 months ago
- This is the official implementation of our paper "QuoTA: Query-oriented Token Assignment via CoT Query Decouple for Long Video Comprehens…☆68Updated last month
- ☆11Updated last week
- [ICLR 2025] Code for Self-Correcting Decoding with Generative Feedback for Mitigating Hallucinations in Large Vision-Language Models☆17Updated last week
- DyCoke: Dynamic Compression of Tokens for Fast Video Large Language Models☆40Updated 2 weeks ago
- The official code of the paper "Deciphering Cross-Modal Alignment in Large Vision-Language Models with Modality Integration Rate".☆98Updated 5 months ago
- [ECCV 2024] API: Attention Prompting on Image for Large Vision-Language Models☆85Updated 6 months ago
- ☆80Updated last month
- [ECCV 2024] The official code for "AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shi…☆57Updated 9 months ago
- [ICLR 2025] SAFREE: Training-Free and Adaptive Guard for Safe Text-to-Image and Video Generation☆36Updated 3 months ago
- The official repository of ECCV 2024 paper "Outlier-Aware Test-time Adaptation with Stable Memory Replay"☆18Updated 7 months ago
- ☆10Updated last week
- [ICCV 2023] Improving Adversarial Robustness of Masked Autoencoders via Test-time Frequency-domain Prompting☆15Updated last year
- [ICLR 2025] PyTorch Implementation of "ETA: Evaluating Then Aligning Safety of Vision Language Models at Inference Time"☆16Updated 2 months ago
- [ICLR'25] Official code for the paper 'MLLMs Know Where to Look: Training-free Perception of Small Visual Details with Multimodal LLMs'☆154Updated last week
- Code for ICLR 2025 Paper: Visual Description Grounding Reduces Hallucinations and Boosts Reasoning in LVLMs☆11Updated 2 weeks ago
- Hyperbolic Safety-Aware Vision-Language Models. CVPR 2025☆15Updated 2 weeks ago
- Elucidated Dataset Condensation (NeurIPS 2024)☆21Updated 6 months ago
- The First to Know: How Token Distributions Reveal Hidden Knowledge in Large Vision-Language Models?☆28Updated 5 months ago
- List of T2I safety papers, updated daily, welcome to discuss using Discussions☆61Updated 8 months ago
- MADTP: Multimodal Alignment-Guided Dynamic Token Pruning for Accelerating Vision-Language Transformer☆41Updated 7 months ago
- Data distillation benchmark☆58Updated last week
- [CVPR 2025] Mitigating Object Hallucinations in Large Vision-Language Models with Assembly of Global and Local Attention☆30Updated 9 months ago
- A pytorch implementation of CVPR24 paper "D4M: Dataset Distillation via Disentangled Diffusion Model"☆29Updated 7 months ago
- [CVPR 2025] VASparse: Towards Efficient Visual Hallucination Mitigation via Visual-Aware Token Sparsification☆23Updated last month
- Official implementation of paper "SparseVLM: Visual Token Sparsification for Efficient Vision-Language Model Inference".☆93Updated last month
- [ECCV2024] Reflective Instruction Tuning: Mitigating Hallucinations in Large Vision-Language Models☆16Updated 9 months ago
- [ECCV'24 Oral] The official GitHub page for ''Images are Achilles' Heel of Alignment: Exploiting Visual Vulnerabilities for Jailbreaking …☆28Updated 6 months ago