XuankunRong / Awesome-LVLM-SafetyView external linksLinks
A curated list of resources dedicated to the safety of Large Vision-Language Models. This repository aligns with our survey titled A Survey of Safety on Large Vision-Language Models: Attacks, Defenses, and Evaluations.
☆196Feb 6, 2026Updated last week
Alternatives and similar repositories for Awesome-LVLM-Safety
Users that are interested in Awesome-LVLM-Safety are comparing it to the libraries listed below
Sorting:
- [NeurIPS'25] Backdoor Cleaning without External Guidance in MLLM Fine-tuning☆17Oct 13, 2025Updated 4 months ago
- MAPO: MIXED ADVANTAGE POLICY OPTIMIZATION☆38Sep 24, 2025Updated 4 months ago
- Multimodal Large Language Model (MLLM) Tuning Survey: Keeping Yourself is Important in Downstream Tuning Multimodal Large Language Model☆94Aug 5, 2025Updated 6 months ago
- [ICLR 2025] PyTorch Implementation of "ETA: Evaluating Then Aligning Safety of Vision Language Models at Inference Time"☆29Jul 20, 2025Updated 6 months ago
- [CVPR'25 Oral] LoRASculpt: Sculpting LoRA for Harmonizing General and Specialized Knowledge in Multimodal Large Language Models☆49Aug 28, 2025Updated 5 months ago
- TPAMI 2024 - Federated Learning for Generalization, Robustness, Fairness: A Survey and Benchmark☆110Aug 5, 2025Updated 6 months ago
- 😎 up-to-date & curated list of awesome Attacks on Large-Vision-Language-Models papers, methods & resources.☆490Jan 27, 2026Updated 3 weeks ago
- [CVPR 2025] Official implementation for "Steering Away from Harm: An Adaptive Approach to Defending Vision Language Model Against Jailbre…☆52Jul 5, 2025Updated 7 months ago
- [CVPR2025] Official Repository for IMMUNE: Improving Safety Against Jailbreaks in Multi-modal LLMs via Inference-Time Alignment☆27Jun 11, 2025Updated 8 months ago
- Safety at Scale: A Comprehensive Survey of Large Model Safety☆227Feb 3, 2026Updated 2 weeks ago
- A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).☆1,860Jan 24, 2026Updated 3 weeks ago
- Accepted by IJCAI-24 Survey Track☆231Aug 25, 2024Updated last year
- Code for Neurips 2024 paper "Shadowcast: Stealthy Data Poisoning Attacks Against Vision-Language Models"☆59Jan 15, 2025Updated last year
- Accepted by ECCV 2024☆187Oct 15, 2024Updated last year
- A Survey on Jailbreak Attacks and Defenses against Multimodal Generative Models☆302Jan 11, 2026Updated last month
- ☆185Nov 17, 2025Updated 3 months ago
- [ICML 2025] X-Transfer Attacks: Towards Super Transferable Adversarial Attacks on CLIP☆37Feb 3, 2026Updated 2 weeks ago
- ☆10Jun 14, 2024Updated last year
- [ECCV 2024] The official code for "AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shi…☆71Feb 9, 2026Updated last week
- ☆74Jan 21, 2026Updated 3 weeks ago
- [ICML'25 Spotlight] Catch Your Emotion: Sharpening Emotion Perception in Multimodal Large Language Models☆46Jan 21, 2026Updated 3 weeks ago
- ☆14Feb 26, 2025Updated 11 months ago
- ☆13Mar 9, 2025Updated 11 months ago
- ☆10Jul 18, 2023Updated 2 years ago
- Evaluating Durability: Benchmark Insights into Multimodal Watermarking☆12Jun 7, 2024Updated last year
- ☆22Sep 5, 2025Updated 5 months ago
- ☆13Jun 17, 2024Updated last year
- [BMVC 2023] Semantic Adversarial Attacks via Diffusion Models☆24Nov 30, 2023Updated 2 years ago
- [NeurIPS 2021] “When does Contrastive Learning Preserve Adversarial Robustness from Pretraining to Finetuning?”☆48Nov 21, 2021Updated 4 years ago
- 武汉大学 iCalendar 校历☆11Feb 11, 2026Updated last week
- Official codebase for Image Hijacks: Adversarial Images can Control Generative Models at Runtime☆54Sep 19, 2023Updated 2 years ago
- ECSO (Make MLLM safe without neither training nor any external models!) (https://arxiv.org/abs/2403.09572)☆35Nov 2, 2024Updated last year
- [EMNLP 2025] Reasoning-to-Defend: Safety-Aware Reasoning Can Defend Large Language Models from Jailbreaking☆12Aug 22, 2025Updated 5 months ago
- (AAAI 24) Step Vulnerability Guided Mean Fluctuation Adversarial Attack against Conditional Diffusion Models☆11Oct 12, 2024Updated last year
- ☆21Jul 25, 2025Updated 6 months ago
- Code of paper "AdvReverb: AdvReverb: Rethinking the Stealthiness of Audio Adversarial Examples to Human Perception"☆18Nov 26, 2023Updated 2 years ago
- A list of research towards security&privacy in AI-Generated Content☆16Jan 10, 2025Updated last year
- DiffusionGuard: A Robust Defense Against Malicious Diffusion-based Image Editing (ICLR 2025)☆43May 18, 2025Updated 8 months ago
- [ICML 2025] An official source code for paper "FlipAttack: Jailbreak LLMs via Flipping".☆163May 2, 2025Updated 9 months ago