A curated list of resources dedicated to the safety of Large Vision-Language Models. This repository aligns with our survey titled A Survey of Safety on Large Vision-Language Models: Attacks, Defenses, and Evaluations.
☆197Feb 6, 2026Updated last month
Alternatives and similar repositories for Awesome-LVLM-Safety
Users that are interested in Awesome-LVLM-Safety are comparing it to the libraries listed below
Sorting:
- [NeurIPS'25] Backdoor Cleaning without External Guidance in MLLM Fine-tuning☆17Oct 13, 2025Updated 4 months ago
- Multimodal Large Language Model (MLLM) Tuning Survey: Keeping Yourself is Important in Downstream Tuning Multimodal Large Language Model☆94Aug 5, 2025Updated 7 months ago
- [ICLR 2025] PyTorch Implementation of "ETA: Evaluating Then Aligning Safety of Vision Language Models at Inference Time"☆30Jul 20, 2025Updated 7 months ago
- TPAMI 2024 - Federated Learning for Generalization, Robustness, Fairness: A Survey and Benchmark☆110Feb 15, 2026Updated 3 weeks ago
- [CVPR'25] EMOE: Modality-Specific Enhanced Dynamic Emotion Experts☆112Jul 12, 2025Updated 7 months ago
- [CVPR 2025] Official implementation for "Steering Away from Harm: An Adaptive Approach to Defending Vision Language Model Against Jailbre…☆53Jul 5, 2025Updated 8 months ago
- [CVPR2025] Official Repository for IMMUNE: Improving Safety Against Jailbreaks in Multi-modal LLMs via Inference-Time Alignment☆27Jun 11, 2025Updated 8 months ago
- Safety at Scale: A Comprehensive Survey of Large Model Safety☆228Feb 3, 2026Updated last month
- A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).☆1,879Updated this week
- Accepted by IJCAI-24 Survey Track☆231Aug 25, 2024Updated last year
- Code for Neurips 2024 paper "Shadowcast: Stealthy Data Poisoning Attacks Against Vision-Language Models"☆59Jan 15, 2025Updated last year
- TPAMI2023 & CVPR2022 - Generalizable Heterogeneous Federated Cross-Correlation and Instance Similarity Learning & Learn From Others and B…☆110Oct 24, 2023Updated 2 years ago
- ☆185Nov 17, 2025Updated 3 months ago
- A curated list of Awesome Personalized Large Multimodal Models resources☆55Feb 4, 2026Updated last month
- ☆10Jun 14, 2024Updated last year
- Code and dataset for the paper: "Can Editing LLMs Inject Harm?"☆21Dec 26, 2025Updated 2 months ago
- ☆75Jan 21, 2026Updated last month
- Official Repository of Personalized Visual Instruct Tuning☆34Mar 6, 2025Updated last year
- [ICML'25 Spotlight] Catch Your Emotion: Sharpening Emotion Perception in Multimodal Large Language Models☆46Jan 21, 2026Updated last month
- [ICML 2025] X-Transfer Attacks: Towards Super Transferable Adversarial Attacks on CLIP☆39Feb 3, 2026Updated last month
- [ECCV 2024] The official code for "AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shi…☆72Feb 9, 2026Updated last month
- ☆14Mar 9, 2025Updated last year
- ☆14Feb 26, 2025Updated last year
- ☆10Jul 18, 2023Updated 2 years ago
- Evaluating Durability: Benchmark Insights into Multimodal Watermarking☆12Jun 7, 2024Updated last year
- ☆13Jun 17, 2024Updated last year
- [BMVC 2023] Semantic Adversarial Attacks via Diffusion Models☆25Nov 30, 2023Updated 2 years ago
- [NeurIPS 2021] “When does Contrastive Learning Preserve Adversarial Robustness from Pretraining to Finetuning?”☆48Nov 21, 2021Updated 4 years ago
- Official codebase for Image Hijacks: Adversarial Images can Control Generative Models at Runtime☆54Sep 19, 2023Updated 2 years ago
- ECSO (Make MLLM safe without neither training nor any external models!) (https://arxiv.org/abs/2403.09572)☆35Nov 2, 2024Updated last year
- [EMNLP 2025] Reasoning-to-Defend: Safety-Aware Reasoning Can Defend Large Language Models from Jailbreaking☆12Aug 22, 2025Updated 6 months ago
- (AAAI 24) Step Vulnerability Guided Mean Fluctuation Adversarial Attack against Conditional Diffusion Models☆11Oct 12, 2024Updated last year
- 武汉大学 iCalendar 校历☆12Updated this week
- ☆21Jul 25, 2025Updated 7 months ago
- A list of research towards security&privacy in AI-Generated Content☆16Jan 10, 2025Updated last year
- Code of paper "AdvReverb: AdvReverb: Rethinking the Stealthiness of Audio Adversarial Examples to Human Perception"☆18Nov 26, 2023Updated 2 years ago
- ☆22Oct 25, 2024Updated last year
- DiffusionGuard: A Robust Defense Against Malicious Diffusion-based Image Editing (ICLR 2025)☆43May 18, 2025Updated 9 months ago
- A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide…☆1,789Updated this week