Trustworthy-AI-Group / Adversarial_Examples_PapersLinks
A list of recent papers about adversarial learning
☆192Updated last week
Alternatives and similar repositories for Adversarial_Examples_Papers
Users that are interested in Adversarial_Examples_Papers are comparing it to the libraries listed below
Sorting:
- A curated list of papers & resources on backdoor attacks and defenses in deep learning.☆216Updated last year
- A curated list of papers & resources linked to data poisoning, backdoor attacks and defenses against them (no longer maintained)☆266Updated 6 months ago
- ☆102Updated last year
- A Survey on Jailbreak Attacks and Defenses against Multimodal Generative Models☆215Updated last month
- A curated list of papers for the transferability of adversarial examples☆72Updated last year
- This is an official repository of ``VLAttack: Multimodal Adversarial Attacks on Vision-Language Tasks via Pre-trained Models'' (NeurIPS 2…☆56Updated 4 months ago
- Safety at Scale: A Comprehensive Survey of Large Model Safety☆183Updated 5 months ago
- Code for ACM MM2024 paper: White-box Multimodal Jailbreaks Against Large Vision-Language Models☆29Updated 7 months ago
- This is the code repository of our submission: Understanding the Dark Side of LLMs’ Intrinsic Self-Correction.☆58Updated 7 months ago
- Revisiting Transferable Adversarial Images (arXiv)☆124Updated 4 months ago
- ☆76Updated last year
- 😎 up-to-date & curated list of awesome Attacks on Large-Vision-Language-Models papers, methods & resources.☆349Updated last week
- TransferAttack is a pytorch framework to boost the adversarial transferability for image classification.☆378Updated 2 weeks ago
- BackdoorLLM: A Comprehensive Benchmark for Backdoor Attacks and Defenses on Large Language Models☆188Updated last month
- Code for paper: "PromptCARE: Prompt Copyright Protection by Watermark Injection and Verification", IEEE S&P 2024.☆32Updated 11 months ago
- WaNet - Imperceptible Warping-based Backdoor Attack (ICLR 2021)☆127Updated 8 months ago
- ☆58Updated 2 months ago
- Invisible Backdoor Attack with Sample-Specific Triggers☆97Updated 3 years ago
- ☆40Updated 3 years ago
- A list of recent adversarial attack and defense papers (including those on large language models)☆42Updated this week
- Official Code for ART: Automatic Red-teaming for Text-to-Image Models to Protect Benign Users (NeurIPS 2024)☆16Updated 9 months ago
- ☆82Updated 4 years ago
- A toolbox for backdoor attacks.☆22Updated 2 years ago
- [NeurIPS-2023] Annual Conference on Neural Information Processing Systems☆209Updated 7 months ago
- CVPR 2025 - Anyattack: Towards Large-scale Self-supervised Adversarial Attacks on Vision-language Models☆41Updated last month
- [ICLR 2024] Inducing High Energy-Latency of Large Vision-Language Models with Verbose Images☆37Updated last year
- [AAAI'25 (Oral)] Jailbreaking Large Vision-language Models via Typographic Visual Prompts☆160Updated last month
- ☆48Updated last year
- ☆112Updated 2 months ago
- [NDSS 2025] Official code for our paper "Explanation as a Watermark: Towards Harmless and Multi-bit Model Ownership Verification via Wate…☆39Updated 9 months ago