[NeurIPS 2025] BackdoorLLM: A Comprehensive Benchmark for Backdoor Attacks and Defenses on Large Language Models
☆276Feb 2, 2026Updated last month
Alternatives and similar repositories for BackdoorLLM
Users that are interested in BackdoorLLM are comparing it to the libraries listed below
Sorting:
- 🔥🔥🔥 Detecting hidden backdoors in Large Language Models with only black-box access☆52Jun 2, 2025Updated 9 months ago
- Code&Data for the paper "Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based Agents" [NeurIPS 2024]☆109Sep 27, 2024Updated last year
- ☆26Aug 21, 2024Updated last year
- [ICLR24] Official Repo of BadChain: Backdoor Chain-of-Thought Prompting for Large Language Models☆48Jul 24, 2024Updated last year
- Safety at Scale: A Comprehensive Survey of Large Model Safety☆228Feb 3, 2026Updated 3 weeks ago
- Code for paper: PoisonPrompt: Backdoor Attack on Prompt-based Large Language Models, IEEE ICASSP 2024. Demo//124.220.228.133:11107☆20Aug 10, 2024Updated last year
- Official code for the ICCV2023 paper ``One-bit Flip is All You Need: When Bit-flip Attack Meets Model Training''☆20Aug 9, 2023Updated 2 years ago
- [ICML 2025] X-Transfer Attacks: Towards Super Transferable Adversarial Attacks on CLIP☆37Feb 3, 2026Updated 3 weeks ago
- [ICLR 2025] BlueSuffix: Reinforced Blue Teaming for Vision-Language Models Against Jailbreak Attacks☆30Nov 2, 2025Updated 4 months ago
- AnyDoor: Test-Time Backdoor Attacks on Multimodal Large Language Models☆60Apr 8, 2024Updated last year
- ☆14Feb 26, 2025Updated last year
- This repo is the official implementation of the ICLR'23 paper "Towards Robustness Certification Against Universal Perturbations." We calc…☆12Feb 14, 2023Updated 3 years ago
- A survey on harmful fine-tuning attack for large language model☆232Updated this week
- Code for Neurips 2024 paper "Shadowcast: Stealthy Data Poisoning Attacks Against Vision-Language Models"☆59Jan 15, 2025Updated last year
- ICL backdoor attack☆17Nov 4, 2024Updated last year
- [ICLR2025] Detecting Backdoor Samples in Contrastive Language Image Pretraining☆19Feb 26, 2025Updated last year
- ☆37Oct 17, 2024Updated last year
- This is the repository that introduces research topics related to protecting intellectual property (IP) of AI from a data-centric perspec…☆23Oct 30, 2023Updated 2 years ago
- Code for paper "Membership Inference Attacks Against Vision-Language Models"☆26Jan 25, 2025Updated last year
- Backdoor Safety Tuning (NeurIPS 2023 & 2024 Spotlight)☆27Nov 18, 2024Updated last year
- Identification of the Adversary from a Single Adversarial Example (ICML 2023)☆10Jul 15, 2024Updated last year
- Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMs. Empirical tricks for LLM Jailbreaking. (NeurIPS 2024)☆163Nov 30, 2024Updated last year
- ☆584Jul 4, 2025Updated 7 months ago
- The implementatin of our ICLR 2021 work: Targeted Attack against Deep Neural Networks via Flipping Limited Weight Bits☆18Jul 20, 2021Updated 4 years ago
- The repo for paper: Exploiting the Index Gradients for Optimization-Based Jailbreaking on Large Language Models.☆13Dec 16, 2024Updated last year
- [NeurIPS 2024] Official implementation for "AgentPoison: Red-teaming LLM Agents via Memory or Knowledge Base Backdoor Poisoning"☆199Apr 12, 2025Updated 10 months ago
- Composite Backdoor Attacks Against Large Language Models☆22Apr 12, 2024Updated last year
- Papers and resources related to the security and privacy of LLMs 🤖☆566Jun 8, 2025Updated 8 months ago
- A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).☆1,870Feb 23, 2026Updated last week
- Imbalanced Gradients: A New Cause of Overestimated Adversarial Robustness. (MD attacks)☆11Aug 29, 2020Updated 5 years ago
- The open-sourced Python toolbox for backdoor attacks and defenses.☆644Sep 27, 2025Updated 5 months ago
- [ICLR2023] Distilling Cognitive Backdoor Patterns within an Image☆36Oct 29, 2025Updated 4 months ago
- Code for Transferable Unlearnable Examples☆22Mar 11, 2023Updated 2 years ago
- A Unified Benchmark and Toolbox for Multimodal Jailbreak Attack–Defense Evaluation☆58Jan 23, 2026Updated last month
- Code for ICCV2025 paper——IDEATOR: Jailbreaking and Benchmarking Large Vision-Language Models Using Themselves☆17Jul 11, 2025Updated 7 months ago
- Implement of Implicit Knowledge Extraction Attack.☆18May 28, 2025Updated 9 months ago
- Pytorch implementation of NPAttack☆12Jul 7, 2020Updated 5 years ago
- this is for the ACM MM paper---Backdoor Attack on Crowd Counting☆17Jul 10, 2022Updated 3 years ago
- ☆77Dec 19, 2024Updated last year