Code for the paper "BadPrompt: Backdoor Attacks on Continuous Prompts"
☆42Jul 8, 2024Updated last year
Alternatives and similar repositories for BadPrompt
Users that are interested in BadPrompt are comparing it to the libraries listed below
Sorting:
- Code for paper: PoisonPrompt: Backdoor Attack on Prompt-based Large Language Models, IEEE ICASSP 2024. Demo//124.220.228.133:11107☆20Aug 10, 2024Updated last year
- Code for the paper "Be Careful about Poisoned Word Embeddings: Exploring the Vulnerability of the Embedding Layers in NLP Models" (NAACL-…☆44Jul 26, 2021Updated 4 years ago
- Implementation of the paper "Exploring the Universal Vulnerability of Prompt-based Learning Paradigm" on Findings of NAACL 2022☆32Jul 11, 2022Updated 3 years ago
- Code for the paper "Rethinking Stealthiness of Backdoor Attack against NLP Models" (ACL-IJCNLP 2021)☆24Dec 9, 2021Updated 4 years ago
- [ICLR 2022] Differentiable Prompt Makes Pre-trained Language Models Better Few-shot Learners☆130Dec 7, 2022Updated 3 years ago
- ☆17Sep 4, 2024Updated last year
- An open-source toolkit for textual backdoor attack and defense (NeurIPS 2022 D&B, Spotlight)☆200Apr 10, 2023Updated 2 years ago
- This repository contains the official code for the paper: "Prompt Injection: Parameterization of Fixed Inputs"☆32Sep 13, 2024Updated last year
- This repository is the official implementation of Generalized Data Weighting via Class-level Gradient Manipulation (NeurIPS 2021)(http://…☆24Oct 8, 2022Updated 3 years ago
- [CVPR 2022] "Quarantine: Sparsity Can Uncover the Trojan Attack Trigger for Free" by Tianlong Chen*, Zhenyu Zhang*, Yihua Zhang*, Shiyu C…☆27Oct 5, 2022Updated 3 years ago
- Code for Backdoor Attacks Against Dataset Distillation☆37Apr 19, 2023Updated 2 years ago
- ☆11Nov 12, 2024Updated last year
- Federated learning with model quantization☆16Nov 28, 2021Updated 4 years ago
- Pytorch implementation of NPAttack☆12Jul 7, 2020Updated 5 years ago
- ☆27Nov 9, 2022Updated 3 years ago
- Code for "Neuron Shapley: Discovering the Responsible Neurons"☆27May 1, 2024Updated last year
- 针对大模型的后门攻击☆12Jun 30, 2024Updated last year
- This is the implementation for CVPR 2022 Oral paper "Better Trigger Inversion Optimization in Backdoor Scanning."☆24Apr 5, 2022Updated 3 years ago
- A minimal PyTorch implementation of Label-Consistent Backdoor Attacks☆29Feb 8, 2021Updated 5 years ago
- ☆18Feb 10, 2022Updated 4 years ago
- ☆58Jul 27, 2022Updated 3 years ago
- Code for the paper "Overconfidence is a Dangerous Thing: Mitigating Membership Inference Attacks by Enforcing Less Confident Prediction" …☆12Sep 6, 2023Updated 2 years ago
- Code Repo for paper Label Leakage and Protection in Two-party Split Learning (ICLR 2022).☆22Mar 12, 2022Updated 4 years ago
- This work corroborates a run-time Trojan detection method exploiting STRong Intentional Perturbation of inputs, is a multi-domain Trojan …☆10Mar 7, 2021Updated 5 years ago
- On the Loss Landscape of Adversarial Training: Identifying Challenges and How to Overcome Them [NeurIPS 2020]☆36Jul 3, 2021Updated 4 years ago
- Red Queen Dataset and data generation template☆27Dec 26, 2025Updated 2 months ago
- The code for AAAI2023 (Generating Transferable 3D Adversarial Point Cloud via Random Perturbation Factorization)☆13Dec 24, 2022Updated 3 years ago
- ☆15Dec 12, 2023Updated 2 years ago
- Official repository for "Robust Prompt Optimization for Defending Language Models Against Jailbreaking Attacks"☆62Aug 8, 2024Updated last year
- Source code for "Neural Anisotropy Directions"☆16Nov 17, 2020Updated 5 years ago
- paper code☆28Oct 5, 2020Updated 5 years ago
- LLM prompt attacks for hacker CTFs via CTFd.☆15Dec 17, 2023Updated 2 years ago
- [CVPR 2024] Not All Prompts Are Secure: A Switchable Backdoor Attack Against Pre-trained Vision Transfomers☆16Oct 24, 2024Updated last year
- Code for paper: "Spinning Language Models: Risks of Propaganda-as-a-Service and Countermeasures"☆21Jun 6, 2022Updated 3 years ago
- The official PyTorch Implementation of the Paper "Adversarial Visual Robustness by Causal Intervention"☆18Oct 6, 2021Updated 4 years ago
- This is an implementation demo of the ICLR 2021 paper [Neural Attention Distillation: Erasing Backdoor Triggers from Deep Neural Networks…☆128Jan 18, 2022Updated 4 years ago
- 华中科技大学网络安全课程设计-Linux下的状态检测防火墙☆11Oct 17, 2022Updated 3 years ago
- [NeurIPS 2021] Better Safe Than Sorry: Preventing Delusive Adversaries with Adversarial Training☆32Jan 9, 2022Updated 4 years ago
- ☆25Jun 17, 2025Updated 9 months ago