Code for the paper "BadPrompt: Backdoor Attacks on Continuous Prompts"
☆40Jul 8, 2024Updated last year
Alternatives and similar repositories for BadPrompt
Users that are interested in BadPrompt are comparing it to the libraries listed below
Sorting:
- Code for paper: PoisonPrompt: Backdoor Attack on Prompt-based Large Language Models, IEEE ICASSP 2024. Demo//124.220.228.133:11107☆20Aug 10, 2024Updated last year
- Implementation of the paper "Exploring the Universal Vulnerability of Prompt-based Learning Paradigm" on Findings of NAACL 2022☆32Jul 11, 2022Updated 3 years ago
- Code for the paper "Be Careful about Poisoned Word Embeddings: Exploring the Vulnerability of the Embedding Layers in NLP Models" (NAACL-…☆44Jul 26, 2021Updated 4 years ago
- Code for the paper "Rethinking Stealthiness of Backdoor Attack against NLP Models" (ACL-IJCNLP 2021)☆24Dec 9, 2021Updated 4 years ago
- ☆17Sep 4, 2024Updated last year
- An open-source toolkit for textual backdoor attack and defense (NeurIPS 2022 D&B, Spotlight)☆200Apr 10, 2023Updated 2 years ago
- [ICLR 2022] Differentiable Prompt Makes Pre-trained Language Models Better Few-shot Learners☆130Dec 7, 2022Updated 3 years ago
- This repository is the official implementation of Generalized Data Weighting via Class-level Gradient Manipulation (NeurIPS 2021)(http://…☆24Oct 8, 2022Updated 3 years ago
- Code for Backdoor Attacks Against Dataset Distillation☆35Apr 19, 2023Updated 2 years ago
- Code for "Label-Consistent Backdoor Attacks"☆57Nov 22, 2020Updated 5 years ago
- ☆11Nov 12, 2024Updated last year
- This work corroborates a run-time Trojan detection method exploiting STRong Intentional Perturbation of inputs, is a multi-domain Trojan …☆10Mar 7, 2021Updated 4 years ago
- A pytorch implementation of the AAAI2021 paper GraCapsNet: Interpretable Graph Capsule Networks for Object Recognition☆10Oct 2, 2022Updated 3 years ago
- Official Implementation for "Purifying Quantization-conditioned Backdoors via Layer-wise Activation Correction with Distribution Approxim…☆12Aug 14, 2024Updated last year
- [CVPR 2022] "Quarantine: Sparsity Can Uncover the Trojan Attack Trigger for Free" by Tianlong Chen*, Zhenyu Zhang*, Yihua Zhang*, Shiyu C…☆27Oct 5, 2022Updated 3 years ago
- ☆26Dec 1, 2022Updated 3 years ago
- This repository contains the official code for the paper: "Prompt Injection: Parameterization of Fixed Inputs"☆32Sep 13, 2024Updated last year
- This is the implementation for CVPR 2022 Oral paper "Better Trigger Inversion Optimization in Backdoor Scanning."☆24Apr 5, 2022Updated 3 years ago
- Code for "Neuron Shapley: Discovering the Responsible Neurons"☆27May 1, 2024Updated last year
- ☆27Nov 9, 2022Updated 3 years ago
- Pytorch implementation of NPAttack☆12Jul 7, 2020Updated 5 years ago
- [ICLR2025] Detecting Backdoor Samples in Contrastive Language Image Pretraining☆19Feb 26, 2025Updated last year
- The code for AAAI2023 (Generating Transferable 3D Adversarial Point Cloud via Random Perturbation Factorization)☆13Dec 24, 2022Updated 3 years ago
- Federated learning with model quantization☆16Nov 28, 2021Updated 4 years ago
- ☆58Jul 27, 2022Updated 3 years ago
- ☆15Dec 12, 2023Updated 2 years ago
- ☆13Nov 8, 2022Updated 3 years ago
- Code&Data for the paper "Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based Agents" [NeurIPS 2024]☆109Sep 27, 2024Updated last year
- A minimal PyTorch implementation of Label-Consistent Backdoor Attacks☆29Feb 8, 2021Updated 5 years ago
- ☆16Aug 14, 2022Updated 3 years ago
- Code for NeurIPS 2024 Paper "Fight Back Against Jailbreaking via Prompt Adversarial Tuning"☆22May 6, 2025Updated 9 months ago
- [CVPR 2024] Not All Prompts Are Secure: A Switchable Backdoor Attack Against Pre-trained Vision Transfomers☆16Oct 24, 2024Updated last year
- Source code for "Neural Anisotropy Directions"☆16Nov 17, 2020Updated 5 years ago
- Red Queen Dataset and data generation template☆26Dec 26, 2025Updated 2 months ago
- A Backdoor Attack against 3D Point Cloud Classifiers (ICCV2021)☆18Oct 20, 2021Updated 4 years ago
- ☆18Aug 15, 2022Updated 3 years ago
- On the Loss Landscape of Adversarial Training: Identifying Challenges and How to Overcome Them [NeurIPS 2020]☆36Jul 3, 2021Updated 4 years ago
- SaTML'23 paper "Backdoor Attacks on Time Series: A Generative Approach" by Yujing Jiang, Xingjun Ma, Sarah Monazam Erfani, and James Bail…☆21Feb 5, 2023Updated 3 years ago
- Implementation of TABOR: A Highly Accurate Approach to Inspecting and Restoring Trojan Backdoors in AI Systems (https://arxiv.org/pdf/190…☆19Apr 13, 2023Updated 2 years ago