grasses / PromptCARELinks
Code for paper: "PromptCARE: Prompt Copyright Protection by Watermark Injection and Verification", IEEE S&P 2024.
☆32Updated 9 months ago
Alternatives and similar repositories for PromptCARE
Users that are interested in PromptCARE are comparing it to the libraries listed below
Sorting:
- Code for paper: PoisonPrompt: Backdoor Attack on Prompt-based Large Language Models, IEEE ICASSP 2024. Demo//124.220.228.133:11107☆17Updated 9 months ago
- [CCS'22] SSLGuard: A Watermarking Scheme for Self-supervised Learning Pre-trained Encoders☆19Updated 2 years ago
- ☆20Updated last year
- ☆20Updated last year
- This is the source code for MEA-Defender. Our paper is accepted by the IEEE Symposium on Security and Privacy (S&P) 2024.☆23Updated last year
- [CVPR 2023] Backdoor Defense via Adaptively Splitting Poisoned Dataset☆49Updated last year
- ☆38Updated 3 years ago
- Text-CRS: A Generalized Certified Robustness Framework against Textual Adversarial Attacks (IEEE S&P 2024)☆34Updated last month
- [AAAI 2024] Data-Free Hard-Label Robustness Stealing Attack☆13Updated last year
- Repository for Towards Codable Watermarking for Large Language Models☆37Updated last year
- [CCS-LAMPS'24] LLM IP Protection Against Model Merging☆15Updated 7 months ago
- Reconstructive Neuron Pruning for Backdoor Defense (ICML 2023)☆37Updated last year
- ☆18Updated 2 years ago
- Code for "Adversarial Illusions in Multi-Modal Embeddings"☆22Updated 10 months ago
- Robust natural language watermarking using invariant features☆25Updated last year
- ☆10Updated 5 months ago
- ☆19Updated last year
- Code and data for paper "A Semantic Invariant Robust Watermark for Large Language Models" accepted by ICLR 2024.☆32Updated 6 months ago
- Backdoor Safety Tuning (NeurIPS 2023 & 2024 Spotlight)☆26Updated 6 months ago
- ☆23Updated 2 years ago
- [MM'23 Oral] "Text-to-image diffusion models can be easily backdoored through multimodal data poisoning"☆28Updated 3 months ago
- [AAAI'21] Deep Feature Space Trojan Attack of Neural Networks by Controlled Detoxification☆29Updated 5 months ago
- A toolbox for backdoor attacks.☆22Updated 2 years ago
- [NDSS 2025] Official code for our paper "Explanation as a Watermark: Towards Harmless and Multi-bit Model Ownership Verification via Wate…☆37Updated 7 months ago
- ☆82Updated 3 years ago
- ☆17Updated 3 years ago
- An Embarrassingly Simple Backdoor Attack on Self-supervised Learning☆16Updated last year
- Official Implementation of ICLR 2022 paper, ``Adversarial Unlearning of Backdoors via Implicit Hypergradient''☆53Updated 2 years ago
- This is the code repository of our submission: Understanding the Dark Side of LLMs’ Intrinsic Self-Correction.☆56Updated 5 months ago
- Codes for NeurIPS 2021 paper "Adversarial Neuron Pruning Purifies Backdoored Deep Models"☆57Updated 2 years ago