dangne / tmdLinks
[EMNLP'22] Textual Manifold-based Defense Against Natural Language Adversarial Examples
☆11Updated 2 years ago
Alternatives and similar repositories for tmd
Users that are interested in tmd are comparing it to the libraries listed below
Sorting:
- ☆18Updated 4 years ago
- ☆23Updated 2 years ago
- code of paper "IMPRESS: Evaluating the Resilience of Imperceptible Perturbations Against Unauthorized Data Usage in Diffusion-Based Gene…☆34Updated last year
- [ICLR 2024] Provable Robust Watermarking for AI-Generated Text☆38Updated 2 years ago
- codes for "Searching for an Effective Defender:Benchmarking Defense against Adversarial Word Substitution"☆31Updated 2 years ago
- Code for the paper "Rethinking Stealthiness of Backdoor Attack against NLP Models" (ACL-IJCNLP 2021)☆24Updated 4 years ago
- [ICML 2024] Prompting4Debugging: Red-Teaming Text-to-Image Diffusion Models by Finding Problematic Prompts (Official Pytorch Implementati…☆52Updated last year
- ACL 2021 - Defense against Adversarial Attacks in NLP via Dirichlet Neighborhood Ensemble☆18Updated 2 years ago
- Official Implementation of the paper "Three Bricks to Consolidate Watermarks for LLMs"☆50Updated last year
- The official implementation of ECCV'24 paper "To Generate or Not? Safety-Driven Unlearned Diffusion Models Are Still Easy To Generate Uns…☆87Updated 9 months ago
- ☆19Updated 2 years ago
- Towards Machine Unlearning Benchmarks: Forgetting the Personal Identities in Facial Recognition Systems☆64Updated 6 months ago
- ☆60Updated 2 years ago
- [ICML 2024] Unsupervised Adversarial Fine-Tuning of Vision Embeddings for Robust Large Vision-Language Models☆149Updated 6 months ago
- AnyDoor: Test-Time Backdoor Attacks on Multimodal Large Language Models☆60Updated last year
- ☆35Updated last year
- [SatML 2024] Shake to Leak: Fine-tuning Diffusion Models Can Amplify the Generative Privacy Risk☆15Updated 9 months ago
- ☆59Updated 3 years ago
- ☆52Updated last year
- [ICLR 2022 official code] Robust Learning Meets Generative Models: Can Proxy Distributions Improve Adversarial Robustness?☆29Updated 3 years ago
- Repo for arXiv preprint "Gradient-based Adversarial Attacks against Text Transformers"☆109Updated 2 years ago
- ☆23Updated 5 years ago
- Source code of paper "An Unforgeable Publicly Verifiable Watermark for Large Language Models" accepted by ICLR 2024☆34Updated last year
- Code and data of the ACL 2021 paper "Turn the Combination Lock: Learnable Textual Backdoor Attacks via Word Substitution"☆16Updated 4 years ago
- Code repo of our paper Towards Understanding Jailbreak Attacks in LLMs: A Representation Space Analysis (https://arxiv.org/abs/2406.10794…☆22Updated last year
- Investigating and Defending Shortcut Learning in Personalized Diffusion Models☆12Updated last year
- [CVPR23W] "A Pilot Study of Query-Free Adversarial Attack against Stable Diffusion" by Haomin Zhuang, Yihua Zhang and Sijia Liu☆26Updated last year
- [ICLR24 (Spotlight)] "SalUn: Empowering Machine Unlearning via Gradient-based Weight Saliency in Both Image Classification and Generation…☆141Updated 6 months ago
- Code for watermarking language models☆84Updated last year
- For Certified Robustness to Text Adversarial Attacks by Randomized [MASK]☆17Updated last year