TooTouch / SIDLinks
pytorch reimplementation for Detecting Adversarial Examples from Sensitivity Inconsistency of Spatial-Transform Domain
☆11Updated 3 years ago
Alternatives and similar repositories for SID
Users that are interested in SID are comparing it to the libraries listed below
Sorting:
- Towards Machine Unlearning Benchmarks: Forgetting the Personal Identities in Facial Recognition Systems☆63Updated 5 months ago
- [NeurIPS 2021] Official PyTorch Implementation for "Distilling Robust and Non-Robust Features in Adversarial Examples by Information Bott…☆48Updated 2 years ago
- Robust natural language watermarking using invariant features☆26Updated 2 years ago
- APBench: A Unified Availability Poisoning Attack and Defenses Benchmark (TMLR 08/2024)☆36Updated 6 months ago
- ☆15Updated 2 years ago
- Code for Friendly Noise against Adversarial Noise: A Powerful Defense against Data Poisoning Attacks (NeurIPS 2022)☆10Updated 2 years ago
- ☆53Updated 2 years ago
- CVPR 2025 - Anyattack: Towards Large-scale Self-supervised Adversarial Attacks on Vision-language Models☆55Updated 2 months ago
- Code repository for CVPR2024 paper 《Pre-trained Model Guided Fine-Tuning for Zero-Shot Adversarial Robustness》☆22Updated last year
- Consistency Regularization for Adversarial Robustness (AAAI 2022)☆53Updated 3 years ago
- ICCV 2021, We find most existing triggers of backdoor attacks in deep learning contain severe artifacts in the frequency domain. This Rep…☆45Updated 3 years ago
- ☆53Updated last year
- ☆20Updated last year
- A curated list of papers for the transferability of adversarial examples☆74Updated last year
- ☆43Updated 2 years ago
- Backdoor Safety Tuning (NeurIPS 2023 & 2024 Spotlight)☆26Updated 11 months ago
- ☆13Updated 4 years ago
- Reconstructive Neuron Pruning for Backdoor Defense (ICML 2023)☆39Updated last year
- [NeurIPS 2023] Code for the paper "Revisiting Adversarial Training for ImageNet: Architectures, Training and Generalization across Threa…☆38Updated 10 months ago
- Repository for the Paper: Refusing Safe Prompts for Multi-modal Large Language Models☆18Updated last year
- Code for the paper "Autoregressive Perturbations for Data Poisoning" (NeurIPS 2022)☆20Updated last year
- An Embarrassingly Simple Backdoor Attack on Self-supervised Learning☆18Updated last year
- ☆32Updated 3 years ago
- ☆31Updated 3 years ago
- [ICLR 2024 Spotlight 🔥 ] - [ Best Paper Award SoCal NLP 2023 🏆] - Jailbreak in pieces: Compositional Adversarial Attacks on Multi-Modal…☆73Updated last year
- Implementation of gradient-based adversarial attack(FGSM,MI-FGSM,PGD)☆100Updated 4 years ago
- Code for identifying natural backdoors in existing image datasets.☆15Updated 3 years ago
- ☆84Updated 4 years ago
- [CVPR 2023] The official implementation of our CVPR 2023 paper "Detecting Backdoors During the Inference Stage Based on Corruption Robust…☆23Updated 2 years ago
- Code for Transferable Unlearnable Examples☆21Updated 2 years ago