YuxinWenRick / canary-in-a-coalmine
☆32Updated 11 months ago
Related projects ⓘ
Alternatives and complementary repositories for canary-in-a-coalmine
- ☆23Updated 2 years ago
- Code for the paper "Autoregressive Perturbations for Data Poisoning" (NeurIPS 2022)☆18Updated 2 months ago
- ☆60Updated 9 months ago
- ☆48Updated 3 years ago
- [ICML 2023] Are Diffusion Models Vulnerable to Membership Inference Attacks?☆31Updated 2 months ago
- [ICLR'21] Dataset Inference for Ownership Resolution in Machine Learning☆31Updated 2 years ago
- Certified robustness "for free" using off-the-shelf diffusion models and classifiers☆36Updated last year
- Github repo for One-shot Neural Backdoor Erasing via Adversarial Weight Masking (NeurIPS 2022)☆14Updated last year
- [ICLR 2023, Spotlight] Indiscriminate Poisoning Attacks on Unsupervised Contrastive Learning☆28Updated 11 months ago
- ☆38Updated 3 months ago
- code release for "Unrolling SGD: Understanding Factors Influencing Machine Unlearning" published at EuroS&P'22☆22Updated 2 years ago
- Code for "Variational Model Inversion Attacks" Wang et al., NeurIPS2021☆20Updated 2 years ago
- ☆11Updated last year
- Code and data for the ICLR 2021 paper "Perceptual Adversarial Robustness: Defense Against Unseen Threat Models".☆54Updated 2 years ago
- [CVPR 2022] "Quarantine: Sparsity Can Uncover the Trojan Attack Trigger for Free" by Tianlong Chen*, Zhenyu Zhang*, Yihua Zhang*, Shiyu C…☆25Updated 2 years ago
- ☆32Updated 2 months ago
- [ICLR2023] Distilling Cognitive Backdoor Patterns within an Image☆31Updated last month
- PyTorch implementation of BPDA+EOT attack to evaluate adversarial defense with an EBM☆23Updated 4 years ago
- Code relative to "Adversarial robustness against multiple and single $l_p$-threat models via quick fine-tuning of robust classifiers"☆15Updated last year
- Backdoor Safety Tuning (NeurIPS 2023 & 2024 Spotlight)☆24Updated this week
- On the effectiveness of adversarial training against common corruptions [UAI 2022]☆30Updated 2 years ago
- ICCV 2021, We find most existing triggers of backdoor attacks in deep learning contain severe artifacts in the frequency domain. This Rep…☆41Updated 2 years ago
- Official repo for An Efficient Membership Inference Attack for the Diffusion Model by Proximal Initialization☆11Updated 8 months ago
- Official implementation of "RelaxLoss: Defending Membership Inference Attacks without Losing Utility" (ICLR 2022)☆46Updated 2 years ago
- Camouflage poisoning via machine unlearning☆15Updated last year
- Official Implementation of ICLR 2022 paper, ``Adversarial Unlearning of Backdoors via Implicit Hypergradient''☆50Updated 2 years ago
- [NeurIPS 2023] Code for the paper "Revisiting Adversarial Training for ImageNet: Architectures, Training and Generalization across Threa…☆38Updated 10 months ago
- Universal Adversarial Perturbations (UAPs) for PyTorch☆46Updated 3 years ago
- A Self-Consistent Robust Error (ICML 2022)☆67Updated last year
- [NeurIPS 2023] Differentially Private Image Classification by Learning Priors from Random Processes☆11Updated last year