Code for the paper Boosting Accuracy and Robustness of Student Models via Adaptive Adversarial Distillation (CVPR 2023).
☆34May 26, 2023Updated 2 years ago
Alternatives and similar repositories for AdaAD
Users that are interested in AdaAD are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- [ICLR 2022] Reliable Adversarial Distillation with Unreliable Teachers☆22Feb 20, 2022Updated 4 years ago
- The Code of ECCV2022:Enhanced Accuracy and Robustness via Multi-Teacher Adversarial Distillation☆36Oct 23, 2025Updated 5 months ago
- The repo for paper: Exploiting the Index Gradients for Optimization-Based Jailbreaking on Large Language Models.☆14Dec 16, 2024Updated last year
- ☆58Jul 27, 2022Updated 3 years ago
- [ICLR 2022 official code] Robust Learning Meets Generative Models: Can Proxy Distributions Improve Adversarial Robustness?☆29Mar 15, 2022Updated 4 years ago
- Accept by CVPR 2025 (highlight)☆24Jun 8, 2025Updated 9 months ago
- Code for the paper "Better Diffusion Models Further Improve Adversarial Training" (ICML 2023)☆145Jul 31, 2023Updated 2 years ago
- This is the official code for "Revisiting Adversarial Robustness Distillation: Robust Soft Labels Make Student Better"☆45Aug 29, 2021Updated 4 years ago
- The code of the ICLR 2024 paper: Adversarial Training on Purification (AToP): Advancing Both Robustness and Generalization☆10Nov 21, 2024Updated last year
- Code repository for the paper "Heuristic Induced Multimodal Risk Distribution Jailbreak Attack for Multimodal Large Language Models"☆15Aug 7, 2025Updated 7 months ago
- Helper-based Adversarial Training: Reducing Excessive Margin to Achieve a Better Accuracy vs. Robustness Trade-off☆33Apr 28, 2022Updated 3 years ago
- Data and code for the paper: Finding Safety Neurons in Large Language Models☆23Jan 29, 2026Updated last month
- Pytorch implementation of Adversarially Robust Distillation (ARD)☆59May 24, 2019Updated 6 years ago
- Reproduces BPDA attack in pytorch☆13Oct 10, 2019Updated 6 years ago
- Code relative to "Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks"☆741May 16, 2024Updated last year
- CVPR 2023 generalist☆16Oct 25, 2023Updated 2 years ago
- ☆14Jun 6, 2023Updated 2 years ago
- Code for ICLR2020 "Improving Adversarial Robustness Requires Revisiting Misclassified Examples"☆153Oct 15, 2020Updated 5 years ago
- ☆20Feb 3, 2025Updated last year
- [ICLR 2022]: Fast AdvProp☆35Mar 21, 2022Updated 4 years ago
- ☆23Dec 14, 2023Updated 2 years ago
- ☆20Oct 5, 2023Updated 2 years ago
- Code for our ICLR 2023 paper Making Substitute Models More Bayesian Can Enhance Transferability of Adversarial Examples.☆18May 31, 2023Updated 2 years ago
- Code for NeurIPS 2024 Paper "Fight Back Against Jailbreaking via Prompt Adversarial Tuning"☆22May 6, 2025Updated 10 months ago
- [IEEE T-BIOM] FaceXBench: Evaluating Multimodal LLMs on Face Understanding☆20Jan 15, 2026Updated 2 months ago
- [ICCV 2023] HybridAugment++: Unified Frequency Spectra Perturbations for Model Robustness☆17Sep 28, 2023Updated 2 years ago
- On the Loss Landscape of Adversarial Training: Identifying Challenges and How to Overcome Them [NeurIPS 2020]☆36Jul 3, 2021Updated 4 years ago
- ☆62Aug 9, 2023Updated 2 years ago
- Convert tensorflow model to pytorch model via [MMdnn](https://github.com/microsoft/MMdnn) for adversarial attacks.☆95Dec 1, 2022Updated 3 years ago
- Understanding and Improving Fast Adversarial Training [NeurIPS 2020]☆96Sep 23, 2021Updated 4 years ago
- [WWW '25] Model Supply Chain Poisoning: Backdooring Pre-trained Models via Embedding Indistinguishability☆18May 30, 2025Updated 9 months ago
- [ICLR 2025] Bridging and Modeling Correlations in Pairwise Data for Direct Preference Optimization☆12Jan 26, 2025Updated last year
- Decoupled Kullback-Leibler Divergence Loss (DKL), NeurIPS 2024 / Generalized Kullback-Leibler Divergence Loss (GKL)☆50Jul 21, 2025Updated 8 months ago
- Unofficial implementation of the DeepMind papers "Uncovering the Limits of Adversarial Training against Norm-Bounded Adversarial Examples…☆99Mar 4, 2022Updated 4 years ago
- [ICML 2025] UDora: A Unified Red Teaming Framework against LLM Agents☆32Jun 24, 2025Updated 8 months ago
- This repository is the official implementation of the paper "ASSET: Robust Backdoor Data Detection Across a Multiplicity of Deep Learning…☆19Jun 7, 2023Updated 2 years ago
- ☆29Nov 29, 2023Updated 2 years ago
- The official code of IEEE S&P 2024 paper "Why Does Little Robustness Help? A Further Step Towards Understanding Adversarial Transferabili…☆20Aug 22, 2024Updated last year
- Guided Adversarial Attack for Evaluating and Enhancing Adversarial Defenses, NeurIPS Spotlight 2020☆27Dec 23, 2020Updated 5 years ago