hbaniecki / adversarial-explainable-ai
π‘ Adversarial attacks on explanations and how to defend them
β314Updated 5 months ago
Alternatives and similar repositories for adversarial-explainable-ai:
Users that are interested in adversarial-explainable-ai are comparing it to the libraries listed below
- RobustBench: a standardized adversarial robustness benchmark [NeurIPS 2021 Benchmarks and Datasets Track]β706Updated last month
- reference implementation for "explanations can be manipulated and geometry is to blame"β36Updated 2 years ago
- Adversarial Attacks on Post Hoc Explanation Techniques (LIME/SHAP)β82Updated 2 years ago
- OpenXAI : Towards a Transparent Evaluation of Model Explanationsβ245Updated 8 months ago
- A Python library for Secure and Explainable Machine Learningβ176Updated 3 months ago
- A curated list of awesome Fairness in AI resourcesβ320Updated last year
- Related papers for robust machine learningβ569Updated last year
- This repository provides simple PyTorch implementations for adversarial training methods on CIFAR-10.β165Updated 4 years ago
- A library for experimenting with, training and evaluating neural networks, with a focus on adversarial robustness.β934Updated last year
- A curated list of papers on adversarial machine learning (adversarial examples and defense methods).β210Updated 2 years ago
- Interesting resources related to Explainable Artificial Intelligence, Interpretable Machine Learning, Interactive Machine Learning, Humanβ¦β73Updated 2 years ago
- [ICLR 2020] A repository for extremely fast adversarial training using FGSMβ443Updated 9 months ago
- LaTeX source for the paper "On Evaluating Adversarial Robustness"β255Updated 4 years ago
- All about explainable AI, algorithmic fairness and moreβ107Updated last year
- Provable adversarial robustness at ImageNet scaleβ387Updated 5 years ago
- β144Updated 6 months ago
- A unified benchmark problem for data poisoning attacksβ155Updated last year
- Code for "On Adaptive Attacks to Adversarial Example Defenses"β87Updated 4 years ago
- Library containing PyTorch implementations of various adversarial attacks and resourcesβ155Updated last week
- Quantus is an eXplainable AI toolkit for responsible evaluation of neural network explanationsβ598Updated 2 months ago
- Code relative to "Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks"β694Updated 11 months ago
- β123Updated 3 years ago
- [NeurIPS 2019] H. Chen*, H. Zhang*, S. Si, Y. Li, D. Boning and C.-J. Hsieh, Robustness Verification of Tree-based Models (*equal contribβ¦β27Updated 5 years ago
- Reference tables to introduce and organize evaluation methods and measures for explainable machine learning systemsβ74Updated 3 years ago
- A curated list of trustworthy deep learning papers. Daily updating...β366Updated last week
- Creating and defending against adversarial examplesβ42Updated 6 years ago
- β157Updated 4 years ago
- Attacks Which Do Not Kill Training Make Adversarial Learning Stronger (ICML2020 Paper)β125Updated last year
- A curated list of academic events on AI Security & Privacyβ150Updated 8 months ago
- PhD/MSc course on Machine Learning Security (Univ. Cagliari)β209Updated 4 months ago