hbaniecki / adversarial-explainable-ai
π‘ Adversarial attacks on explanations and how to defend them
β309Updated 2 months ago
Alternatives and similar repositories for adversarial-explainable-ai:
Users that are interested in adversarial-explainable-ai are comparing it to the libraries listed below
- OpenXAI : Towards a Transparent Evaluation of Model Explanationsβ239Updated 6 months ago
- reference implementation for "explanations can be manipulated and geometry is to blame"β36Updated 2 years ago
- A curated list of papers on adversarial machine learning (adversarial examples and defense methods).β210Updated 2 years ago
- Related papers for robust machine learningβ566Updated last year
- Adversarial Attacks on Post Hoc Explanation Techniques (LIME/SHAP)β82Updated 2 years ago
- Interesting resources related to Explainable Artificial Intelligence, Interpretable Machine Learning, Interactive Machine Learning, Humanβ¦β73Updated 2 years ago
- All about explainable AI, algorithmic fairness and moreβ107Updated last year
- A curated list of awesome Fairness in AI resourcesβ314Updated last year
- Code for "On Adaptive Attacks to Adversarial Example Defenses"β86Updated 4 years ago
- RobustBench: a standardized adversarial robustness benchmark [NeurIPS 2021 Benchmarks and Datasets Track]β693Updated 2 weeks ago
- A Python library for Secure and Explainable Machine Learningβ170Updated 3 weeks ago
- This repository provides simple PyTorch implementations for adversarial training methods on CIFAR-10.β162Updated 4 years ago
- List of relevant resources for machine learning from explanatory supervisionβ155Updated last month
- β120Updated 3 years ago
- Using / reproducing ACD from the paper "Hierarchical interpretations for neural network predictions" π§ (ICLR 2019)β128Updated 3 years ago
- CARLA: A Python Library to Benchmark Algorithmic Recourse and Counterfactual Explanation Algorithmsβ286Updated last year
- A library for experimenting with, training and evaluating neural networks, with a focus on adversarial robustness.β926Updated last year
- Quantus is an eXplainable AI toolkit for responsible evaluation of neural network explanationsβ584Updated 2 weeks ago
- β142Updated 4 months ago
- Library containing PyTorch implementations of various adversarial attacks and resourcesβ150Updated last week
- A unified benchmark problem for data poisoning attacksβ152Updated last year
- Repository of the paper "Imperceptible Adversarial Attacks on Tabular Data" presented at NeurIPS 2019 Workshop on Robust AI in Financial β¦β15Updated 3 years ago
- Provable adversarial robustness at ImageNet scaleβ383Updated 5 years ago
- Code for using CDEP from the paper "Interpretations are useful: penalizing explanations to align neural networks with prior knowledge" htβ¦β127Updated 3 years ago
- Empirical tricks for training robust models (ICLR 2021)β253Updated last year
- Reference tables to introduce and organize evaluation methods and measures for explainable machine learning systemsβ74Updated 2 years ago
- π± A curated list of data valuation (DV) to design your next data marketplaceβ115Updated 3 months ago
- Papers and code of Explainable AI esp. w.r.t. Image classificiationβ203Updated 2 years ago
- Code relative to "Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks"β679Updated 9 months ago
- A curated list of trustworthy deep learning papers. Daily updating...β354Updated last week