hbaniecki / adversarial-explainable-ai
π‘ Adversarial attacks on explanations and how to defend them
β304Updated last month
Alternatives and similar repositories for adversarial-explainable-ai:
Users that are interested in adversarial-explainable-ai are comparing it to the libraries listed below
- OpenXAI : Towards a Transparent Evaluation of Model Explanationsβ237Updated 5 months ago
- All about explainable AI, algorithmic fairness and moreβ107Updated last year
- Adversarial Attacks on Post Hoc Explanation Techniques (LIME/SHAP)β81Updated 2 years ago
- Related papers for robust machine learningβ566Updated last year
- Quantus is an eXplainable AI toolkit for responsible evaluation of neural network explanationsβ573Updated 2 months ago
- A curated list of papers on adversarial machine learning (adversarial examples and defense methods).β211Updated 2 years ago
- A Python library for Secure and Explainable Machine Learningβ162Updated 2 months ago
- RobustBench: a standardized adversarial robustness benchmark [NeurIPS 2021 Benchmarks and Datasets Track]β689Updated 3 weeks ago
- Interesting resources related to Explainable Artificial Intelligence, Interpretable Machine Learning, Interactive Machine Learning, Humanβ¦β72Updated 2 years ago
- Datasets derived from US census dataβ247Updated 8 months ago
- Library containing PyTorch implementations of various adversarial attacks and resourcesβ149Updated 3 weeks ago
- β140Updated 3 months ago
- Code for "On Adaptive Attacks to Adversarial Example Defenses"β85Updated 3 years ago
- A curated list of trustworthy deep learning papers. Daily updating...β349Updated last week
- A library for experimenting with, training and evaluating neural networks, with a focus on adversarial robustness.β923Updated last year
- XAI-Bench is a library for benchmarking feature attribution explainability techniquesβ60Updated last year
- Reference tables to introduce and organize evaluation methods and measures for explainable machine learning systemsβ74Updated 2 years ago
- A unified benchmark problem for data poisoning attacksβ152Updated last year
- A curated list of awesome Fairness in AI resourcesβ313Updated last year
- List of relevant resources for machine learning from explanatory supervisionβ155Updated 6 months ago
- Code for using CDEP from the paper "Interpretations are useful: penalizing explanations to align neural networks with prior knowledge" htβ¦β128Updated 3 years ago
- Papers and code of Explainable AI esp. w.r.t. Image classificiationβ200Updated 2 years ago
- π± A curated list of data valuation (DV) to design your next data marketplaceβ114Updated 2 months ago
- Uncertainty Quantification 360 (UQ360) is an extensible open-source toolkit that can help you estimate, communicate and use uncertainty iβ¦β257Updated 4 months ago
- reference implementation for "explanations can be manipulated and geometry is to blame"β36Updated 2 years ago
- LOcal Rule-based Exlanationsβ50Updated last year
- β303Updated last month
- A library for running membership inference attacks against ML modelsβ141Updated 2 years ago
- Using / reproducing ACD from the paper "Hierarchical interpretations for neural network predictions" π§ (ICLR 2019)β128Updated 3 years ago
- β119Updated 3 years ago