hbaniecki / adversarial-explainable-aiLinks
π‘ Adversarial attacks on explanations and how to defend them
β315Updated 6 months ago
Alternatives and similar repositories for adversarial-explainable-ai
Users that are interested in adversarial-explainable-ai are comparing it to the libraries listed below
Sorting:
- A curated list of papers on adversarial machine learning (adversarial examples and defense methods).β210Updated 3 years ago
- Adversarial Attacks on Post Hoc Explanation Techniques (LIME/SHAP)β82Updated 2 years ago
- A Python library for Secure and Explainable Machine Learningβ176Updated 4 months ago
- reference implementation for "explanations can be manipulated and geometry is to blame"β36Updated 2 years ago
- OpenXAI : Towards a Transparent Evaluation of Model Explanationsβ247Updated 9 months ago
- RobustBench: a standardized adversarial robustness benchmark [NeurIPS 2021 Benchmarks and Datasets Track]β713Updated 2 months ago
- Interesting resources related to Explainable Artificial Intelligence, Interpretable Machine Learning, Interactive Machine Learning, Humanβ¦β72Updated 2 years ago
- Reference tables to introduce and organize evaluation methods and measures for explainable machine learning systemsβ74Updated 3 years ago
- Related papers for robust machine learningβ568Updated 2 years ago
- A curated list of awesome Fairness in AI resourcesβ321Updated last year
- A library for experimenting with, training and evaluating neural networks, with a focus on adversarial robustness.β937Updated last year
- All about explainable AI, algorithmic fairness and moreβ108Updated last year
- CARLA: A Python Library to Benchmark Algorithmic Recourse and Counterfactual Explanation Algorithmsβ289Updated last year
- Quantus is an eXplainable AI toolkit for responsible evaluation of neural network explanationsβ599Updated 3 months ago
- Library containing PyTorch implementations of various adversarial attacks and resourcesβ157Updated last month
- LOcal Rule-based Exlanationsβ53Updated last year
- Repository of the paper "Imperceptible Adversarial Attacks on Tabular Data" presented at NeurIPS 2019 Workshop on Robust AI in Financial β¦β15Updated 3 years ago
- pyDVL is a library of stable implementations of algorithms for data valuation and influence function computationβ129Updated 3 weeks ago
- A curated list of trustworthy deep learning papers. Daily updating...β367Updated last week
- β144Updated 7 months ago
- A repository to quickly generate synthetic data and associated trojaned deep learning modelsβ77Updated last year
- This repository provides simple PyTorch implementations for adversarial training methods on CIFAR-10.β167Updated 4 years ago
- A unified benchmark problem for data poisoning attacksβ155Updated last year
- Detect model's attentionβ165Updated 4 years ago
- Code for "On Adaptive Attacks to Adversarial Example Defenses"β87Updated 4 years ago
- An awesome list of papers on privacy attacks against machine learningβ608Updated last year
- β33Updated 11 months ago
- π Influenciae is a Tensorflow Toolbox for Influence Functionsβ63Updated last year
- LaTeX source for the paper "On Evaluating Adversarial Robustness"β255Updated 4 years ago
- Code relative to "Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks"β702Updated last year