hbaniecki / adversarial-explainable-aiLinks
π‘ Adversarial attacks on explanations and how to defend them
β318Updated 6 months ago
Alternatives and similar repositories for adversarial-explainable-ai
Users that are interested in adversarial-explainable-ai are comparing it to the libraries listed below
Sorting:
- Interesting resources related to Explainable Artificial Intelligence, Interpretable Machine Learning, Interactive Machine Learning, Humanβ¦β73Updated 2 years ago
- A Python library for Secure and Explainable Machine Learningβ180Updated 4 months ago
- A curated list of papers on adversarial machine learning (adversarial examples and defense methods).β210Updated 3 years ago
- RobustBench: a standardized adversarial robustness benchmark [NeurIPS 2021 Benchmarks and Datasets Track]β720Updated 2 months ago
- Adversarial Attacks on Post Hoc Explanation Techniques (LIME/SHAP)β82Updated 2 years ago
- OpenXAI : Towards a Transparent Evaluation of Model Explanationsβ247Updated 10 months ago
- All about explainable AI, algorithmic fairness and moreβ109Updated last year
- β125Updated 3 years ago
- Related papers for robust machine learningβ566Updated 2 years ago
- Quantus is an eXplainable AI toolkit for responsible evaluation of neural network explanationsβ603Updated 4 months ago
- reference implementation for "explanations can be manipulated and geometry is to blame"β36Updated 2 years ago
- β145Updated 8 months ago
- Library containing PyTorch implementations of various adversarial attacks and resourcesβ158Updated last month
- Code for "On Adaptive Attacks to Adversarial Example Defenses"β87Updated 4 years ago
- A unified benchmark problem for data poisoning attacksβ156Updated last year
- A library for experimenting with, training and evaluating neural networks, with a focus on adversarial robustness.β938Updated last year
- Code relative to "Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks"β707Updated last year
- This repository provides simple PyTorch implementations for adversarial training methods on CIFAR-10.β167Updated 4 years ago
- PhD/MSc course on Machine Learning Security (Univ. Cagliari)β210Updated 2 weeks ago
- This is a collection of papers and other resources related to fairness.β94Updated 2 years ago
- [ICLR 2020] A repository for extremely fast adversarial training using FGSMβ444Updated 10 months ago
- CARLA: A Python Library to Benchmark Algorithmic Recourse and Counterfactual Explanation Algorithmsβ290Updated last year
- Papers and code of Explainable AI esp. w.r.t. Image classificiationβ212Updated 2 years ago
- TrojanZoo provides a universal pytorch platform to conduct security researches (especially backdoor attacks/defenses) of image classificaβ¦β289Updated 10 months ago
- Witches' Brew: Industrial Scale Data Poisoning via Gradient Matchingβ103Updated 10 months ago
- Empirical tricks for training robust models (ICLR 2021)β254Updated 2 years ago
- A repository to quickly generate synthetic data and associated trojaned deep learning modelsβ77Updated 2 years ago
- LOcal Rule-based Exlanationsβ51Updated last year
- Provable adversarial robustness at ImageNet scaleβ389Updated 6 years ago
- A repo for transfer learning with deep tabular modelsβ104Updated 2 years ago