pankessel / adv_explanation_ref
reference implementation for "explanations can be manipulated and geometry is to blame"
☆36Updated 2 years ago
Alternatives and similar repositories for adv_explanation_ref:
Users that are interested in adv_explanation_ref are comparing it to the libraries listed below
- code release for the paper "On Completeness-aware Concept-Based Explanations in Deep Neural Networks"☆53Updated 3 years ago
- ☆51Updated 4 years ago
- This repository provides a PyTorch implementation of "Fooling Neural Network Interpretations via Adversarial Model Manipulation". Our pap…☆22Updated 4 years ago
- Python implementation for evaluating explanations presented in "On the (In)fidelity and Sensitivity for Explanations" in NeurIPS 2019 for…☆25Updated 3 years ago
- Adversarially Robust Neural Network on MNIST.☆64Updated 3 years ago
- The Pitfalls of Simplicity Bias in Neural Networks [NeurIPS 2020] (http://arxiv.org/abs/2006.07710v2)☆39Updated last year
- Code for using CDEP from the paper "Interpretations are useful: penalizing explanations to align neural networks with prior knowledge" ht…☆127Updated 4 years ago
- Code and data for the ICLR 2021 paper "Perceptual Adversarial Robustness: Defense Against Unseen Threat Models".☆55Updated 3 years ago
- Original dataset release for CIFAR-10H☆82Updated 4 years ago
- ☆37Updated 2 years ago
- ☆73Updated 5 years ago
- Explaining Image Classifiers by Counterfactual Generation☆28Updated 3 years ago
- Implemented CURE algorithm from robustness via curvature regularization and vice versa☆31Updated 2 years ago
- Invertible Concept-based Explanation (ICE)☆18Updated 3 years ago
- Source code for "Neural Anisotropy Directions"☆15Updated 4 years ago
- Code for "Testing Robustness Against Unforeseen Adversaries"☆81Updated 9 months ago
- ☆109Updated 2 years ago
- Understanding and Improving Fast Adversarial Training [NeurIPS 2020]☆95Updated 3 years ago
- ☆55Updated 4 years ago
- Code for the ICLR 2022 paper. Salient Imagenet: How to discover spurious features in deep learning?☆40Updated 2 years ago
- Using / reproducing ACD from the paper "Hierarchical interpretations for neural network predictions" 🧠 (ICLR 2019)☆128Updated 3 years ago
- Towards Automatic Concept-based Explanations☆159Updated last year
- Interpretation of Neural Network is Fragile☆36Updated last year
- 💡 Adversarial attacks on explanations and how to defend them☆314Updated 5 months ago
- Semisupervised learning for adversarial robustness https://arxiv.org/pdf/1905.13736.pdf☆141Updated 5 years ago
- Reference tables to introduce and organize evaluation methods and measures for explainable machine learning systems☆74Updated 3 years ago
- Code for "On Adaptive Attacks to Adversarial Example Defenses"☆87Updated 4 years ago
- ☆140Updated 4 years ago
- Unofficial implementation of the DeepMind papers "Uncovering the Limits of Adversarial Training against Norm-Bounded Adversarial Examples…☆96Updated 3 years ago
- ☆38Updated 3 years ago