wwoods / adversarial-explanations-cifar
Code example for the paper, "Adversarial Explanations for Understanding Image Classification Decisions and Improved Neural Network Robustness."
☆23Updated 10 months ago
Alternatives and similar repositories for adversarial-explanations-cifar:
Users that are interested in adversarial-explanations-cifar are comparing it to the libraries listed below
- Figures & code from the paper "Shortcut Learning in Deep Neural Networks" (Nature Machine Intelligence 2020)☆95Updated 2 years ago
- 'Robust Semantic Interpretability: Revisiting Concept Activation Vectors' Official Implementation☆11Updated 4 years ago
- Adversarial Defense by Restricting the Hidden Space of Deep Neural Networks, in ICCV 2019☆59Updated 5 years ago
- Official PyTorch implementation for our ICCV 2019 paper - Fooling Network Interpretation in Image Classification☆24Updated 5 years ago
- Project page for our paper: Interpreting Adversarially Trained Convolutional Neural Networks☆65Updated 5 years ago
- [TNNLS 2019] Gaussian-based softmax: Improving Intra-class Compactness and Inter-class Separability of Features☆8Updated 5 years ago
- PyTorch Implementation of CVPR'19 (oral) - Mitigating Information Leakage in Image Representations: A Maximum Entropy Approach☆27Updated 5 years ago
- A PyTorch Implementation of a Large Margin Deep Networks for Classification☆23Updated 5 years ago
- This code reproduces the results of the paper, "Measuring Data Leakage in Machine-Learning Models with Fisher Information"☆49Updated 3 years ago
- Explaining Image Classifiers by Counterfactual Generation☆28Updated 2 years ago
- Accompanying code for the paper "Zero-shot Knowledge Transfer via Adversarial Belief Matching"☆140Updated 4 years ago
- Information Bottlenecks for Attribution☆77Updated 2 years ago
- CVPR'19 experiments with (on-manifold) adversarial examples.☆44Updated 4 years ago
- ☆39Updated 6 years ago
- Code for Fong and Vedaldi 2017, "Interpretable Explanations of Black Boxes by Meaningful Perturbation"☆30Updated 5 years ago
- Code for the Paper 'On the Connection Between Adversarial Robustness and Saliency Map Interpretability' by C. Etmann, S. Lunz, P. Maass, …☆16Updated 5 years ago
- Interpretation of Neural Network is Fragile☆36Updated 8 months ago
- Pre-Training Buys Better Robustness and Uncertainty Estimates (ICML 2019)☆100Updated 2 years ago
- Visual Explanation using Uncertainty based Class Activation Maps☆22Updated 4 years ago
- Code for the paper "Adversarial Training and Robustness for Multiple Perturbations", NeurIPS 2019☆47Updated 2 years ago
- REPresentAtion bIas Removal (REPAIR) of datasets☆56Updated last year
- Code implementing the experiments described in the NeurIPS 2018 paper "With Friends Like These, Who Needs Adversaries?".☆13Updated 4 years ago
- Code for Net2Vec: Quantifying and Explaining how Concepts are Encoded by Filters in Deep Neural Networks☆31Updated 6 years ago
- Code for "Learning Perceptually-Aligned Representations via Adversarial Robustness"☆160Updated 4 years ago
- Reverse Cross Entropy for Adversarial Detection (NeurIPS 2018)☆45Updated 3 years ago
- ☆73Updated 4 years ago
- Code for AAAI 2018 accepted paper: "Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing the…☆55Updated 2 years ago
- code release for the paper "On Completeness-aware Concept-Based Explanations in Deep Neural Networks"☆52Updated 2 years ago
- Code for the CVPR 2021 paper: Understanding Failures of Deep Networks via Robust Feature Extraction☆35Updated 2 years ago
- reference implementation for "explanations can be manipulated and geometry is to blame"☆36Updated 2 years ago