Ighina / CERTIFAILinks
A python implementation of CERTIFAI framework for machine learning models' explainability as discussed in https://www.aies-conference.com/2020/wp-content/papers/099.pdf
☆11Updated 3 years ago
Alternatives and similar repositories for CERTIFAI
Users that are interested in CERTIFAI are comparing it to the libraries listed below
Sorting:
- reference implementation for "explanations can be manipulated and geometry is to blame"☆37Updated 3 years ago
- ☆16Updated 4 years ago
- Adversarial Attacks on Post Hoc Explanation Techniques (LIME/SHAP)☆84Updated 3 years ago
- Codes for reproducing the experimental results in "Proper Network Interpretability Helps Adversarial Robustness in Classification", publi…☆13Updated 5 years ago
- Attack benchmark repository☆21Updated 2 months ago
- KNN Defense Against Clean Label Poisoning Attacks☆13Updated 4 years ago
- Adversarial detection and defense for deep learning systems using robust feature alignment☆18Updated 5 years ago
- ☆31Updated 4 years ago
- Creating and defending against adversarial examples☆41Updated 7 years ago
- Implementation of Adversarial Debiasing in PyTorch to address Gender Bias☆31Updated 5 years ago
- Code for "Differential Privacy Has Disparate Impact on Model Accuracy" NeurIPS'19☆33Updated 4 years ago
- Invertible Concept-based Explanation (ICE)☆19Updated 3 months ago
- code release for the paper "On Completeness-aware Concept-Based Explanations in Deep Neural Networks"☆54Updated 3 years ago
- Codes for ICCV 2021 paper "AGKD-BML: Defense Against Adversarial Attack by Attention Guided Knowledge Distillation and Bi-directional Met…☆12Updated 3 years ago
- CLEVER (Cross-Lipschitz Extreme Value for nEtwork Robustness) is a robustness metric for deep neural networks☆63Updated 4 years ago
- ☆37Updated 2 years ago
- Detection of adversarial examples using influence functions and nearest neighbors☆37Updated 3 years ago
- code for model-targeted poisoning☆12Updated 2 years ago
- Python implementation for evaluating explanations presented in "On the (In)fidelity and Sensitivity for Explanations" in NeurIPS 2019 for…☆25Updated 3 years ago
- Implementation for "Defense-VAE: A Fast and Accurate Defense against Adversarial Attacks"☆14Updated 5 years ago
- Craft poisoned data using MetaPoison☆54Updated 4 years ago
- ConvexPolytopePosioning☆37Updated 6 years ago
- Code for "On Adaptive Attacks to Adversarial Example Defenses"☆87Updated 4 years ago
- 💡 Adversarial attacks on explanations and how to defend them☆334Updated last year
- Implemented CURE algorithm from robustness via curvature regularization and vice versa☆32Updated 3 years ago
- Fighting Gradients with Gradients: Dynamic Defenses against Adversarial Attacks☆38Updated 4 years ago
- Foolbox implementation for NeurIPS 2021 Paper: "Fast Minimum-norm Adversarial Attacks through Adaptive Norm Constraints".☆24Updated 3 years ago
- PrivGAN: Protecting GANs from membership inference attacks at low cost☆36Updated last year
- Code for our paper☆13Updated 3 years ago
- FairBatch: Batch Selection for Model Fairness (ICLR 2021)☆19Updated 2 years ago