Ighina / CERTIFAI
A python implementation of CERTIFAI framework for machine learning models' explainability as discussed in https://www.aies-conference.com/2020/wp-content/papers/099.pdf
☆9Updated 2 years ago
Alternatives and similar repositories for CERTIFAI:
Users that are interested in CERTIFAI are comparing it to the libraries listed below
- Explore/examine/explain/expose your model with the explabox!☆15Updated 2 months ago
- Codes for reproducing the experimental results in "Proper Network Interpretability Helps Adversarial Robustness in Classification", publi…☆13Updated 4 years ago
- ☆11Updated 4 years ago
- [NeurIPS 2019] H. Chen*, H. Zhang*, S. Si, Y. Li, D. Boning and C.-J. Hsieh, Robustness Verification of Tree-based Models (*equal contrib…☆26Updated 5 years ago
- Adversarial Attacks on Post Hoc Explanation Techniques (LIME/SHAP)☆82Updated 2 years ago
- Adversarial detection and defense for deep learning systems using robust feature alignment☆15Updated 4 years ago
- ☆24Updated 4 years ago
- bayesian lime☆17Updated 5 months ago
- Fighting Gradients with Gradients: Dynamic Defenses against Adversarial Attacks☆38Updated 3 years ago
- This repository provides a PyTorch implementation of "Fooling Neural Network Interpretations via Adversarial Model Manipulation". Our pap…☆22Updated 4 years ago
- Library for training globally-robust neural networks.☆28Updated last year
- Code for "On the Trade-off between Adversarial and Backdoor Robustness" (NIPS 2020)☆17Updated 4 years ago
- ☆21Updated last year
- Implementation of Adversarial Debiasing in PyTorch to address Gender Bias☆30Updated 4 years ago
- Keeps track of popular provable training and verification approaches towards robust neural networks, including leaderboards on popular da…☆18Updated 7 months ago
- Implementation of https://github.com/PurduePAML/TrojanNN☆9Updated 6 years ago
- Codes for ICCV 2021 paper "AGKD-BML: Defense Against Adversarial Attack by Attention Guided Knowledge Distillation and Bi-directional Met…☆12Updated 2 years ago
- Source code of "Hold me tight! Influence of discriminative features on deep network boundaries"☆22Updated 3 years ago
- code release for the paper "On Completeness-aware Concept-Based Explanations in Deep Neural Networks"☆52Updated 2 years ago
- ☆25Updated 6 years ago
- ☆36Updated last year
- KNN Defense Against Clean Label Poisoning Attacks☆12Updated 3 years ago
- This is a benchmark to evaluate machine learning local explanaitons quality generated from any explainer for text and image data☆30Updated 3 years ago
- Detection of adversarial examples using influence functions and nearest neighbors☆33Updated 2 years ago
- HCOMP '22 -- Eliciting and Learning with Soft Labels from Every Annotator☆10Updated 2 years ago
- Adversarially Robust Transfer Learning with LWF loss applied to the deep feature representation (penultimate) layer☆18Updated 4 years ago
- Code relative to "Adversarial robustness against multiple and single $l_p$-threat models via quick fine-tuning of robust classifiers"☆18Updated 2 years ago
- ⚖️ Code for the paper "Ethical Adversaries: Towards Mitigating Unfairness with Adversarial Machine Learning".☆11Updated 2 years ago
- code for model-targeted poisoning☆12Updated last year
- Creating and defending against adversarial examples☆42Updated 6 years ago