dmitrykazhdan / CMELinks
CME: Concept-based Model Extraction
☆12Updated 5 years ago
Alternatives and similar repositories for CME
Users that are interested in CME are comparing it to the libraries listed below
Sorting:
- DISSECT: Disentangled Simultaneous Explanations via Concept Traversals☆12Updated last year
- code release for the paper "On Completeness-aware Concept-Based Explanations in Deep Neural Networks"☆54Updated 3 years ago
- Library implementing state-of-the-art Concept-based and Disentanglement Learning methods for Explainable AI☆55Updated 3 years ago
- ☆38Updated 4 years ago
- ☆46Updated 2 years ago
- LISA for ICML 2022☆52Updated 2 years ago
- CVPR'19 experiments with (on-manifold) adversarial examples.☆45Updated 5 years ago
- Explaining Image Classifiers by Counterfactual Generation☆28Updated 3 years ago
- Code for Environment Inference for Invariant Learning (ICML 2021 Paper)☆51Updated 4 years ago
- 'Robust Semantic Interpretability: Revisiting Concept Activation Vectors' Official Implementation☆11Updated 5 years ago
- [ICLR'22] Self-supervised learning optimally robust representations for domain shift.☆24Updated 3 years ago
- Self-Explaining Neural Networks☆43Updated 5 years ago
- ☆36Updated 4 years ago
- ☆31Updated 4 years ago
- Code for the ICLR 2022 paper "Attention-based interpretability with Concept Transformers"☆42Updated 2 months ago
- Code for "Generative causal explanations of black-box classifiers"☆35Updated 4 years ago
- Code for "Interpretable Image Recognition with Hierarchical Prototypes"☆19Updated 6 years ago
- Do input gradients highlight discriminative features? [NeurIPS 2021] (https://arxiv.org/abs/2102.12781)☆13Updated 2 years ago
- Repository for our NeurIPS 2022 paper "Concept Embedding Models", our NeurIPS 2023 paper "Learning to Receive Help", and our ICML 2025 pa…☆71Updated 2 months ago
- Repo for the paper: "Agree to Disagree: Diversity through Disagreement for Better Transferability"☆36Updated 3 years ago
- ☆72Updated 5 years ago
- This repository provides a PyTorch implementation of "Fooling Neural Network Interpretations via Adversarial Model Manipulation". Our pap…☆23Updated 4 years ago
- This repository contains the implementation of Concept Activation Regions, a new framework to explain deep neural networks with human con…☆14Updated 3 years ago
- Official repository for the AAAI-21 paper 'Explainable Models with Consistent Interpretations'☆18Updated 3 years ago
- Distributional Shapley: A Distributional Framework for Data Valuation☆30Updated last year
- Self-Explaining Neural Networks☆13Updated 2 years ago
- Learning perturbation sets for robust machine learning☆65Updated 4 years ago
- ☆38Updated 5 years ago
- On the Loss Landscape of Adversarial Training: Identifying Challenges and How to Overcome Them [NeurIPS 2020]☆36Updated 4 years ago
- Improving the Fairness of Chest X-ray Classifiers☆14Updated 3 years ago