rachtibat / LRP-eXplains-Transformers
Layer-Wise Relevance Propagation for Large Language Models and Vision Transformers [ICML 2024]
☆128Updated 2 weeks ago
Alternatives and similar repositories for LRP-eXplains-Transformers:
Users that are interested in LRP-eXplains-Transformers are comparing it to the libraries listed below
- An eXplainable AI toolkit with Concept Relevance Propagation and Relevance Maximization☆124Updated 8 months ago
- Official Code Implementation of the paper : XAI for Transformers: Better Explanations through Conservative Propagation☆63Updated 3 years ago
- Zennit is a high-level framework in Python using PyTorch for explaining/exploring neural networks using attribution methods like LRP.☆210Updated 7 months ago
- CoSy: Evaluating Textual Explanations☆14Updated last month
- MetaQuantus is an XAI performance tool to identify reliable evaluation metrics☆33Updated 10 months ago
- A toolkit for quantitative evaluation of data attribution methods.☆40Updated last week
- Reveal to Revise: An Explainable AI Life Cycle for Iterative Bias Correction of Deep Models. Paper presented at MICCAI 2023 conference.☆19Updated last year
- Code for the paper: Discover-then-Name: Task-Agnostic Concept Bottlenecks via Automated Concept Discovery. ECCV 2024.☆35Updated 4 months ago
- A basic implementation of Layer-wise Relevance Propagation (LRP) in PyTorch.☆87Updated 2 years ago
- A simple PyTorch implementation of influence functions.☆84Updated 8 months ago
- ☆88Updated 2 weeks ago
- A repository for summaries of recent explainable AI/Interpretable ML approaches☆73Updated 5 months ago
- Dataset and code for the CLEVR-XAI dataset.☆31Updated last year
- 👋 Code for : "CRAFT: Concept Recursive Activation FacTorization for Explainability" (CVPR 2023)☆62Updated last year
- Code for the paper "Post-hoc Concept Bottleneck Models". Spotlight @ ICLR 2023☆73Updated 9 months ago
- This repository collects all relevant resources about interpretability in LLMs☆322Updated 4 months ago
- A new framework to transform any neural networks into an interpretable concept-bottleneck-model (CBM) without needing labeled concept dat…☆88Updated 11 months ago
- Using sparse coding to find distributed representations used by neural networks.☆217Updated last year
- ☆58Updated this week
- A fast, effective data attribution method for neural networks in PyTorch☆195Updated 3 months ago
- Pruning By Explaining Revisited: Optimizing Attribution Methods to Prune CNNs and Transformers, Paper accepted at eXCV workshop of ECCV 2…☆18Updated last month
- A resource repository for representation engineering in large language models☆107Updated 3 months ago
- DataInf: Efficiently Estimating Data Influence in LoRA-tuned LLMs and Diffusion Models (ICLR 2024)☆62Updated 5 months ago
- OpenXAI : Towards a Transparent Evaluation of Model Explanations☆239Updated 6 months ago
- Sparse Autoencoder for Mechanistic Interpretability☆217Updated 7 months ago
- Repository for PURE: Turning Polysemantic Neurons Into Pure Features by Identifying Relevant Circuits, accepted at CVPR 2024 XAI4CV Works…☆12Updated 9 months ago
- Concept Bottleneck Models, ICML 2020☆191Updated 2 years ago
- Interpretability for sequence generation models 🐛 🔍☆405Updated 3 months ago