rachtibat / LRP-eXplains-Transformers
Layer-Wise Relevance Propagation for Large Language Models and Vision Transformers [ICML 2024]
☆149Updated last month
Alternatives and similar repositories for LRP-eXplains-Transformers:
Users that are interested in LRP-eXplains-Transformers are comparing it to the libraries listed below
- An eXplainable AI toolkit with Concept Relevance Propagation and Relevance Maximization☆124Updated 10 months ago
- Zennit is a high-level framework in Python using PyTorch for explaining/exploring neural networks using attribution methods like LRP.☆222Updated 9 months ago
- ☆11Updated last week
- CoSy: Evaluating Textual Explanations☆16Updated 3 months ago
- MetaQuantus is an XAI performance tool to identify reliable evaluation metrics☆34Updated last year
- A toolkit for quantitative evaluation of data attribution methods.☆44Updated last week
- Official Code Implementation of the paper : XAI for Transformers: Better Explanations through Conservative Propagation☆63Updated 3 years ago
- A basic implementation of Layer-wise Relevance Propagation (LRP) in PyTorch.☆89Updated 2 years ago
- ☆85Updated last week
- Using sparse coding to find distributed representations used by neural networks.☆236Updated last year
- Reveal to Revise: An Explainable AI Life Cycle for Iterative Bias Correction of Deep Models. Paper presented at MICCAI 2023 conference.☆19Updated last year
- Code for the paper: Discover-then-Name: Task-Agnostic Concept Bottlenecks via Automated Concept Discovery. ECCV 2024.☆39Updated 5 months ago
- Concept Relevance Propagation for Localization Models, accepted at SAIAD workshop at CVPR 2023.☆14Updated last year
- A fast, effective data attribution method for neural networks in PyTorch☆204Updated 5 months ago
- Pruning By Explaining Revisited: Optimizing Attribution Methods to Prune CNNs and Transformers, Paper accepted at eXCV workshop of ECCV 2…☆20Updated 3 months ago
- Materials for EACL2024 tutorial: Transformer-specific Interpretability☆50Updated last year
- 👋 Overcomplete is a Vision-based SAE Toolbox☆51Updated 3 weeks ago
- Interpretability for sequence generation models 🐛 🔍☆412Updated 5 months ago
- [ICLR 23 spotlight] An automatic and efficient tool to describe functionalities of individual neurons in DNNs☆48Updated last year
- DataInf: Efficiently Estimating Data Influence in LoRA-tuned LLMs and Diffusion Models (ICLR 2024)☆63Updated 6 months ago
- Quantus is an eXplainable AI toolkit for responsible evaluation of neural network explanations☆598Updated 2 months ago
- Sparse Autoencoder for Mechanistic Interpretability☆241Updated 9 months ago
- A repository for summaries of recent explainable AI/Interpretable ML approaches☆74Updated 6 months ago
- 👋 Code for : "CRAFT: Concept Recursive Activation FacTorization for Explainability" (CVPR 2023)☆62Updated last year
- This repository collects all relevant resources about interpretability in LLMs☆341Updated 5 months ago
- AI Logging for Interpretability and Explainability🔬☆111Updated 10 months ago
- ☆91Updated 2 months ago
- Sparse probing paper full code.☆55Updated last year
- Code for the paper "Post-hoc Concept Bottleneck Models". Spotlight @ ICLR 2023☆77Updated 11 months ago
- NeuroSurgeon is a package that enables researchers to uncover and manipulate subnetworks within models in Huggingface Transformers☆41Updated 2 months ago