rachtibat / LRP-eXplains-TransformersLinks
Layer-wise Relevance Propagation for Large Language Models and Vision Transformers [ICML 2024]
☆211Updated 5 months ago
Alternatives and similar repositories for LRP-eXplains-Transformers
Users that are interested in LRP-eXplains-Transformers are comparing it to the libraries listed below
Sorting:
- Mechanistic understanding and validation of large AI models with SemanticLens☆47Updated last week
- [NeurIPS 2024] CoSy is an automatic evaluation framework for textual explanations of neurons.☆18Updated 5 months ago
- An eXplainable AI toolkit with Concept Relevance Propagation and Relevance Maximization☆139Updated last year
- Code for the paper: Discover-then-Name: Task-Agnostic Concept Bottlenecks via Automated Concept Discovery. ECCV 2024.☆52Updated last year
- A repository for summaries of recent explainable AI/Interpretable ML approaches☆87Updated last year
- MetaQuantus is an XAI performance tool to identify reliable evaluation metrics☆39Updated last year
- A toolkit for quantitative evaluation of data attribution methods.☆54Updated 4 months ago
- Zennit is a high-level framework in Python using PyTorch for explaining/exploring neural networks using attribution methods like LRP.☆238Updated 4 months ago
- Using sparse coding to find distributed representations used by neural networks.☆287Updated 2 years ago
- A resource repository for representation engineering in large language models☆143Updated last year
- 👋 Overcomplete is a Vision-based SAE Toolbox☆106Updated last week
- ☆28Updated last year
- ☆32Updated last year
- ☆136Updated 3 weeks ago
- DataInf: Efficiently Estimating Data Influence in LoRA-tuned LLMs and Diffusion Models (ICLR 2024)☆78Updated last year
- Official Code Implementation of the paper : XAI for Transformers: Better Explanations through Conservative Propagation☆67Updated 3 years ago
- This repository collects all relevant resources about interpretability in LLMs☆387Updated last year
- Code for the paper "Post-hoc Concept Bottleneck Models". Spotlight @ ICLR 2023☆89Updated last year
- [ICLR 23] A new framework to transform any neural networks into an interpretable concept-bottleneck-model (CBM) without needing labeled c…☆126Updated last year
- Reveal to Revise: An Explainable AI Life Cycle for Iterative Bias Correction of Deep Models. Paper presented at MICCAI 2023 conference.☆20Updated last year
- Sparse Autoencoder for Mechanistic Interpretability☆285Updated last year
- A basic implementation of Layer-wise Relevance Propagation (LRP) in PyTorch.☆100Updated 3 years ago
- A fast, effective data attribution method for neural networks in PyTorch☆222Updated last year
- [NeurIPS 2025 MechInterp Workshop - Spotlight] Official implementation of the paper "RelP: Faithful and Efficient Circuit Discovery in La…☆21Updated last month
- Repository for PURE: Turning Polysemantic Neurons Into Pure Features by Identifying Relevant Circuits, accepted at CVPR 2024 XAI4CV Works…☆19Updated last year
- A simple PyTorch implementation of influence functions.☆91Updated last year
- Sparse probing paper full code.☆65Updated last year
- OpenXAI : Towards a Transparent Evaluation of Model Explanations☆252Updated last year
- Concept Bottleneck Models, ICML 2020☆233Updated 2 years ago
- ☆366Updated 3 months ago