rachtibat / LRP-eXplains-TransformersLinks
Layer-wise Relevance Propagation for Large Language Models and Vision Transformers [ICML 2024]
☆192Updated 3 months ago
Alternatives and similar repositories for LRP-eXplains-Transformers
Users that are interested in LRP-eXplains-Transformers are comparing it to the libraries listed below
Sorting:
- Using sparse coding to find distributed representations used by neural networks.☆274Updated last year
- MetaQuantus is an XAI performance tool to identify reliable evaluation metrics☆39Updated last year
- An eXplainable AI toolkit with Concept Relevance Propagation and Relevance Maximization☆133Updated last year
- A fast, effective data attribution method for neural networks in PyTorch☆218Updated 10 months ago
- Official Code Implementation of the paper : XAI for Transformers: Better Explanations through Conservative Propagation☆65Updated 3 years ago
- ☆32Updated 10 months ago
- This repository collects all relevant resources about interpretability in LLMs☆374Updated 11 months ago
- Official implementation of the paper "RelP: Faithful and Efficient Circuit Discovery via Relevance Patching"☆15Updated last month
- ☆127Updated last week
- Code for the paper: Discover-then-Name: Task-Agnostic Concept Bottlenecks via Automated Concept Discovery. ECCV 2024.☆49Updated 11 months ago
- A toolkit for quantitative evaluation of data attribution methods.☆53Updated 2 months ago
- A repository for summaries of recent explainable AI/Interpretable ML approaches☆84Updated last year
- A resource repository for representation engineering in large language models☆138Updated 10 months ago
- [NeurIPS 2024] CoSy is an automatic evaluation framework for textual explanations of neurons.☆18Updated 3 months ago
- Zennit is a high-level framework in Python using PyTorch for explaining/exploring neural networks using attribution methods like LRP.☆233Updated 2 months ago
- Mechanistic understanding and validation of large AI models with SemanticLens☆37Updated 3 weeks ago
- DataInf: Efficiently Estimating Data Influence in LoRA-tuned LLMs and Diffusion Models (ICLR 2024)☆76Updated last year
- A simple PyTorch implementation of influence functions.☆91Updated last year
- Sparse Autoencoder for Mechanistic Interpretability☆269Updated last year
- Materials for EACL2024 tutorial: Transformer-specific Interpretability☆60Updated last year
- ☆151Updated 2 years ago
- 👋 Overcomplete is a Vision-based SAE Toolbox☆90Updated 2 months ago
- Code for the paper "Post-hoc Concept Bottleneck Models". Spotlight @ ICLR 2023☆84Updated last year
- [ICLR 2025] General-purpose activation steering library☆108Updated 3 weeks ago
- Reveal to Revise: An Explainable AI Life Cycle for Iterative Bias Correction of Deep Models. Paper presented at MICCAI 2023 conference.☆20Updated last year
- ☆202Updated 10 months ago
- Conformal Language Modeling☆32Updated last year
- A Python Data Valuation Package☆30Updated 2 years ago
- ☆348Updated last month
- ☆28Updated 11 months ago