MadryLab / trakLinks
A fast, effective data attribution method for neural networks in PyTorch
☆227Updated last year
Alternatives and similar repositories for trak
Users that are interested in trak are comparing it to the libraries listed below
Sorting:
- Influence Functions with (Eigenvalue-corrected) Kronecker-Factored Approximate Curvature☆177Updated 7 months ago
- A simple PyTorch implementation of influence functions.☆92Updated last year
- `dattri` is a PyTorch library for developing, benchmarking, and deploying efficient data attribution algorithms.☆104Updated last week
- Influence Analysis and Estimation - Survey, Papers, and Taxonomy☆84Updated last year
- AI Logging for Interpretability and Explainability🔬☆138Updated last year
- ☆80Updated 3 years ago
- DataInf: Efficiently Estimating Data Influence in LoRA-tuned LLMs and Diffusion Models (ICLR 2024)☆79Updated last year
- Data for "Datamodels: Predicting Predictions with Training Data"☆97Updated 2 years ago
- ☆241Updated last year
- ☆103Updated last year
- ☆63Updated 4 years ago
- Source code of "Task arithmetic in the tangent space: Improved editing of pre-trained models".☆108Updated 2 years ago
- ☆103Updated 2 years ago
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆85Updated 10 months ago
- ☆115Updated 11 months ago
- NeuroSurgeon is a package that enables researchers to uncover and manipulate subnetworks within models in Huggingface Transformers☆42Updated 11 months ago
- ☆32Updated 11 months ago
- Experiments and code to generate the GINC small-scale in-context learning dataset from "An Explanation for In-context Learning as Implici…☆105Updated 2 years ago
- Conformal Language Modeling☆32Updated 2 years ago
- ☆60Updated 2 years ago
- ☆32Updated last year
- ☆32Updated 2 years ago
- ☆245Updated last year
- ☆24Updated last year
- This is an official repository for "LAVA: Data Valuation without Pre-Specified Learning Algorithms" (ICLR2023).☆52Updated last year
- Röttger et al. (NAACL 2024): "XSTest: A Test Suite for Identifying Exaggerated Safety Behaviours in Large Language Models"☆126Updated 11 months ago
- ☆51Updated 2 years ago
- Stanford NLP Python library for benchmarking the utility of LLM interpretability methods☆163Updated 7 months ago
- ☆51Updated 2 years ago
- A library for efficient patching and automatic circuit discovery.☆88Updated last month