TransformerLensOrg / TransformerLens
A library for mechanistic interpretability of GPT-style language models
☆1,599Updated this week
Related projects ⓘ
Alternatives and complementary repositories for TransformerLens
- Training Sparse Autoencoders on Language Models☆481Updated this week
- The nnsight package enables interpreting and manipulating the internals of deep learned models.☆406Updated this week
- Stanford NLP Python Library for Understanding and Improving PyTorch Models via Interventions☆649Updated 2 weeks ago
- ☆354Updated this week
- Representation Engineering: A Top-Down Approach to AI Transparency☆730Updated 3 months ago
- Sparse autoencoders☆345Updated this week
- Resources for skilling up in AI alignment research engineering. Covers basics of deep learning, mechanistic interpretability, and RL.☆200Updated 9 months ago
- A bibliography and survey of the papers surrounding o1☆795Updated last week
- Tools for understanding how transformer predictions are built layer-by-layer☆432Updated 5 months ago
- ☆331Updated 4 months ago
- Mechanistic Interpretability Visualizations using React☆200Updated 4 months ago
- Deep learning for dummies. All the practical details and useful utilities that go into working with real models.☆717Updated last month
- This repository collects all relevant resources about interpretability in LLMs☆289Updated 3 weeks ago
- What would you do with 1000 H100s...☆910Updated 10 months ago
- Minimalistic large language model 3D-parallelism training☆1,265Updated this week
- ReFT: Representation Finetuning for Language Models☆1,164Updated 2 weeks ago
- utilities for decoding deep representations (like sentence embeddings) back to text☆737Updated 2 months ago
- A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).☆745Updated this week
- Cramming the training of a (BERT-type) language model into limited compute.☆1,296Updated 5 months ago
- The hub for EleutherAI's work on interpretability and learning dynamics☆2,285Updated 3 weeks ago
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends☆813Updated this week
- Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).☆158Updated last month
- Sparse Autoencoder for Mechanistic Interpretability☆189Updated 4 months ago
- Locating and editing factual associations in GPT (NeurIPS 2022)☆576Updated 7 months ago
- ☆188Updated last month
- Using sparse coding to find distributed representations used by neural networks.☆188Updated last year
- A benchmark to evaluate language models on questions I've previously asked them to solve.☆916Updated 2 weeks ago
- ☆1,385Updated this week
- Inference-Time Intervention: Eliciting Truthful Answers from a Language Model☆471Updated last month