jbloomAus / SAELens
Training Sparse Autoencoders on Language Models
☆594Updated this week
Alternatives and similar repositories for SAELens:
Users that are interested in SAELens are comparing it to the libraries listed below
- Sparse autoencoders☆414Updated last week
- Sparse Autoencoder for Mechanistic Interpretability☆214Updated 6 months ago
- The nnsight package enables interpreting and manipulating the internals of deep learned models.☆469Updated this week
- Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).☆177Updated last month
- Mechanistic Interpretability Visualizations using React☆223Updated last month
- ☆413Updated 6 months ago
- ☆220Updated last week
- Using sparse coding to find distributed representations used by neural networks.☆210Updated last year
- ☆202Updated 3 months ago
- ☆139Updated this week
- ☆116Updated last year
- This repository collects all relevant resources about interpretability in LLMs☆309Updated 2 months ago
- ☆422Updated this week
- Tools for understanding how transformer predictions are built layer-by-layer☆461Updated 7 months ago
- Resources for skilling up in AI alignment research engineering. Covers basics of deep learning, mechanistic interpretability, and RL.☆208Updated 11 months ago
- A library for mechanistic interpretability of GPT-style language models☆1,783Updated this week
- ☆135Updated this week
- Representation Engineering: A Top-Down Approach to AI Transparency☆778Updated 5 months ago
- Stanford NLP Python Library for Understanding and Improving PyTorch Models via Interventions☆688Updated this week
- ☆109Updated 5 months ago
- Steering vectors for transformer language models in Pytorch / Huggingface☆81Updated 2 months ago
- Steering Llama 2 with Contrastive Activation Addition☆119Updated 8 months ago
- Erasing concepts from neural representations with provable guarantees☆221Updated this week
- Extract full next-token probabilities via language model APIs☆229Updated 11 months ago
- Code and results accompanying the paper "Refusal in Language Models Is Mediated by a Single Direction".☆163Updated 3 months ago
- For OpenMOSS Mechanistic Interpretability Team's Sparse Autoencoder (SAE) research.☆87Updated this week
- ☆45Updated this week
- ☆187Updated 11 months ago
- Improving Alignment and Robustness with Circuit Breakers☆176Updated 4 months ago
- ☆54Updated 2 months ago