ai-safety-foundation / sparse_autoencoder
Sparse Autoencoder for Mechanistic Interpretability
☆233Updated 8 months ago
Alternatives and similar repositories for sparse_autoencoder:
Users that are interested in sparse_autoencoder are comparing it to the libraries listed below
- Using sparse coding to find distributed representations used by neural networks.☆224Updated last year
- ☆258Updated last month
- Mechanistic Interpretability Visualizations using React☆235Updated 3 months ago
- Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).☆190Updated 3 months ago
- ☆213Updated 5 months ago
- Sparsify transformers with SAEs and transcoders☆494Updated this week
- ☆121Updated last year
- ☆150Updated 2 weeks ago
- Delphi was the home of a temple to Phoebus Apollo, which famously had the inscription, 'Know Thyself.' This library lets language models …☆163Updated this week
- Training Sparse Autoencoders on Language Models☆686Updated this week
- The nnsight package enables interpreting and manipulating the internals of deep learned models.☆522Updated this week
- ☆78Updated last week
- ☆444Updated 8 months ago
- ☆61Updated 4 months ago
- Steering Llama 2 with Contrastive Activation Addition☆131Updated 10 months ago
- ☆131Updated 4 months ago
- ☆113Updated 7 months ago
- Steering vectors for transformer language models in Pytorch / Huggingface☆90Updated last month
- For OpenMOSS Mechanistic Interpretability Team's Sparse Autoencoder (SAE) research.☆105Updated last week
- This repository collects all relevant resources about interpretability in LLMs☆327Updated 4 months ago
- Code and results accompanying the paper "Refusal in Language Models Is Mediated by a Single Direction".☆195Updated 5 months ago
- LLM experiments done during SERI MATS - focusing on activation steering / interpreting activation spaces☆91Updated last year
- ☆90Updated last month
- ☆196Updated last year
- Tools for studying developmental interpretability in neural networks.☆86Updated 2 months ago
- ☆26Updated 11 months ago
- A toolkit for describing model features and intervening on those features to steer behavior.☆163Updated 4 months ago
- ☆32Updated 4 months ago
- Resources for skilling up in AI alignment research engineering. Covers basics of deep learning, mechanistic interpretability, and RL.☆206Updated last year
- Open source replication of Anthropic's Crosscoders for Model Diffing☆46Updated 5 months ago