EleutherAI / sparsify
Sparsify transformers with SAEs and transcoders
☆459Updated this week
Alternatives and similar repositories for sparsify:
Users that are interested in sparsify are comparing it to the libraries listed below
- Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).☆182Updated 2 months ago
- Training Sparse Autoencoders on Language Models☆616Updated this week
- ☆239Updated last week
- Sparse Autoencoder for Mechanistic Interpretability☆216Updated 6 months ago
- Using sparse coding to find distributed representations used by neural networks.☆213Updated last year
- ☆150Updated this week
- Mechanistic Interpretability Visualizations using React☆231Updated 2 months ago
- ☆421Updated 7 months ago
- The nnsight package enables interpreting and manipulating the internals of deep learned models.☆489Updated this week
- ☆116Updated last year
- ☆203Updated 4 months ago
- ☆141Updated 3 weeks ago
- This repository collects all relevant resources about interpretability in LLMs☆320Updated 3 months ago
- Extract full next-token probabilities via language model APIs☆228Updated 11 months ago
- Code and results accompanying the paper "Refusal in Language Models Is Mediated by a Single Direction".☆180Updated 4 months ago
- Tools for understanding how transformer predictions are built layer-by-layer☆473Updated 8 months ago
- ☆109Updated 6 months ago
- ☆54Updated 3 months ago
- Steering vectors for transformer language models in Pytorch / Huggingface☆87Updated 2 months ago
- ViT Prisma is a mechanistic interpretability library for Vision Transformers (ViTs).☆199Updated this week
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆184Updated 8 months ago
- ☆121Updated last week
- Steering Llama 2 with Contrastive Activation Addition☆122Updated 8 months ago
- For OpenMOSS Mechanistic Interpretability Team's Sparse Autoencoder (SAE) research.☆92Updated this week
- Resources for skilling up in AI alignment research engineering. Covers basics of deep learning, mechanistic interpretability, and RL.☆208Updated last year
- A toolkit for describing model features and intervening on those features to steer behavior.☆159Updated 3 months ago
- Improving Alignment and Robustness with Circuit Breakers☆184Updated 4 months ago
- Stanford NLP Python library for understanding and improving PyTorch models via interventions☆696Updated this week
- Erasing concepts from neural representations with provable guarantees☆222Updated 3 weeks ago
- ☆190Updated 11 months ago