Prisma-Multimodal / ViT-PrismaLinks
ViT Prisma is a mechanistic interpretability library for Vision and Video Transformers (ViTs).
☆282Updated last week
Alternatives and similar repositories for ViT-Prisma
Users that are interested in ViT-Prisma are comparing it to the libraries listed below
Sorting:
- Sparsify transformers with SAEs and transcoders☆583Updated this week
- ☆314Updated last month
- Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).☆206Updated 6 months ago
- Sparse Autoencoder for Mechanistic Interpretability☆255Updated 11 months ago
- ☆122Updated last year
- The nnsight package enables interpreting and manipulating the internals of deep learned models.☆605Updated this week
- ☆231Updated 9 months ago
- Mechanistic Interpretability Visualizations using React☆260Updated 6 months ago
- Using sparse coding to find distributed representations used by neural networks.☆259Updated last year
- ☆105Updated last month
- Delphi was the home of a temple to Phoebus Apollo, which famously had the inscription, 'Know Thyself.' This library lets language models …☆192Updated this week
- ☆500Updated 11 months ago
- ☆572Updated 3 months ago
- Training Sparse Autoencoders on Language Models☆864Updated this week
- Tools for understanding how transformer predictions are built layer-by-layer☆503Updated last year
- ☆99Updated 5 months ago
- ☆181Updated 3 months ago
- Resources for skilling up in AI alignment research engineering. Covers basics of deep learning, mechanistic interpretability, and RL.☆216Updated last year
- ☆273Updated last year
- Steering vectors for transformer language models in Pytorch / Huggingface☆112Updated 4 months ago
- This repository collects all relevant resources about interpretability in LLMs☆362Updated 8 months ago
- Sparse Autoencoder Training Library☆53Updated 2 months ago
- ☆47Updated 7 months ago
- Official implementation of MAIA, A Multimodal Automated Interpretability Agent☆82Updated 3 weeks ago
- Notebooks accompanying Anthropic's "Toy Models of Superposition" paper☆127Updated 2 years ago
- 🧠 Starter templates for doing interpretability research☆72Updated last year
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆190Updated last year
- ☆78Updated 4 months ago
- Tools for studying developmental interpretability in neural networks.☆99Updated 2 weeks ago
- Applying SAEs for fine-grained control☆22Updated 6 months ago