callummcdougall / ARENA_2.0
Resources for skilling up in AI alignment research engineering. Covers basics of deep learning, mechanistic interpretability, and RL.
☆194Updated 9 months ago
Related projects ⓘ
Alternatives and complementary repositories for ARENA_2.0
- Mechanistic Interpretability Visualizations using React☆195Updated 3 months ago
- ☆337Updated this week
- ☆186Updated last month
- The nnsight package enables interpreting and manipulating the internals of deep learned models.☆399Updated this week
- Tools for studying developmental interpretability in neural networks.☆74Updated this week
- Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).☆157Updated last month
- Training Sparse Autoencoders on Language Models☆449Updated this week
- Sparse Autoencoder for Mechanistic Interpretability☆187Updated 3 months ago
- METR Task Standard☆122Updated last week
- Machine Learning for Alignment Bootcamp☆63Updated 2 years ago
- ☆102Updated last month
- ☆141Updated 2 weeks ago
- ☆108Updated last year
- Using sparse coding to find distributed representations used by neural networks.☆181Updated 11 months ago
- ☆96Updated 3 months ago
- Keeping language models honest by directly eliciting knowledge encoded in their activations.☆186Updated this week
- This repository collects all relevant resources about interpretability in LLMs☆282Updated last week
- Sparse autoencoders☆333Updated 2 weeks ago
- ☆252Updated 8 months ago
- Steering vectors for transformer language models in Pytorch / Huggingface☆64Updated last month
- Notebooks accompanying Anthropic's "Toy Models of Superposition" paper☆95Updated 2 years ago
- Tools for understanding how transformer predictions are built layer-by-layer☆429Updated 5 months ago
- ☆43Updated 4 months ago
- ☆99Updated this week
- Extract full next-token probabilities via language model APIs☆228Updated 8 months ago
- Machine Learning for Alignment Bootcamp (MLAB).☆22Updated 2 years ago
- Steering Llama 2 with Contrastive Activation Addition☆94Updated 5 months ago
- Mechanistic Interpretability for Transformer Models☆49Updated 2 years ago
- 🧠 Starter templates for doing interpretability research☆63Updated last year
- Emergent world representations: Exploring a sequence model trained on a synthetic task☆168Updated last year