redwoodresearch / remix_public
☆19Updated 2 years ago
Alternatives and similar repositories for remix_public:
Users that are interested in remix_public are comparing it to the libraries listed below
- Tools for studying developmental interpretability in neural networks.☆86Updated last month
- ☆26Updated 11 months ago
- ☆29Updated 10 months ago
- Keeping language models honest by directly eliciting knowledge encoded in their activations.☆197Updated this week
- Vivaria is METR's tool for running evaluations and conducting agent elicitation research.☆83Updated this week
- (Model-written) LLM evals library☆18Updated 3 months ago
- ☆15Updated 3 weeks ago
- Mechanistic Interpretability for Transformer Models☆50Updated 2 years ago
- Redwood Research's transformer interpretability tools☆14Updated 2 years ago
- ☆61Updated 4 months ago
- Machine Learning for Alignment Bootcamp☆72Updated 2 years ago
- ☆10Updated 8 months ago
- ☆62Updated 2 years ago
- Mechanistic Interpretability Visualizations using React☆233Updated 3 months ago
- Measuring the situational awareness of language models☆34Updated last year
- ☆211Updated 5 months ago
- ☆19Updated 2 years ago
- Resources for skilling up in AI alignment research engineering. Covers basics of deep learning, mechanistic interpretability, and RL.☆206Updated last year
- Machine Learning for Alignment Bootcamp (MLAB).☆28Updated 3 years ago
- Notebooks accompanying Anthropic's "Toy Models of Superposition" paper☆116Updated 2 years ago
- A library for bridging Python and HTML/Javascript (via Svelte) for creating interactive visualizations☆14Updated 11 months ago
- ☆65Updated last month
- A library for efficient patching and automatic circuit discovery.☆59Updated last month
- Steering vectors for transformer language models in Pytorch / Huggingface☆90Updated last month
- Code for my NeurIPS 2024 ATTRIB paper titled "Attribution Patching Outperforms Automated Circuit Discovery"☆30Updated 9 months ago
- See the issue board for the current status of active and prospective projects!☆65Updated 3 years ago
- Erasing concepts from neural representations with provable guarantees☆226Updated last month
- 🧠 Starter templates for doing interpretability research☆67Updated last year
- Algebraic value editing in pretrained language models☆63Updated last year
- PyTorch and NNsight implementation of AtP* (Kramar et al 2024, DeepMind)☆18Updated 2 months ago