safety-research / safety-examplesLinks
☆20Updated 2 months ago
Alternatives and similar repositories for safety-examples
Users that are interested in safety-examples are comparing it to the libraries listed below
Sorting:
- Inference API for many LLMs and other useful tools for empirical research☆68Updated last week
- ☆34Updated last year
- METR Task Standard☆159Updated 7 months ago
- A TinyStories LM with SAEs and transcoders☆13Updated 5 months ago
- ControlArena is a collection of settings, model organisms and protocols - for running control experiments.☆93Updated this week
- Resources for skilling up in AI alignment research engineering. Covers basics of deep learning, mechanistic interpretability, and RL.☆226Updated last month
- Machine Learning for Alignment Bootcamp☆78Updated 3 years ago
- A tiny easily hackable implementation of a feature dashboard.☆13Updated 2 months ago
- Attribution-based Parameter Decomposition☆30Updated 3 months ago
- Open source replication of Anthropic's Crosscoders for Model Diffing☆59Updated 10 months ago
- Redwood Research's transformer interpretability tools☆14Updated 3 years ago
- Notebooks accompanying Anthropic's "Toy Models of Superposition" paper☆128Updated 3 years ago
- Sparse Autoencoder Training Library☆54Updated 4 months ago
- ☆68Updated 2 weeks ago
- Investigating the generalization behavior of LM probes trained to predict truth labels: (1) from one annotator to another, and (2) from e…☆28Updated last year
- Mechanistic Interpretability Visualizations using React☆289Updated 9 months ago
- Improving Steering Vectors by Targeting Sparse Autoencoder Features☆24Updated 10 months ago
- Tools for studying developmental interpretability in neural networks.☆103Updated 2 months ago
- ☆54Updated 10 months ago
- ☆103Updated 6 months ago
- A collection of different ways to implement accessing and modifying internal model activations for LLMs☆19Updated 11 months ago
- ☆99Updated 4 months ago
- Applying SAEs for fine-grained control☆23Updated 9 months ago
- ☆240Updated 11 months ago
- Official Code for our paper: "Language Models Learn to Mislead Humans via RLHF""☆17Updated 11 months ago
- ☆142Updated last week
- A library for efficient patching and automatic circuit discovery.☆76Updated last month
- Code repo for the model organisms and convergent directions of EM papers.☆28Updated last month
- A toolkit for describing model features and intervening on those features to steer behavior.☆201Updated 10 months ago
- ☆17Updated 5 months ago