goodfire-ai / goodfire-sdkLinks
Ember is a hosted API/SDK that lets you shape AI model behavior by directly controlling a model's internal units of computation, or "features". With Ember, you can modify features to precisely control model outputs, or use them as building blocks for tasks like classification.
☆29Updated 3 weeks ago
Alternatives and similar repositories for goodfire-sdk
Users that are interested in goodfire-sdk are comparing it to the libraries listed below
Sorting:
- METR Task Standard☆157Updated 6 months ago
- Mechanistic Interpretability Visualizations using React☆273Updated 7 months ago
- Inference API for many LLMs and other useful tools for empirical research☆63Updated this week
- ControlArena is a collection of settings, model organisms and protocols - for running control experiments.☆82Updated this week
- Resources for skilling up in AI alignment research engineering. Covers basics of deep learning, mechanistic interpretability, and RL.☆221Updated last year
- (Model-written) LLM evals library☆18Updated 7 months ago
- ☆234Updated 10 months ago
- Vivaria is METR's tool for running evaluations and conducting agent elicitation research.☆103Updated this week
- A tiny easily hackable implementation of a feature dashboard.☆12Updated last month
- Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).☆210Updated 7 months ago
- ☆19Updated last month
- Unified access to Large Language Model modules using NNsight☆38Updated 2 weeks ago
- ☆99Updated 4 months ago
- ☆16Updated this week
- ☆51Updated 8 months ago
- ☆125Updated last year
- Tools for studying developmental interpretability in neural networks.☆100Updated last month
- Machine Learning for Alignment Bootcamp☆76Updated 3 years ago
- Open source interpretability artefacts for R1.☆157Updated 3 months ago
- Decoder only transformer, built from scratch with PyTorch☆31Updated last year
- Sparse Autoencoder for Mechanistic Interpretability☆258Updated last year
- Notebooks accompanying Anthropic's "Toy Models of Superposition" paper☆128Updated 2 years ago
- ☆72Updated 2 months ago
- The nnsight package enables interpreting and manipulating the internals of deep learned models.☆622Updated last week
- ☆289Updated last year
- Steering vectors for transformer language models in Pytorch / Huggingface☆120Updated 5 months ago
- Collection of evals for Inspect AI☆201Updated this week
- ☆182Updated 5 months ago
- A toolkit for describing model features and intervening on those features to steer behavior.☆195Updated 9 months ago
- Contains random samples referenced in the paper "Sleeper Agents: Training Robustly Deceptive LLMs that Persist Through Safety Training".☆113Updated last year