wattenberg / superposition
Code associated to papers on superposition (in ML interpretability)
☆26Updated 2 years ago
Alternatives and similar repositories for superposition:
Users that are interested in superposition are comparing it to the libraries listed below
- Proof-of-concept of global switching between numpy/jax/pytorch in a library.☆18Updated 7 months ago
- Code Release for "Broken Neural Scaling Laws" (BNSL) paper☆58Updated last year
- Understanding how features learned by neural networks evolve throughout training☆32Updated 3 months ago
- ☆26Updated last year
- Meta-learning inductive biases in the form of useful conserved quantities.☆37Updated 2 years ago
- The Energy Transformer block, in JAX☆55Updated last year
- Engineering the state of RNN language models (Mamba, RWKV, etc.)☆32Updated 8 months ago
- Official repository for the paper "Can You Learn an Algorithm? Generalizing from Easy to Hard Problems with Recurrent Networks"☆60Updated 2 years ago
- ☆53Updated last year
- Evaluation of neuro-symbolic engines☆34Updated 6 months ago
- Minimal but scalable implementation of large language models in JAX☆31Updated 3 months ago
- Measuring the situational awareness of language models☆34Updated last year
- ☆24Updated this week
- Code for NeurIPS 2024 Spotlight: "Scaling Laws and Compute-Optimal Training Beyond Fixed Training Durations"☆70Updated 3 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆94Updated 2 months ago
- ☆49Updated last year
- Code accompanying our paper "Feature Learning in Infinite-Width Neural Networks" (https://arxiv.org/abs/2011.14522)☆59Updated 3 years ago
- ☆15Updated 5 months ago
- Jax/Flax rewrite of Karpathy's nanoGPT☆55Updated 2 years ago
- Sparse and discrete interpretability tool for neural networks☆58Updated last year
- A programming language for formal/informal computation.☆41Updated this week
- A case study of efficient training of large language models using commodity hardware.☆68Updated 2 years ago
- Transformer with Mu-Parameterization, implemented in Jax/Flax. Supports FSDP on TPU pods.☆30Updated 2 months ago
- Open source replication of Anthropic's Crosscoders for Model Diffing☆36Updated 3 months ago
- gzip Predicts Data-dependent Scaling Laws☆34Updated 8 months ago
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆82Updated last year
- An attempt to merge ESBN with Transformers, to endow Transformers with the ability to emergently bind symbols☆15Updated 3 years ago
- ☆29Updated 4 months ago
- A mechanistic approach for understanding and detecting factual errors of large language models.☆40Updated 7 months ago
- A MAD laboratory to improve AI architecture designs 🧪☆102Updated last month