fjzzq2002 / pizzaLinks
Code repository for "The Clock and the Pizza: Two Stories in Mechanistic Explanation of Neural Networks"
☆18Updated 2 years ago
Alternatives and similar repositories for pizza
Users that are interested in pizza are comparing it to the libraries listed below
Sorting:
- ☆28Updated 2 years ago
- Official implementation of the transformer (TF) architecture suggested in a paper entitled "Looped Transformers as Programmable Computers…☆30Updated 2 years ago
- ☆46Updated 2 years ago
- Omnigrok: Grokking Beyond Algorithmic Data☆62Updated 2 years ago
- ☆78Updated 3 years ago
- ☆31Updated 10 months ago
- ZeroC is a neuro-symbolic method that trained with elementary visual concepts and relations, can zero-shot recognize and acquire more com…☆33Updated 2 years ago
- This is the official repository for the "Towards Vision-Language Mechanistic Interpretability: A Causal Tracing Tool for BLIP" paper acce…☆25Updated last year
- ☆116Updated last year
- This repository includes code to reproduce the tables in "Loss Landscapes are All You Need: Neural Network Generalization Can Be Explaine…☆40Updated 2 years ago
- Official repository for our paper, Transformers Learn Higher-Order Optimization Methods for In-Context Learning: A Study with Linear Mode…☆19Updated last year
- ☆13Updated last year
- Deep Learning & Information Bottleneck☆63Updated 2 years ago
- ☆34Updated 2 years ago
- [NeurIPS 2023 Spotlight] Temperature Balancing, Layer-wise Weight Analysis, and Neural Network Training☆36Updated 10 months ago
- ModelDiff: A Framework for Comparing Learning Algorithms☆58Updated 2 years ago
- Code for "Can We Scale Transformers to Predict Parameters of Diverse ImageNet Models?" [ICML 2023]☆38Updated last year
- The Energy Transformer block, in JAX☆63Updated 2 years ago
- Sparse and discrete interpretability tool for neural networks☆64Updated 2 years ago
- Unofficial Implementation of Selective Attention Transformer☆20Updated last year
- ☆73Updated last year
- Code for GFlowNet-EM, a novel algorithm for fitting latent variable models with compositional latents and an intractable true posterior.☆42Updated 2 years ago
- Code for "Theoretical Foundations of Deep Selective State-Space Models" (NeurIPS 2024)☆15Updated last year
- ☆24Updated last year
- ☆108Updated last year
- Hrrformer: A Neuro-symbolic Self-attention Model (ICML23)☆61Updated 4 months ago
- ☆16Updated last year
- Interpretating the latent space representations of attention head outputs for LLMs☆36Updated last year
- Universal Neurons in GPT2 Language Models☆30Updated last year
- Code for NeurIPS 2024 Spotlight: "Scaling Laws and Compute-Optimal Training Beyond Fixed Training Durations"☆89Updated last year