PaulPauls / llama3_interpretability_sae

A complete end-to-end pipeline for LLM interpretability with sparse autoencoders (SAEs) using Llama 3.2, written in pure PyTorch and fully reproducible.
461Updated this week

Related projects

Alternatives and complementary repositories for llama3_interpretability_sae