PaulPauls / llama3_interpretability_saeView on GitHub
A complete end-to-end pipeline for LLM interpretability with sparse autoencoders (SAEs) using Llama 3.2, written in pure PyTorch and fully reproducible.
633Mar 23, 2025Updated last year

Alternatives and similar repositories for llama3_interpretability_sae

Users that are interested in llama3_interpretability_sae are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.

Sorting:

Are these results useful?