scienceetonnante / grokking
Demonstration of the grokking phenomenon in machine learning in a simple case
☆57Updated last month
Alternatives and similar repositories for grokking:
Users that are interested in grokking are comparing it to the libraries listed below
- 🧠Starter templates for doing interpretability research☆67Updated last year
- PyTorch implementation of "Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets"☆36Updated 3 years ago
- unofficial re-implementation of "Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets"☆77Updated 2 years ago
- Delphi was the home of a temple to Phoebus Apollo, which famously had the inscription, 'Know Thyself.' This library lets language models …☆165Updated this week
- Tools for studying developmental interpretability in neural networks.☆87Updated 2 months ago
- Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).☆192Updated 3 months ago
- SMIT: A Simple Modality Integration Tool☆16Updated last year
- Mechanistic Interpretability Visualizations using React☆238Updated 3 months ago
- Omnigrok: Grokking Beyond Algorithmic Data☆54Updated 2 years ago
- Tools for understanding how transformer predictions are built layer-by-layer☆481Updated 10 months ago
- Sparse and discrete interpretability tool for neural networks☆60Updated last year
- ☆267Updated last month
- Editing Models with Task Arithmetic☆460Updated last year
- Sparse Autoencoder for Mechanistic Interpretability☆234Updated 8 months ago
- nanoGPT-like codebase for LLM training☆91Updated this week
- Resources for skilling up in AI alignment research engineering. Covers basics of deep learning, mechanistic interpretability, and RL.☆207Updated last year
- ☆214Updated 6 months ago
- Access to free kaggle compute power from your command line☆28Updated last year
- Notebooks accompanying Anthropic's "Toy Models of Superposition" paper☆118Updated 2 years ago
- Redwood Research's transformer interpretability tools☆14Updated 2 years ago
- The boundary of neural network trainability is fractal☆197Updated last year
- 🔬 Interpretability for Leela Chess Zero networks.☆12Updated this week
- Parameter-Free Optimizers for Pytorch☆122Updated 11 months ago
- ☆61Updated 2 years ago
- ☆55Updated 4 months ago
- Neural Networks and the Chomsky Hierarchy☆205Updated 11 months ago
- Backend ressources for Albert. Albert is a conversational agent that uses official French data sources to answer administrative agents qu…☆120Updated 2 months ago
- Official JAX implementation of xLSTM including fast and efficient training and inference code. 7B model available at https://huggingface.…☆91Updated 2 months ago
- Benchmarks for the Evaluation of LLM Supervision☆31Updated this week
- Using sparse coding to find distributed representations used by neural networks.☆226Updated last year