RylanSchaeffer / Stanford-AI-Alignment-Double-Descent-TutorialLinks
Code for Arxiv Double Descent Demystified: Identifying, Interpreting & Ablating the Sources of a Deep Learning Puzzle
☆30Updated last year
Alternatives and similar repositories for Stanford-AI-Alignment-Double-Descent-Tutorial
Users that are interested in Stanford-AI-Alignment-Double-Descent-Tutorial are comparing it to the libraries listed below
Sorting:
- This repository includes code to reproduce the tables in "Loss Landscapes are All You Need: Neural Network Generalization Can Be Explaine…☆39Updated 2 years ago
- ☆27Updated 2 years ago
- Deep Networks Grok All the Time and Here is Why☆37Updated last year
- ModelDiff: A Framework for Comparing Learning Algorithms☆59Updated 2 years ago
- Portfolio REgret for Confidence SEquences☆20Updated 9 months ago
- Sparse and discrete interpretability tool for neural networks☆63Updated last year
- ☆32Updated last year
- Unofficial implementation of Conformal Language Modeling by Quach et al☆29Updated 2 years ago
- ☆37Updated 3 years ago
- Understanding how features learned by neural networks evolve throughout training☆39Updated 11 months ago
- 👋 Overcomplete is a Vision-based SAE Toolbox☆90Updated 2 months ago
- Omnigrok: Grokking Beyond Algorithmic Data☆62Updated 2 years ago
- Google Research☆46Updated 2 years ago
- Official code for the paper "Compositional Generalization from First Principles" (NeurIPS 2023)☆12Updated 2 years ago
- ☆58Updated last year
- Code Release for "Broken Neural Scaling Laws" (BNSL) paper☆59Updated last year
- A centralized place for deep thinking code and experiments☆87Updated 2 years ago
- Repository for the PopulAtion Parameter Averaging (PAPA) paper☆27Updated last year
- Recycling diverse models☆45Updated 2 years ago
- Code for minimum-entropy coupling.☆32Updated last year
- Automatic Integration for Neural Spatio-Temporal Point Process models (AI-STPP) is a new paradigm for exact, efficient, non-parametric inf…☆25Updated 11 months ago
- Code for "Counterfactual Token Generation in Large Language Models", Arxiv 2024.☆29Updated 11 months ago
- Official repo of Progressive Data Expansion: data, code and evaluation☆29Updated last year
- Official Code Repository for the paper "Key-value memory in the brain"☆28Updated 7 months ago
- Fast Axiomatic Attribution for Neural Networks (NeurIPS*2021)☆16Updated 2 years ago
- Implementation of the general framework for AMIE, from the paper "Towards Conversational Diagnostic AI", out of Google Deepmind☆68Updated last year
- Repository for the paper Do SSL Models Have Déjà Vu? A Case of Unintended Memorization in Self-supervised Learning☆36Updated 2 years ago
- Quantification of Uncertainty with Adversarial Models☆29Updated 2 years ago
- Minimum Description Length probing for neural network representations☆20Updated 8 months ago
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆67Updated last year