edwardjhu / TP4
Code accompanying our paper "Feature Learning in Infinite-Width Neural Networks" (https://arxiv.org/abs/2011.14522)
☆58Updated 3 years ago
Related projects ⓘ
Alternatives and complementary repositories for TP4
- Code for the paper: "Tensor Programs II: Neural Tangent Kernel for Any Architecture"☆97Updated 4 years ago
- ☆97Updated 2 years ago
- Official repository for the paper "Can You Learn an Algorithm? Generalizing from Easy to Hard Problems with Recurrent Networks"☆60Updated 2 years ago
- Pytorch implementation of preconditioned stochastic gradient descent (affine group preconditioner, low-rank approximation preconditioner …☆127Updated last month
- Silly twitter torch implementations.☆46Updated 2 years ago
- The official repository for our paper "The Devil is in the Detail: Simple Tricks Improve Systematic Generalization of Transformers". We s…☆66Updated last year
- A centralized place for deep thinking code and experiments☆77Updated last year
- A library to create and manage configuration files, especially for machine learning projects.☆77Updated 2 years ago
- Implementations and checkpoints for ResNet, Wide ResNet, ResNeXt, ResNet-D, and ResNeSt in JAX (Flax).☆104Updated 2 years ago
- Code Release for "Broken Neural Scaling Laws" (BNSL) paper☆57Updated last year
- Hessian spectral density estimation in TF and Jax☆115Updated 4 years ago
- Easy-to-use AdaHessian optimizer (PyTorch)☆77Updated 4 years ago
- unofficial re-implementation of "Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets"☆63Updated 2 years ago
- ☆58Updated 2 years ago
- PyTorch implementation of "Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets"☆32Updated 2 years ago
- [NeurIPS'19] Deep Equilibrium Models Jax Implementation☆37Updated 4 years ago
- DeepOBS: A Deep Learning Optimizer Benchmark Suite☆103Updated 11 months ago
- Meta-learning inductive biases in the form of useful conserved quantities.☆37Updated 2 years ago
- Fast Discounted Cumulative Sums in PyTorch☆95Updated 3 years ago
- Fast training of unitary deep network layers from low-rank updates☆28Updated last year
- Official code repository of the paper Linear Transformers Are Secretly Fast Weight Programmers.☆100Updated 3 years ago
- This repository contains the Julia code for the paper "Competitive Gradient Descent"☆23Updated 4 years ago
- Structured matrices for compressing neural networks☆67Updated last year
- Source code of "What can linearized neural networks actually say about generalization?☆18Updated 3 years ago
- ☆105Updated 2 weeks ago
- ☆59Updated 3 years ago
- JAX Synergistic Memory Inspector☆164Updated 4 months ago
- This repository includes code to reproduce the tables in "Loss Landscapes are All You Need: Neural Network Generalization Can Be Explaine…☆34Updated last year
- JMP is a Mixed Precision library for JAX.☆187Updated 6 months ago