google-deepmind / dks
Multi-framework implementation of Deep Kernel Shaping and Tailored Activation Transformations, which are methods that modify neural network models (and their initializations) to make them easier to train.
☆68Updated last week
Alternatives and similar repositories for dks:
Users that are interested in dks are comparing it to the libraries listed below
- Pytorch implementation of preconditioned stochastic gradient descent (Kron and affine preconditioner, low-rank approximation precondition…☆171Updated this week
- ☆52Updated 6 months ago
- ☆60Updated 3 years ago
- Meta-learning inductive biases in the form of useful conserved quantities.☆37Updated 2 years ago
- Official repository for the paper "Can You Learn an Algorithm? Generalizing from Easy to Hard Problems with Recurrent Networks"☆59Updated 3 years ago
- Implementations and checkpoints for ResNet, Wide ResNet, ResNeXt, ResNet-D, and ResNeSt in JAX (Flax).☆108Updated 2 years ago
- A selection of neural network models ported from torchvision for JAX & Flax.☆44Updated 4 years ago
- Transformers with doubly stochastic attention☆45Updated 2 years ago
- ☆87Updated 3 weeks ago
- Codes accompanying the paper "LaProp: a Better Way to Combine Momentum with Adaptive Gradient"☆28Updated 4 years ago
- Open source code for EigenGame.☆30Updated last year
- JAX implementation of Learning to learn by gradient descent by gradient descent☆27Updated 5 months ago
- A functional training loops library for JAX☆86Updated last year
- Experiment of using Tangent to autodiff triton☆78Updated last year
- FID computation in Jax/Flax.☆27Updated 8 months ago
- Implementation of PSGD optimizer in JAX☆30Updated 3 months ago
- ☆30Updated 5 months ago
- The Energy Transformer block, in JAX☆56Updated last year
- Fast training of unitary deep network layers from low-rank updates☆29Updated 2 years ago
- Tensor Parallelism with JAX + Shard Map☆11Updated last year
- A collection of optimizers, some arcane others well known, for Flax.☆29Updated 3 years ago
- Neural Networks for JAX☆83Updated 6 months ago
- [ICML 2024] SIRFShampoo: Structured inverse- and root-free Shampoo in PyTorch (https://arxiv.org/abs/2402.03496)☆14Updated 4 months ago
- ☆101Updated 9 months ago
- JMP is a Mixed Precision library for JAX.☆193Updated 2 months ago
- A simple library for scaling up JAX programs☆134Updated 5 months ago
- Usable implementation of Emerging Symbol Binding Network (ESBN), in Pytorch☆24Updated 4 years ago
- ☆27Updated last year
- Automatically take good care of your preemptible TPUs☆36Updated last year
- DiCE: The Infinitely Differentiable Monte-Carlo Estimator☆31Updated last year