google-deepmind / dksLinks
Multi-framework implementation of Deep Kernel Shaping and Tailored Activation Transformations, which are methods that modify neural network models (and their initializations) to make them easier to train.
☆74Updated 5 months ago
Alternatives and similar repositories for dks
Users that are interested in dks are comparing it to the libraries listed below
Sorting:
- Meta-learning inductive biases in the form of useful conserved quantities.☆38Updated 3 years ago
- Pytorch implementation of preconditioned stochastic gradient descent (Kron and affine preconditioner, low-rank approximation precondition…☆188Updated this week
- ☆62Updated last year
- Official repository for the paper "Can You Learn an Algorithm? Generalizing from Easy to Hard Problems with Recurrent Networks"☆60Updated 3 years ago
- ☆60Updated 3 years ago
- Automatically take good care of your preemptible TPUs☆37Updated 2 years ago
- ☆31Updated last month
- Open source code for EigenGame.☆34Updated 2 years ago
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆46Updated 2 years ago
- ☆118Updated 2 weeks ago
- JAX implementation of Learning to learn by gradient descent by gradient descent☆28Updated 4 months ago
- Experiment of using Tangent to autodiff triton☆81Updated last year
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆46Updated last year
- Portfolio REgret for Confidence SEquences☆20Updated last year
- Code accompanying our paper "Feature Learning in Infinite-Width Neural Networks" (https://arxiv.org/abs/2011.14522)☆63Updated 4 years ago
- Jax like function transformation engine but micro, microjax☆34Updated last year
- ☆29Updated last year
- Latent Diffusion Language Models☆70Updated 2 years ago
- ☆33Updated last year
- A case study of efficient training of large language models using commodity hardware.☆68Updated 3 years ago
- Implementation of GateLoop Transformer in Pytorch and Jax☆91Updated last year
- This repository includes code to reproduce the tables in "Loss Landscapes are All You Need: Neural Network Generalization Can Be Explaine…☆40Updated 2 years ago
- A collection of optimizers, some arcane others well known, for Flax.☆29Updated 4 years ago
- Implementation of Gradient Agreement Filtering, from Chaubard et al. of Stanford, but for single machine microbatches, in Pytorch☆25Updated 11 months ago
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆92Updated last year
- ☆34Updated last year
- ☆192Updated 6 months ago
- DiCE: The Infinitely Differentiable Monte-Carlo Estimator☆32Updated 2 years ago
- Deep Networks Grok All the Time and Here is Why☆38Updated last year
- Pytorch implementation of a simple way to enable (Stochastic) Frame Averaging for any network☆51Updated last year