formll / dogLinks
DoG is SGD's Best Friend: A Parameter-Free Dynamic Step Size Schedule
☆63Updated last year
Alternatives and similar repositories for dog
Users that are interested in dog are comparing it to the libraries listed below
Sorting:
- Replicating and dissecting the git-re-basin project in one-click-replication Colabs☆36Updated 2 years ago
- Parameter-Free Optimizers for Pytorch☆130Updated last year
- ☆53Updated 9 months ago
- ☆26Updated 2 weeks ago
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆84Updated last year
- ☆32Updated 9 months ago
- A State-Space Model with Rational Transfer Function Representation.☆79Updated last year
- unofficial re-implementation of "Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets"☆77Updated 3 years ago
- Sequence Modeling with Multiresolution Convolutional Memory (ICML 2023)☆124Updated last year
- Implementation of GateLoop Transformer in Pytorch and Jax☆89Updated last year
- Pytorch implementation of preconditioned stochastic gradient descent (Kron and affine preconditioner, low-rank approximation precondition…☆179Updated last month
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆66Updated 9 months ago
- Omnigrok: Grokking Beyond Algorithmic Data☆58Updated 2 years ago
- ☆197Updated 7 months ago
- Code accompanying our paper "Feature Learning in Infinite-Width Neural Networks" (https://arxiv.org/abs/2011.14522)☆62Updated 4 years ago
- ☆70Updated 7 months ago
- Codes for the paper The emergence of clusters in self-attention dynamics.☆16Updated last year
- nanoGPT-like codebase for LLM training☆100Updated 2 months ago
- LoRA for arbitrary JAX models and functions☆140Updated last year
- Deep Networks Grok All the Time and Here is Why☆37Updated last year
- Transformers with doubly stochastic attention☆46Updated 2 years ago
- ☆26Updated 2 years ago
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆80Updated last year
- Lightning-like training API for JAX with Flax☆42Updated 7 months ago
- 🧱 Modula software package☆204Updated 3 months ago
- ☆32Updated last year
- Code for the paper "Function-Space Learning Rates"☆20Updated last month
- Automatically take good care of your preemptible TPUs☆36Updated 2 years ago
- One Initialization to Rule them All: Fine-tuning via Explained Variance Adaptation☆40Updated 9 months ago
- [ICML 2024] SINGD: KFAC-like Structured Inverse-Free Natural Gradient Descent (http://arxiv.org/abs/2312.05705)☆22Updated 8 months ago