facebookresearch / dadaptationLinks
D-Adaptation for SGD, Adam and AdaGrad
☆523Updated 7 months ago
Alternatives and similar repositories for dadaptation
Users that are interested in dadaptation are comparing it to the libraries listed below
Sorting:
- optimizer & lr scheduler & loss function collections in PyTorch☆344Updated last week
- Code release for "Git Re-Basin: Merging Models modulo Permutation Symmetries"☆487Updated 2 years ago
- Code for our NeurIPS 2022 paper☆369Updated 2 years ago
- Implementation of a memory efficient multi-head attention as proposed in the paper, "Self-attention Does Not Need O(n²) Memory"☆379Updated 2 years ago
- minLoRA: a minimal PyTorch library that allows you to apply LoRA to any PyTorch model.☆474Updated 2 years ago
- Effortless plugin and play Optimizer to cut model training costs by 50%. New optimizer that is 2x faster than Adam on LLMs.☆379Updated last year
- ☆783Updated 2 months ago
- A simple way to keep track of an Exponential Moving Average (EMA) version of your Pytorch model☆601Updated 8 months ago
- The Prodigy optimizer and its variants for training neural networks.☆411Updated 7 months ago
- Official Implementation of "ADOPT: Modified Adam Can Converge with Any β2 with the Optimal Rate"☆428Updated 8 months ago
- Implementation of the Adan (ADAptive Nesterov momentum algorithm) Optimizer in Pytorch☆251Updated 2 years ago
- ☆307Updated last year
- 🦁 Lion, new optimizer discovered by Google Brain using genetic algorithms that is purportedly better than Adam(w), in Pytorch☆2,155Updated 9 months ago
- Memory Efficient Attention (O(sqrt(n)) for Jax and PyTorch☆184Updated 2 years ago
- For optimization algorithm research and development.☆530Updated this week
- A library to inspect and extract intermediate layers of PyTorch models.☆473Updated 3 years ago
- The official implementation of “Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training”☆967Updated last year
- Library for Jacobian descent with PyTorch. It enables the optimization of neural networks with multiple losses (e.g. multi-task learning)…☆264Updated this week
- Code release for "Dropout Reduces Underfitting"☆314Updated 2 years ago
- Named tensors with first-class dimensions for PyTorch☆331Updated 2 years ago
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆390Updated last week
- maximal update parametrization (µP)☆1,590Updated last year
- Efficient optimizers☆256Updated 3 weeks ago
- TensorDict is a pytorch dedicated tensor container.☆955Updated last week
- ☆208Updated 2 years ago
- Annotated version of the Mamba paper☆488Updated last year
- Implementing the Denoising Diffusion Probabilistic Model in Flax☆149Updated 2 years ago
- Pretrained deep learning models for Jax/Flax: StyleGAN2, GPT2, VGG, ResNet, etc.☆258Updated 5 months ago
- Unofficial Implementation of Consistency Models in pytorch☆259Updated 2 years ago
- Language Modeling with the H3 State Space Model☆519Updated last year