Official Implementation of "ADOPT: Modified Adam Can Converge with Any β2 with the Optimal Rate"
☆435Dec 12, 2024Updated last year
Alternatives and similar repositories for adopt
Users that are interested in adopt are comparing it to the libraries listed below
Sorting:
- Schedule-Free Optimization in PyTorch☆2,257May 21, 2025Updated 9 months ago
- ☆252Dec 2, 2024Updated last year
- For optimization algorithm research and development.☆557Updated this week
- ☆15Mar 2, 2025Updated last year
- Grams: Gradient Descent with Adaptive Momentum Scaling (ICLR 2025 Workshop)☆17Mar 6, 2025Updated last year
- Official Code for MIMETIC^2☆13Nov 19, 2024Updated last year
- ☆307Apr 23, 2025Updated 10 months ago
- Efficient optimizers☆285Dec 20, 2025Updated 2 months ago
- Getting crystal-like representations with harmonic loss☆194Apr 2, 2025Updated 11 months ago
- ☆39Oct 31, 2025Updated 4 months ago
- The AdEMAMix Optimizer: Better, Faster, Older.☆186Sep 12, 2024Updated last year
- When it comes to optimizers, it's always better to be safe than sorry☆407Sep 26, 2025Updated 5 months ago
- MiSS is a novel PEFT method that features a low-rank structure but introduces a new update mechanism distinct from LoRA, achieving an exc…☆31Jan 28, 2026Updated last month
- Train VAE like a boss☆313Oct 21, 2024Updated last year
- Erwin: A Tree-based Hierarchical Transformer for Large-scale Physical Systems [ICML'25]☆111Oct 11, 2025Updated 4 months ago
- Solution of Kaggle competition: MAP - Charting Student Math Misunderstandings☆24Oct 25, 2025Updated 4 months ago
- Code for NeurIPS 2024 Spotlight: "Scaling Laws and Compute-Optimal Training Beyond Fixed Training Durations"☆92Oct 30, 2024Updated last year
- Codes accompanying the paper "LaProp: a Better Way to Combine Momentum with Adaptive Gradient"☆29Jul 30, 2020Updated 5 years ago
- Official repository of "LiNeS: Post-training Layer Scaling Prevents Forgetting and Enhances Model Merging"☆32Nov 4, 2024Updated last year
- DeMo: Decoupled Momentum Optimization☆198Dec 2, 2024Updated last year
- Muon is an optimizer for hidden layers in neural networks☆2,350Jan 19, 2026Updated last month
- ☆70Nov 15, 2024Updated last year
- Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793☆453May 13, 2025Updated 9 months ago
- albumentations test☆11Jun 23, 2020Updated 5 years ago
- [NeurIPS 2024] Goldfish Loss: Mitigating Memorization in Generative LLMs☆94Nov 17, 2024Updated last year
- ☆40Jan 6, 2025Updated last year
- lightweight, fast and robust columnar dataframe for data analytics with online update☆23Aug 14, 2021Updated 4 years ago
- Helpful tools and examples for working with flex-attention☆1,140Feb 8, 2026Updated 3 weeks ago
- ☆22Nov 9, 2024Updated last year
- Bringing BERT into modernity via both architecture changes and scaling☆1,632Updated this week
- PyTorch Implementation of TecNets (Task-Embedded Control Networks)☆10Dec 8, 2022Updated 3 years ago
- 2nd Place Solution for the Google Research - Identify Contrails to Reduce Global Warming Competition☆14Aug 15, 2023Updated 2 years ago
- ☆10May 24, 2021Updated 4 years ago
- Benchmark Benchmark Benchmark!☆51May 31, 2023Updated 2 years ago
- Adan: Adaptive Nesterov Momentum Algorithm for Faster Optimizing Deep Models