facebookresearch / optimizersLinks
For optimization algorithm research and development.
☆521Updated this week
Alternatives and similar repositories for optimizers
Users that are interested in optimizers are comparing it to the libraries listed below
Sorting:
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆386Updated last week
- Annotated version of the Mamba paper☆486Updated last year
- TensorDict is a pytorch dedicated tensor container.☆937Updated this week
- Scalable and Performant Data Loading☆287Updated this week
- Implementation of Diffusion Transformer (DiT) in JAX☆279Updated last year
- ☆303Updated last year
- Efficient optimizers☆232Updated last week
- Universal Tensor Operations in Einstein-Inspired Notation for Python.☆385Updated 3 months ago
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆561Updated 3 weeks ago
- ☆273Updated 11 months ago
- CIFAR-10 speedruns: 94% in 2.6 seconds and 96% in 27 seconds☆256Updated 4 months ago
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆607Updated this week
- Best practices & guides on how to write distributed pytorch training code☆444Updated 4 months ago
- Named tensors with first-class dimensions for PyTorch☆332Updated 2 years ago
- Helpful tools and examples for working with flex-attention☆865Updated 2 weeks ago
- ☆440Updated 8 months ago
- Puzzles for exploring transformers☆354Updated 2 years ago
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.☆290Updated 10 months ago
- PyTorch Single Controller☆296Updated this week
- Transform datasets at scale. Optimize datasets for fast AI model training.☆505Updated this week
- Home for "How To Scale Your Model", a short blog-style textbook about scaling LLMs on TPUs☆424Updated this week
- ☆160Updated last year
- Implementation of Flash Attention in Jax☆213Updated last year
- The AdEMAMix Optimizer: Better, Faster, Older.☆183Updated 10 months ago
- Official Implementation of "ADOPT: Modified Adam Can Converge with Any β2 with the Optimal Rate"☆426Updated 7 months ago
- ☆195Updated 7 months ago
- 🧱 Modula software package☆202Updated 3 months ago
- Effortless plugin and play Optimizer to cut model training costs by 50%. New optimizer that is 2x faster than Adam on LLMs.☆380Updated last year
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆359Updated last week
- An implementation of PSGD Kron second-order optimizer for PyTorch☆92Updated 3 months ago