ethansmith2000 / fsdp_optimizers
supporting pytorch FSDP for optimizers
☆75Updated last month
Alternatives and similar repositories for fsdp_optimizers:
Users that are interested in fsdp_optimizers are comparing it to the libraries listed below
- Efficient optimizers☆144Updated this week
- ☆53Updated 11 months ago
- ☆146Updated last month
- Focused on fast experimentation and simplicity☆64Updated 3 weeks ago
- ☆75Updated 6 months ago
- WIP☆92Updated 5 months ago
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆121Updated 9 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆90Updated last month
- 🧱 Modula software package☆132Updated this week
- Simple implementation of muP, based on Spectral Condition for Feature Learning. The implementation is SGD only, dont use it for Adam☆73Updated 5 months ago
- Fast, Modern, Memory Efficient, and Low Precision PyTorch Optimizers☆77Updated 6 months ago
- DeMo: Decoupled Momentum Optimization☆170Updated last month
- ☆33Updated 4 months ago
- Experiment of using Tangent to autodiff triton☆74Updated 11 months ago
- ☆49Updated 10 months ago
- Pytorch implementation of the PEER block from the paper, Mixture of A Million Experts, by Xu Owen He at Deepmind☆115Updated 4 months ago
- Muon optimizer for neural networks: >30% extra sample efficiency, <3% wallclock overhead☆210Updated last week
- Normalized Transformer (nGPT)☆145Updated last month
- Implementation of PSGD optimizer in JAX☆26Updated 2 weeks ago
- ☆78Updated 9 months ago
- A MAD laboratory to improve AI architecture designs 🧪☆102Updated last month
- ☆50Updated 3 months ago
- Understand and test language model architectures on synthetic tasks.☆175Updated this week
- An implementation of the Llama architecture, to instruct and delight☆21Updated this week
- seqax = sequence modeling + JAX☆136Updated 6 months ago
- Minimal but scalable implementation of large language models in JAX☆28Updated 2 months ago
- Explorations into the proposal from the paper "Grokfast, Accelerated Grokking by Amplifying Slow Gradients"☆95Updated 3 weeks ago
- ☆24Updated last month
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆182Updated 7 months ago
- ☆52Updated 2 months ago