mlcommons / algorithmic-efficiencyLinks
MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvements in both training algorithms and models.
☆389Updated this week
Alternatives and similar repositories for algorithmic-efficiency
Users that are interested in algorithmic-efficiency are comparing it to the libraries listed below
Sorting:
- For optimization algorithm research and development.☆524Updated this week
- ☆232Updated 5 months ago
- ☆275Updated last year
- jax-triton contains integrations between JAX and OpenAI Triton☆412Updated last month
- 🧱 Modula software package☆216Updated last week
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆630Updated this week
- Universal Tensor Operations in Einstein-Inspired Notation for Python.☆392Updated 3 months ago
- Named tensors with first-class dimensions for PyTorch☆332Updated 2 years ago
- CIFAR-10 speedruns: 94% in 2.6 seconds and 96% in 27 seconds☆275Updated 3 weeks ago
- Home for "How To Scale Your Model", a short blog-style textbook about scaling LLMs on TPUs☆445Updated last week
- ☆442Updated 9 months ago
- Pytorch implementation of preconditioned stochastic gradient descent (Kron and affine preconditioner, low-rank approximation precondition…☆180Updated last week
- ☆323Updated this week
- Annotated version of the Mamba paper☆487Updated last year
- JAX Synergistic Memory Inspector☆177Updated last year
- Orbax provides common checkpointing and persistence utilities for JAX users☆409Updated this week
- A library for unit scaling in PyTorch☆128Updated 3 weeks ago
- TensorDict is a pytorch dedicated tensor container.☆949Updated last week
- Unofficial JAX implementations of deep learning research papers☆156Updated 3 years ago
- Library for reading and processing ML training data.☆487Updated this week
- CLU lets you write beautiful training loops in JAX.☆351Updated last month
- Puzzles for exploring transformers☆356Updated 2 years ago
- ☆187Updated last week
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.☆290Updated 11 months ago
- A simple library for scaling up JAX programs☆140Updated 9 months ago
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆522Updated this week
- Efficient optimizers☆252Updated last week
- ☆304Updated last year
- JMP is a Mixed Precision library for JAX.☆207Updated 6 months ago
- Implementation of Flash Attention in Jax☆215Updated last year