devpatelio / micromlLinks
☆15Updated 6 months ago
Alternatives and similar repositories for microml
Users that are interested in microml are comparing it to the libraries listed below
Sorting:
- ☆532Updated 4 months ago
- Implementation of Diffusion Transformer (DiT) in JAX☆298Updated last year
- Dion optimizer algorithm☆403Updated last week
- Home for "How To Scale Your Model", a short blog-style textbook about scaling LLMs on TPUs☆724Updated 2 weeks ago
- Minimal yet performant LLM examples in pure JAX☆207Updated last week
- ☆285Updated last year
- ☆547Updated last year
- seqax = sequence modeling + JAX☆168Updated 4 months ago
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆685Updated 3 weeks ago
- ☆128Updated 3 weeks ago
- Library for reading and processing ML training data.☆621Updated this week
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆105Updated 2 months ago
- For optimization algorithm research and development.☆552Updated this week
- Efficient optimizers☆277Updated last month
- Async RL Training at Scale☆938Updated this week
- Accelerate, Optimize performance with streamlined training and serving options with JAX.☆325Updated this week
- Simple MPI implementation for prototyping or learning☆292Updated 4 months ago
- Tutorials on tinygrad☆444Updated 2 months ago
- UNet diffusion model in pure CUDA☆655Updated last year
- ☆108Updated last week
- Solve puzzles. Learn CUDA.☆64Updated 2 years ago
- ☆224Updated 3 weeks ago
- Cost aware hyperparameter tuning algorithm☆176Updated last year
- Platform for evaluating reinforcement learning (RL) algorithms on a physical Atari system.☆134Updated 3 months ago
- 🧱 Modula software package☆315Updated 3 months ago
- Deep learning for dummies. All the practical details and useful utilities that go into working with real models.☆829Updated 4 months ago
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆196Updated 6 months ago
- Where GPUs get cooked 👩🍳🔥☆326Updated 2 months ago
- ☆460Updated last year
- FlexAttention based, minimal vllm-style inference engine for fast Gemma 2 inference.☆320Updated last month