kimbochen / md-blogsLinks
A blog where I write about research papers and blog posts I read.
☆12Updated last year
Alternatives and similar repositories for md-blogs
Users that are interested in md-blogs are comparing it to the libraries listed below
Sorting:
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆155Updated 2 years ago
- Experiment of using Tangent to autodiff triton☆82Updated 2 years ago
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆86Updated 2 years ago
- ☆47Updated 2 years ago
- Solve puzzles. Learn CUDA.☆63Updated 2 years ago
- ☆92Updated last year
- seqax = sequence modeling + JAX☆170Updated 6 months ago
- ☆27Updated last year
- MoE training for Me and You and maybe other people☆335Updated last month
- Write a fast kernel and run it on Discord. See how you compare against the best!☆71Updated this week
- Custom triton kernels for training Karpathy's nanoGPT.☆19Updated last year
- PTX-Tutorial Written Purely By AIs (Deep Research of Openai and Claude 3.7)☆66Updated 10 months ago
- train with kittens!☆63Updated last year
- ☆91Updated last year
- Minimal but scalable implementation of large language models in JAX☆35Updated 2 months ago
- ML/DL Math and Method notes☆66Updated 2 years ago
- A really tiny autograd engine☆99Updated 8 months ago
- ☆178Updated 2 years ago
- Mixed precision training from scratch with Tensors and CUDA☆28Updated last year
- Compiling useful links, papers, benchmarks, ideas, etc.☆46Updated 10 months ago
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆198Updated 8 months ago
- PyTorch centric eager mode debugger☆48Updated last year
- A puzzle to learn about prompting☆135Updated 2 years ago
- ☆22Updated last year
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆115Updated last month
- A set of Python scripts that makes your experience on TPU better☆56Updated 4 months ago
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆132Updated last year
- An implementation of the Llama architecture, to instruct and delight☆21Updated 8 months ago
- gzip Predicts Data-dependent Scaling Laws☆34Updated last year
- Simple Transformer in Jax☆142Updated last year