dshah3 / GPU-PuzzlesLinks
Solve puzzles. Learn CUDA.
☆64Updated last year
Alternatives and similar repositories for GPU-Puzzles
Users that are interested in GPU-Puzzles are comparing it to the libraries listed below
Sorting:
- ☆91Updated last year
- ☆176Updated last year
- ☆285Updated last year
- seqax = sequence modeling + JAX☆168Updated 3 months ago
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆132Updated last year
- ☆89Updated last year
- A puzzle to learn about prompting☆135Updated 2 years ago
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆148Updated 2 years ago
- Puzzles for exploring transformers☆376Updated 2 years ago
- JAX implementation of the Llama 2 model☆215Updated last year
- supporting pytorch FSDP for optimizers☆83Updated 11 months ago
- A really tiny autograd engine☆97Updated 5 months ago
- A set of Python scripts that makes your experience on TPU better☆54Updated last month
- Minimal yet performant LLM examples in pure JAX☆198Updated last month
- An implementation of the Llama architecture, to instruct and delight☆21Updated 5 months ago
- Fast bare-bones BPE for modern tokenizer training☆168Updated 4 months ago
- Distributed pretraining of large language models (LLMs) on cloud TPU slices, with Jax and Equinox.☆24Updated last year
- Experiment of using Tangent to autodiff triton☆79Updated last year
- torchax is a PyTorch frontend for JAX. It gives JAX the ability to author JAX programs using familiar PyTorch syntax. It also provides JA…☆117Updated this week
- Minimal but scalable implementation of large language models in JAX☆35Updated 2 months ago
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆104Updated last month
- Simple Transformer in Jax☆139Updated last year
- A simple library for scaling up JAX programs☆144Updated last week
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆87Updated last year
- JAX bindings for Flash Attention v2☆97Updated last week
- An implementation of the transformer architecture onto an Nvidia CUDA kernel☆192Updated 2 years ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆171Updated 4 months ago
- ☆53Updated last year
- Custom triton kernels for training Karpathy's nanoGPT.☆19Updated last year
- Accelerated First Order Parallel Associative Scan☆189Updated last year