lucasdelimanogueira / PyNorch
Recreating PyTorch from scratch (C/C++, CUDA, NCCL and Python, with multi-GPU support and automatic differentiation!)
☆114Updated 5 months ago
Related projects ⓘ
Alternatives and complementary repositories for PyNorch
- Alex Krizhevsky's original code from Google Code☆190Updated 8 years ago
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆483Updated 3 weeks ago
- ring-attention experiments☆97Updated last month
- ☆133Updated 9 months ago
- Simple Byte pair Encoding mechanism used for tokenization process . written purely in C☆120Updated last week
- UNet diffusion model in pure CUDA☆584Updated 4 months ago
- Implementation of Diffusion Transformer (DiT) in JAX☆252Updated 5 months ago
- Cataloging released Triton kernels.☆134Updated 2 months ago
- ☆44Updated last week
- The Tensor (or Array)☆411Updated 3 months ago
- Mixed precision training from scratch with Tensors and CUDA☆20Updated 6 months ago
- Triton implementation of GPT/LLAMA☆16Updated 2 months ago
- An implementation of the transformer architecture onto an Nvidia CUDA kernel☆157Updated last year
- ☆152Updated this week
- A really tiny autograd engine☆87Updated 7 months ago
- Deep learning library implemented from scratch in numpy. Mixtral, Mamba, LLaMA, GPT, ResNet, and other experiments.☆48Updated 7 months ago
- Best practices & guides on how to write distributed pytorch training code☆286Updated 2 weeks ago
- Documented and Unit Tested educational Deep Learning framework with Autograd from scratch.☆105Updated 7 months ago
- From zero to hero CUDA for accelerating maths and machine learning on GPU.☆171Updated 3 months ago
- The Autograd Engine☆534Updated 2 months ago
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆107Updated last year
- LLaMA 2 implemented from scratch in PyTorch☆254Updated last year
- Distributed training (multi-node) of a Transformer model☆43Updated 7 months ago
- This code repository contains the code used for my "Optimizing Memory Usage for Training LLMs and Vision Transformers in PyTorch" blog po…☆86Updated last year
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆193Updated this week
- Fast bare-bones BPE for modern tokenizer training☆142Updated last month
- Solve puzzles. Learn CUDA.☆61Updated 11 months ago
- For optimization algorithm research and development.☆449Updated this week
- High-Performance FP32 Matrix Multiplication on CPU☆301Updated this week
- The simplest but fast implementation of matrix multiplication in CUDA.☆33Updated 3 months ago