joey00072 / Tinytorch
A really tiny autograd engine
☆90Updated 11 months ago
Alternatives and similar repositories for Tinytorch:
Users that are interested in Tinytorch are comparing it to the libraries listed below
- Solve puzzles. Learn CUDA.☆63Updated last year
- Following master Karpathy with GPT-2 implementation and training, writing lots of comments cause I have memory of a goldfish☆169Updated 7 months ago
- Simple Transformer in Jax☆136Updated 8 months ago
- Just large language models. Hackable, with as little abstraction as possible. Done for my own purposes, feel free to rip.☆44Updated last year
- ☆75Updated 8 months ago
- Simple Byte pair Encoding mechanism used for tokenization process . written purely in C☆128Updated 4 months ago
- ☆98Updated 10 months ago
- An implementation of the transformer architecture onto an Nvidia CUDA kernel☆173Updated last year
- Alex Krizhevsky's original code from Google Code☆190Updated 9 years ago
- Collection of autoregressive model implementation☆83Updated 3 weeks ago
- ☆148Updated last year
- A puzzle to learn about prompting☆124Updated last year
- Andrej Kapathy's micrograd implemented in c☆28Updated 7 months ago
- Cerule - A Tiny Mighty Vision Model☆67Updated 6 months ago
- supporting pytorch FSDP for optimizers☆79Updated 3 months ago
- ☆60Updated last year
- Learning about CUDA by writing PTX code.☆123Updated last year
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆125Updated last year
- Tensor library with autograd using only Rust's standard library☆65Updated 8 months ago
- pytorch from scratch in pure C/CUDA and python☆40Updated 5 months ago
- An introduction to LLM Sampling☆76Updated 2 months ago
- Fast bare-bones BPE for modern tokenizer training☆148Updated 4 months ago
- ☆23Updated 7 months ago
- Full finetuning of large language models without large memory requirements☆93Updated last year
- ☆27Updated 8 months ago
- Implementation of the Llama architecture with RLHF + Q-learning☆163Updated last month
- A curated list of resources for learning and exploring Triton, OpenAI's programming language for writing efficient GPU code.☆300Updated this week
- Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wi…☆343Updated 7 months ago