jsbaan / transformer-from-scratch
Well documented, unit tested, type checked and formatted implementation of a vanilla transformer - for educational purposes.
☆211Updated 5 months ago
Related projects: ⓘ
- LLaMA 2 implemented from scratch in PyTorch☆216Updated 11 months ago
- A Simplified PyTorch Implementation of Vision Transformer (ViT)☆123Updated 3 months ago
- Annotated version of the Mamba paper☆445Updated 6 months ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆248Updated 10 months ago
- Tutorial for how to build BERT from scratch☆81Updated 3 months ago
- Puzzles for exploring transformers☆293Updated last year
- Fast bare-bones BPE for modern tokenizer training☆138Updated 3 weeks ago
- An implementation of the transformer architecture onto an Nvidia CUDA kernel☆152Updated 11 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆89Updated last week
- Code repository for the paper - "Matryoshka Representation Learning"☆398Updated 7 months ago
- ☆124Updated 7 months ago
- GPT-2 (124M) quality in 5B tokens☆227Updated last week
- Interpretability for sequence generation models 🐛 🔍☆361Updated 3 weeks ago
- Recreating PyTorch from scratch (C/C++, CUDA and Python, with multi-GPU support and automatic differentiation!)☆89Updated 3 months ago
- An open collection of implementation tips, tricks and resources for training large language models☆455Updated last year
- MinT: Minimal Transformer Library and Tutorials☆247Updated 2 years ago
- Outlining techniques for improving the training performance of your PyTorch model without compromising its accuracy☆124Updated last year
- I will build Transformer from scratch☆48Updated 4 months ago
- An interactive exploration of Transformer programming.☆243Updated 10 months ago
- A walkthrough of transformer architecture code☆295Updated 7 months ago
- Code implementation from my blog post: https://fkodom.substack.com/p/transformers-from-scratch-in-pytorch☆90Updated last year
- Llama from scratch, or How to implement a paper without crying☆499Updated 3 months ago
- ☆160Updated last year
- Implementation of Diffusion Transformer (DiT) in JAX☆246Updated 3 months ago
- ☆325Updated 11 months ago
- Implementation of https://srush.github.io/annotated-s4☆457Updated last year
- Project 2 (Building Large Language Models) for Stanford CS324: Understanding and Developing Large Language Models (Winter 2022)☆101Updated last year
- Building blocks for foundation models.☆347Updated 8 months ago
- Training small GPT-2 style models using Kolmogorov-Arnold networks.☆105Updated 3 months ago
- Repo for "Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture"☆530Updated 4 months ago