Laz4rz / GPT-2Links
Following Karpathy with GPT-2 implementation and training, writing lots of comments cause I have memory of a goldfish
☆172Updated last year
Alternatives and similar repositories for GPT-2
Users that are interested in GPT-2 are comparing it to the libraries listed below
Sorting:
- small auto-grad engine inspired from Karpathy's micrograd and PyTorch☆276Updated last year
- Simple Transformer in Jax☆140Updated last year
- in this repository, i'm going to implement increasingly complex llm inference optimizations☆79Updated 7 months ago
- High Quality Resources on GPU Programming/Architecture☆589Updated last year
- Learnings and programs related to CUDA☆432Updated 6 months ago
- This repo has all the basic things you'll need in-order to understand complete vision transformer architecture and its various implementa…☆229Updated last year
- (WIP) A small but powerful, homemade PyTorch from scratch.☆666Updated 2 weeks ago
- A really tiny autograd engine☆98Updated 7 months ago
- could we make an ml stack in 100,000 lines of code?☆46Updated last year
- a tiny multidimensional array implementation in C similar to numpy, but only one file.☆225Updated last year
- ☆96Updated last year
- a tiny vectorstore implementation built with numpy.☆63Updated last year
- ☆537Updated 5 months ago
- Solve puzzles to improve your tinygrad skills!☆175Updated 2 months ago
- ☆95Updated 11 months ago
- ☆92Updated last year
- Deep learning for dummies. All the practical details and useful utilities that go into working with real models.☆829Updated 5 months ago
- Simple Byte pair Encoding mechanism used for tokenization process . written purely in C☆142Updated last year
- Tutorials on tinygrad☆449Updated 3 months ago
- Fast bare-bones BPE for modern tokenizer training☆174Updated 6 months ago
- A comprehensive deep dive into the world of tokens☆227Updated last year
- Alex Krizhevsky's original code from Google Code☆198Updated 9 years ago
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆112Updated 2 weeks ago
- rl from zero pretrain, can it be done? yes.☆286Updated 3 months ago
- A simple MLX implementation for pretraining LLMs on Apple Silicon.☆85Updated 4 months ago
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆195Updated 7 months ago
- UNet diffusion model in pure CUDA☆659Updated last year
- Compiling useful links, papers, benchmarks, ideas, etc.☆46Updated 9 months ago
- Gradient descent is cool and all, but what if we could delete it?☆105Updated 4 months ago
- ComplexTensor: Machine Learning By Bridging Classical and Quantum Computation☆79Updated last year