KrishM123 / transformer.cppLinks
TransformerCPP is a minimal C++ machine learning library with autograd and tensor ops, inspired by PyTorch. It includes a from-scratch Transformer model demo, optimized for CPU and multithreaded performance.
☆47Updated 2 months ago
Alternatives and similar repositories for transformer.cpp
Users that are interested in transformer.cpp are comparing it to the libraries listed below
Sorting:
- Accelerated General (FP32) Matrix Multiplication from scratch in CUDA☆181Updated last year
- small auto-grad engine inspired from Karpathy's micrograd and PyTorch☆276Updated last year
- A minimal Tensor Processing Unit (TPU) inspired by Google's TPUv1.☆194Updated last year
- Learnings and programs related to CUDA☆432Updated 7 months ago
- Competitive GPU kernel optimization platform.☆149Updated 2 weeks ago
- Learning about CUDA by writing PTX code.☆151Updated last year
- 100 days of CUDA Challenge☆47Updated 6 months ago
- pytorch from scratch in pure C/CUDA and python☆41Updated last year
- GPT-2 in C☆78Updated last year
- Some CUDA example code with READMEs.☆179Updated 2 months ago
- Multi-Threaded FP32 Matrix Multiplication on x86 CPUs☆376Updated 9 months ago
- Visualization of cache-optimized matrix multiplication☆157Updated 10 months ago
- An implement of deep learning framework and models in C☆47Updated 10 months ago
- Quantized LLM training in pure CUDA/C++.☆235Updated 2 weeks ago
- A curriculum for learning about gpu performance engineering, from scratch to what the frontier AI labs do☆341Updated 3 weeks ago
- ☆96Updated last year
- (WIP) A small but powerful, homemade PyTorch from scratch.☆672Updated last week
- a tiny multidimensional array implementation in C similar to numpy, but only one file.☆225Updated last year
- SIMD quantization kernels☆94Updated 4 months ago
- Solve puzzles to improve your tinygrad skills!☆178Updated 3 months ago
- Tensor library with autograd using only Rust's standard library☆71Updated last year
- Simple MPI implementation for prototyping or learning☆300Updated 6 months ago
- could we make an ml stack in 100,000 lines of code?☆46Updated last year
- in this repository, i'm going to implement increasingly complex llm inference optimizations☆81Updated 8 months ago
- Complete solutions to the Programming Massively Parallel Processors Edition 4☆655Updated 7 months ago
- ☆118Updated last month
- High Quality Resources on GPU Programming/Architecture☆591Updated last year
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆155Updated 2 years ago
- An implementation of the transformer architecture onto an Nvidia CUDA kernel☆202Updated 2 years ago
- Notes on "Programming Massively Parallel Processors" by Hwu, Kirk, and Hajj (4th ed.)☆53Updated last year