salykova / sgemm.c
Multi-Threaded FP32 Matrix Multiplication on x86 CPUs
☆348Updated this week
Alternatives and similar repositories for sgemm.c:
Users that are interested in sgemm.c are comparing it to the libraries listed below
- Learning about CUDA by writing PTX code.☆128Updated last year
- LLM training in simple, raw C/CUDA☆92Updated 11 months ago
- Alex Krizhevsky's original code from Google Code☆191Updated 9 years ago
- High-Performance SGEMM on CUDA devices☆90Updated 3 months ago
- throwaway GPT inference☆138Updated 10 months ago
- (WIP) A small but powerful, homemade PyTorch from scratch.☆543Updated last week
- ☆241Updated last year
- Tutorials on tinygrad☆370Updated last month
- pytorch from scratch in pure C/CUDA and python☆40Updated 6 months ago
- Fastest kernels written from scratch☆236Updated 3 weeks ago
- Fast CUDA matrix multiplication from scratch☆697Updated last year
- Accelerated General (FP32) Matrix Multiplication from scratch in CUDA☆114Updated 3 months ago
- Nvidia Instruction Set Specification Generator☆256Updated 9 months ago
- small auto-grad engine inspired from Karpathy's micrograd and PyTorch☆252Updated 5 months ago
- Learnings and programs related to CUDA☆379Updated 2 months ago
- Custom PTX Instruction Benchmark☆122Updated last month
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆534Updated this week
- Tensor library with autograd using only Rust's standard library☆67Updated 9 months ago
- UNet diffusion model in pure CUDA☆602Updated 9 months ago
- An implementation of the transformer architecture onto an Nvidia CUDA kernel☆179Updated last year
- Recreating PyTorch from scratch (C/C++, CUDA, NCCL and Python, with multi-GPU support and automatic differentiation!)☆150Updated 10 months ago
- A curated list of resources for learning and exploring Triton, OpenAI's programming language for writing efficient GPU code.☆339Updated last month
- ☆47Updated 3 weeks ago
- CUDA/Metal accelerated language model inference☆541Updated 2 weeks ago
- Yet Another Language Model: LLM inference in C++/CUDA, no libraries except for I/O☆283Updated 3 months ago
- Some CUDA example code with READMEs.☆94Updated last month
- Reference Kernels for the Leaderboard☆33Updated last week
- Visualization of cache-optimized matrix multiplication☆120Updated last month
- NVIDIA tools guide☆129Updated 3 months ago
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆130Updated last year