leo811121 / UIUC-CS-483-Parallel-Programming
☆18Updated 4 years ago
Alternatives and similar repositories for UIUC-CS-483-Parallel-Programming:
Users that are interested in UIUC-CS-483-Parallel-Programming are comparing it to the libraries listed below
- ring-attention experiments☆123Updated 3 months ago
- A minimal implementation of vllm.☆33Updated 6 months ago
- Learning about CUDA by writing PTX code.☆35Updated 11 months ago
- ☆141Updated last year
- Cataloging released Triton kernels.☆164Updated last month
- Collection of kernels written in Triton language☆97Updated this week
- Examples and exercises from the book Programming Massively Parallel Processors - A Hands-on Approach. David B. Kirk and Wen-mei W. Hwu (T…☆63Updated 4 years ago
- Tritonbench is a collection of PyTorch custom operators with example inputs to measure their performance.☆86Updated this week
- ☆175Updated this week
- Mixed precision training from scratch with Tensors and CUDA☆21Updated 9 months ago
- ☆156Updated last year
- 📚FFPA: Yet antother Faster Flash Prefill Attention with O(1)⚡️SRAM complexity for headdim > 256, 1.8x~3x↑🎉faster than SDPA EA.☆96Updated this week
- [ICLR2025] Breaking Throughput-Latency Trade-off for Long Sequences with Speculative Decoding☆107Updated 2 months ago
- Course materials for MIT6.5940: TinyML and Efficient Deep Learning Computing☆26Updated last month
- ☆16Updated 10 months ago
- extensible collectives library in triton☆82Updated 4 months ago
- KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems☆166Updated this week
- This repository contains the experimental PyTorch native float8 training UX☆221Updated 6 months ago
- Implement Flash Attention using Cute.☆69Updated last month
- Puzzles for learning Triton, play it with minimal environment configuration!☆222Updated 2 months ago
- Step-by-step optimization of CUDA SGEMM☆280Updated 2 years ago
- Odysseus: Playground of LLM Sequence Parallelism☆64Updated 7 months ago
- torch.compile artifacts for common deep learning models, can be used as a learning resource for torch.compile☆16Updated last year
- The simplest but fast implementation of matrix multiplication in CUDA.☆34Updated 6 months ago
- Experiment of using Tangent to autodiff triton☆75Updated last year
- ☆58Updated 2 months ago
- ☆64Updated last year
- A sparse attention kernel supporting mix sparse patterns☆108Updated this week
- TORCH_LOGS parser for PT2☆31Updated this week