Shigangli / Magicube
Magicube is a high-performance library for quantized sparse matrix operations (SpMM and SDDMM) of deep learning on Tensor Cores.
☆79Updated last year
Related projects: ⓘ
- ☆30Updated 2 years ago
- ☆15Updated 3 months ago
- Automatic Mapping Generation, Verification, and Exploration for ISA-based Spatial Accelerators☆100Updated last year
- ☆44Updated 5 years ago
- Mirror of http://gitlab.hpcrl.cse.ohio-state.edu/chong/ppopp19_ae, refactoring for understanding☆12Updated 2 years ago
- ☆73Updated 5 months ago
- A Row Decomposition-based Approach for Sparse Matrix Multiplication on GPUs☆10Updated 9 months ago
- ☆72Updated last year
- ☆40Updated 4 months ago
- ☆15Updated last year
- A Vectorized N:M Format for Unleashing the Power of Sparse Tensor Cores☆31Updated 9 months ago
- ☆19Updated 5 months ago
- Dissecting NVIDIA GPU Architecture☆78Updated 2 years ago
- Artifact for PPoPP22 QGTC: Accelerating Quantized GNN via GPU Tensor Core.☆27Updated 2 years ago
- ☆38Updated 4 years ago
- Source code of the PPoPP '22 paper: "TileSpGEMM: A Tiled Algorithm for Parallel Sparse General Matrix-Matrix Multiplication on GPUs" by Y…☆35Updated 3 months ago
- SparseTIR: Sparse Tensor Compiler for Deep Learning☆129Updated last year
- DietCode Code Release☆59Updated 2 years ago
- MAGIS: Memory Optimization via Coordinated Graph Transformation and Scheduling for DNN (ASPLOS'24)☆35Updated 3 months ago
- Code for paper "Design Principles for Sparse Matrix Multiplication on the GPU" accepted to Euro-Par 2018☆70Updated 3 years ago
- ☆15Updated 2 months ago
- PyTorch-Based Fast and Efficient Processing for Various Machine Learning Applications with Diverse Sparsity☆95Updated last month
- TileFlow is a performance analysis tool based on Timeloop for fusion dataflows☆53Updated 5 months ago
- Artifacts of EVT ASPLOS'24☆12Updated 6 months ago
- Repository for artifact evaluation of ASPLOS 2023 paper "SparseTIR: Composable Abstractions for Sparse Compilation in Deep Learning"☆23Updated last year
- Source code of the SC '23 paper: "DASP: Specific Dense Matrix Multiply-Accumulate Units Accelerated General Sparse Matrix-Vector Multipli…☆16Updated 3 months ago
- Matrix Multiply-Accumulate with CUDA and WMMA( Tensor Core)☆109Updated 4 years ago
- ☆81Updated 4 months ago
- ☆9Updated 2 years ago
- Implementation of TSM2L and TSM2R -- High-Performance Tall-and-Skinny Matrix-Matrix Multiplication Algorithms for CUDA☆31Updated 4 years ago