Implementation of FusedMM method for IPDPS 2021 paper titled "FusedMM: A Unified SDDMM-SpMM Kernel for Graph Embedding and Graph Neural Networks"
☆31Aug 12, 2022Updated 3 years ago
Alternatives and similar repositories for FusedMM
Users that are interested in FusedMM are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- ☆12Mar 14, 2023Updated 3 years ago
- ☆114Jul 3, 2021Updated 4 years ago
- New batched algorithm for sparse matrix-matrix multiplication (SpMM)☆16May 7, 2019Updated 6 years ago
- NeuraChip Accelerator Simulator☆16Apr 26, 2024Updated 2 years ago
- [HPCA 2022] GCoD: Graph Convolutional Network Acceleration via Dedicated Algorithm and Accelerator Co-Design☆39Mar 30, 2022Updated 4 years ago
- Managed Kubernetes at scale on DigitalOcean • AdDigitalOcean Kubernetes includes the control plane, bandwidth allowance, container registry, automatic updates, and more for free.
- [Mlsys'22] Understanding gnn computational graph: A coordinated computation, io, and memory perspective☆22Sep 11, 2023Updated 2 years ago
- Code for paper "Design Principles for Sparse Matrix Multiplication on the GPU" accepted to Euro-Par 2018☆73Oct 5, 2020Updated 5 years ago
- ☆48Jan 30, 2026Updated 2 months ago
- Source code for the evaluated benchmarks and proposed cache management technique, GRASP, in [Faldu et al., HPCA'20].☆18Jan 23, 2020Updated 6 years ago
- A GPU algorithm for sparse matrix-matrix multiplication☆75Oct 1, 2020Updated 5 years ago
- A simulation framework for modeling efficiency of Graph Neural Network Dataflows☆24Feb 14, 2025Updated last year
- An HBM FPGA based SpMV Accelerator☆18Aug 29, 2024Updated last year
- ☆69Jun 16, 2021Updated 4 years ago
- ☆37Jan 20, 2022Updated 4 years ago
- Wordpress hosting with auto-scaling - Free Trial Offer • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- Reference implementation of the draft C++ GraphBLAS specification.☆32Feb 19, 2025Updated last year
- Artifact for USENIX ATC'23: TC-GNN: Bridging Sparse GNN Computation and Dense Tensor Cores on GPUs.☆57Oct 16, 2023Updated 2 years ago
- [TECS'23] A project on the co-design of Accelerators and CNNs.☆21Dec 10, 2022Updated 3 years ago
- ☆49Apr 11, 2025Updated last year
- Magicube is a high-performance library for quantized sparse matrix operations (SpMM and SDDMM) of deep learning on Tensor Cores.☆92Nov 23, 2022Updated 3 years ago
- HLS project modeling various sparse accelerators.☆12Jan 11, 2022Updated 4 years ago
- AdaBatch: Adaptive Batch Sizes for Training Deep Neural Networks☆42Dec 16, 2017Updated 8 years ago
- CSR-based SpGEMM on nVidia and AMD GPUs☆48Apr 9, 2016Updated 10 years ago
- [DATE'23] The official code for paper <CLAP: Locality Aware and Parallel Triangle Counting with Content Addressable Memory>☆24Mar 16, 2026Updated last month
- GPUs on demand by Runpod - Special Offer Available • AdRun AI, ML, and HPC workloads on powerful cloud GPUs—without limits or wasted spend. Deploy GPUs in under a minute and pay by the second.
- SpV8 is a SpMV kernel written in AVX-512. Artifact for our SpV8 paper @ DAC '21.☆29Mar 16, 2021Updated 5 years ago
- Efficient SpGEMM on GPU using CUDA and CSR☆60Jul 18, 2023Updated 2 years ago
- The simulator for SPADA, an SpGEMM accelerator with adaptive dataflow☆47Jan 26, 2023Updated 3 years ago
- Sparse matrix computation library for GPU☆59Jul 12, 2020Updated 5 years ago
- Instructions and templates for SC authors☆17Aug 22, 2021Updated 4 years ago
- ☆11Aug 7, 2023Updated 2 years ago
- ☆11Aug 4, 2022Updated 3 years ago
- ☆21Aug 21, 2023Updated 2 years ago
- including compiler to encode DGL GNN model to instructions, runtime software to transfer data and control the accelerator, and hardware v…☆14Nov 19, 2023Updated 2 years ago
- Wordpress hosting with auto-scaling - Free Trial Offer • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- [ISCA'25] LIA: A Single-GPU LLM Inference Acceleration with Cooperative AMX-Enabled CPU-GPU Computation and CXL Offloading☆12Jun 28, 2025Updated 10 months ago
- ☆14Aug 3, 2024Updated last year
- SparseTIR: Sparse Tensor Compiler for Deep Learning☆144Mar 31, 2023Updated 3 years ago
- A source-to-source compiler for optimizing CUDA dynamic parallelism by aggregating launches☆15Jun 21, 2019Updated 6 years ago
- Multivariate cumulants of any order☆15May 17, 2023Updated 2 years ago
- This repo is to collect the state-of-the-art GNN hardware acceleration paper☆55Jun 8, 2021Updated 4 years ago
- ☆16Nov 22, 2022Updated 3 years ago