xiezhq-hermann / graphilerLinks
Graphiler is a compiler stack built on top of DGL and TorchScript which compiles GNNs defined using user-defined functions (UDFs) into efficient execution plans.
☆60Updated 2 years ago
Alternatives and similar repositories for graphiler
Users that are interested in graphiler are comparing it to the libraries listed below
Sorting:
- ☆70Updated 4 years ago
- ☆107Updated 3 years ago
- Artifact for OSDI'23: MGG: Accelerating Graph Neural Networks with Fine-grained intra-kernel Communication-Computation Pipelining on Mult…☆39Updated last year
- Artifact for PPoPP20 "Understanding and Bridging the Gaps in Current GNN Performance Optimizations"☆39Updated 3 years ago
- Distributed Multi-GPU GNN Framework☆36Updated 4 years ago
- Artifact for OSDI'21 GNNAdvisor: An Adaptive and Efficient Runtime System for GNN Acceleration on GPUs.☆66Updated 2 years ago
- ☆45Updated 2 years ago
- FGNN's artifact evaluation (EuroSys 2022)☆17Updated 3 years ago
- Artifact for USENIX ATC'23: TC-GNN: Bridging Sparse GNN Computation and Dense Tensor Cores on GPUs.☆47Updated last year
- [ICLR 2022] "PipeGCN: Efficient Full-Graph Training of Graph Convolutional Networks with Pipelined Feature Communication" by Cheng Wan, Y…☆33Updated 2 years ago
- Artifact for PPoPP22 QGTC: Accelerating Quantized GNN via GPU Tensor Core.☆29Updated 3 years ago
- Repo for the IISWC 2018 submission☆9Updated 3 years ago
- A Framework for Graph Sampling and Random Walk on GPUs.☆39Updated 4 months ago
- PyTorch-Direct code on top of PyTorch-1.8.0nightly (e152ca5) for Large Graph Convolutional Network Training with GPU-Oriented Data Commun…☆44Updated last year
- ☆27Updated 10 months ago
- Graph Sampling using GPU☆52Updated 3 years ago
- Artifact evaluation of the paper "Accelerating Training and Inference of Graph Neural Networks with Fast Sampling and Pipelining"☆23Updated 3 years ago
- A Factored System for Sample-based GNN Training over GPUs☆42Updated last year
- Ginex: SSD-enabled Billion-scale Graph Neural Network Training on a Single Machine via Provably Optimal In-memory Caching☆38Updated 11 months ago
- Set of datasets for the deep learning recommendation model (DLRM).☆47Updated 2 years ago
- Repository for artifact evaluation of ASPLOS 2023 paper "SparseTIR: Composable Abstractions for Sparse Compilation in Deep Learning"☆25Updated 2 years ago
- ☆9Updated 3 years ago
- Dorylus: Affordable, Scalable, and Accurate GNN Training☆76Updated 4 years ago
- [MLSys 2022] "BNS-GCN: Efficient Full-Graph Training of Graph Convolutional Networks with Partition-Parallelism and Random Boundary Node …☆56Updated last year
- Adaptive Message Quantization and Parallelization for Distributed Full-graph GNN Training☆24Updated last year
- An extention of TVMScript to write simple and high performance GPU kernels with tensorcore.☆50Updated 11 months ago
- ☆9Updated last year
- SoCC'20 and TPDS'21: Scaling GNN Training on Large Graphs via Computation-aware Caching and Partitioning.☆50Updated 2 years ago
- Tacker: Tensor-CUDA Core Kernel Fusion for Improving the GPU Utilization while Ensuring QoS☆27Updated 4 months ago
- MAGIS: Memory Optimization via Coordinated Graph Transformation and Scheduling for DNN (ASPLOS'24)☆52Updated last year