SJTU-IPADS / fgnn-artifactsLinks
FGNN's artifact evaluation (EuroSys 2022)
☆17Updated 3 years ago
Alternatives and similar repositories for fgnn-artifacts
Users that are interested in fgnn-artifacts are comparing it to the libraries listed below
Sorting:
- A Factored System for Sample-based GNN Training over GPUs☆42Updated last year
- Adaptive Message Quantization and Parallelization for Distributed Full-graph GNN Training☆24Updated last year
- ☆27Updated 9 months ago
- Dorylus: Affordable, Scalable, and Accurate GNN Training☆76Updated 4 years ago
- Artifact for OSDI'21 GNNAdvisor: An Adaptive and Efficient Runtime System for GNN Acceleration on GPUs.☆66Updated 2 years ago
- A Framework for Graph Sampling and Random Walk on GPUs.☆39Updated 3 months ago
- Artifact for OSDI'23: MGG: Accelerating Graph Neural Networks with Fine-grained intra-kernel Communication-Computation Pipelining on Mult…☆40Updated last year
- Graphiler is a compiler stack built on top of DGL and TorchScript which compiles GNNs defined using user-defined functions (UDFs) into ef…☆61Updated 2 years ago
- Compiler for Dynamic Neural Networks☆46Updated last year
- Distributed Multi-GPU GNN Framework☆37Updated 4 years ago
- ☆22Updated last year
- Ginex: SSD-enabled Billion-scale Graph Neural Network Training on a Single Machine via Provably Optimal In-memory Caching☆38Updated 10 months ago
- My paper/code reading notes in Chinese☆46Updated last year
- Graph Sampling using GPU☆52Updated 3 years ago
- ☆46Updated 2 years ago
- DISB is a new DNN inference serving benchmark with diverse workloads and models, as well as real-world traces.☆52Updated 9 months ago
- ☆33Updated last week
- Artifact evaluation of the paper "Accelerating Training and Inference of Graph Neural Networks with Fast Sampling and Pipelining"☆23Updated 3 years ago
- SHADE: Enable Fundamental Cacheability for Distributed Deep Learning Training☆33Updated 2 years ago
- Artifacts for our ASPLOS'23 paper ElasticFlow☆51Updated last year
- SoCC'20 and TPDS'21: Scaling GNN Training on Large Graphs via Computation-aware Caching and Partitioning.☆50Updated 2 years ago
- FlashMob is a shared-memory random walk system.☆32Updated last year
- Bamboo is a system for running large pipeline-parallel DNNs affordably, reliably, and efficiently using spot instances.☆50Updated 2 years ago
- ☆53Updated 4 years ago
- ☆18Updated 4 years ago
- A GPU-accelerated DNN inference serving system that supports instant kernel preemption and biased concurrent execution in GPU scheduling.☆42Updated 3 years ago
- ☆34Updated 11 months ago
- Artifact for PPoPP20 "Understanding and Bridging the Gaps in Current GNN Performance Optimizations"☆38Updated 3 years ago
- ☆8Updated 3 years ago
- High performance RDMA-based distributed feature collection component for training GNN model on EXTREMELY large graph☆54Updated 2 years ago