quiver-team / quiver-feature
High performance RDMA-based distributed feature collection component for training GNN model on EXTREMELY large graph
☆52Updated 2 years ago
Alternatives and similar repositories for quiver-feature:
Users that are interested in quiver-feature are comparing it to the libraries listed below
- DISB is a new DNN inference serving benchmark with diverse workloads and models, as well as real-world traces.☆52Updated 8 months ago
- My paper/code reading notes in Chinese☆46Updated 11 months ago
- Tacker: Tensor-CUDA Core Kernel Fusion for Improving the GPU Utilization while Ensuring QoS☆25Updated 2 months ago
- ☆16Updated 2 years ago
- A Factored System for Sample-based GNN Training over GPUs☆42Updated last year
- FGNN's artifact evaluation (EuroSys 2022)☆17Updated 3 years ago
- Adaptive Message Quantization and Parallelization for Distributed Full-graph GNN Training☆23Updated last year
- ☆22Updated last year
- Artifact of ASPLOS'23 paper entitled: GRACE: A Scalable Graph-Based Approach to Accelerating Recommendation Model Inference☆18Updated 2 years ago
- PipeSwitch: Fast Pipelined Context Switching for Deep Learning Applications☆126Updated 3 years ago
- Vector search with bounded performance.☆34Updated last year
- REEF is a GPU-accelerated DNN inference serving system that enables instant kernel preemption and biased concurrent execution in GPU sche…☆94Updated 2 years ago
- Paella: Low-latency Model Serving with Virtualized GPU Scheduling☆58Updated last year
- Dorylus: Affordable, Scalable, and Accurate GNN Training☆77Updated 3 years ago
- SOTA Learning-augmented Systems☆36Updated 2 years ago
- A GPU-accelerated DNN inference serving system that supports instant kernel preemption and biased concurrent execution in GPU scheduling.☆42Updated 2 years ago
- Artifact for OSDI'23: MGG: Accelerating Graph Neural Networks with Fine-grained intra-kernel Communication-Computation Pipelining on Mult…☆40Updated last year
- Analysis for the traces from byteprofile☆31Updated last year
- Artifact of OSDI '24 paper, ”Llumnix: Dynamic Scheduling for Large Language Model Serving“☆60Updated 11 months ago
- Graph Sampling using GPU☆52Updated 3 years ago
- SHADE: Enable Fundamental Cacheability for Distributed Deep Learning Training☆32Updated 2 years ago
- ☆53Updated 4 years ago
- Official resporitory for "IPDPS' 24 QSync: Quantization-Minimized Synchronous Distributed Training Across Hybrid Devices".☆19Updated last year
- A lightweight design for computation-communication overlap.☆67Updated last week
- Stateful LLM Serving☆65Updated last month
- A prefill & decode disaggregated LLM serving framework with shared GPU memory and fine-grained compute isolation.☆70Updated last week
- A high-performance distributed deep learning system targeting large-scale and automated distributed training. If you have any interests, …☆111Updated last year
- Bamboo is a system for running large pipeline-parallel DNNs affordably, reliably, and efficiently using spot instances.☆49Updated 2 years ago
- ☆79Updated 2 years ago
- An efficient concurrent graph processing system☆46Updated 3 years ago