bytedance / QSync
Official resporitory for "IPDPS' 24 QSync: Quantization-Minimized Synchronous Distributed Training Across Hybrid Devices".
☆19Updated 11 months ago
Alternatives and similar repositories for QSync:
Users that are interested in QSync are comparing it to the libraries listed below
- An external memory allocator example for PyTorch.☆14Updated 3 years ago
- An Attention Superoptimizer☆21Updated last month
- ThrillerFlow is a Dataflow Analysis and Codegen Framework written in Rust.☆14Updated 2 months ago
- ☆24Updated last year
- TileFusion is a highly efficient kernel template library designed to elevate the level of abstraction in CUDA C for processing tiles.☆56Updated this week
- SOTA Learning-augmented Systems☆34Updated 2 years ago
- Official repository for the paper DynaPipe: Optimizing Multi-task Training through Dynamic Pipelines☆17Updated last year
- Tacker: Tensor-CUDA Core Kernel Fusion for Improving the GPU Utilization while Ensuring QoS☆19Updated last week
- Artifact for "Apparate: Rethinking Early Exits to Tame Latency-Throughput Tensions in ML Serving" [SOSP '24]☆22Updated 2 months ago
- Supplemental materials for The ASPLOS 2025 / EuroSys 2025 Contest on Intra-Operator Parallelism for Distributed Deep Learning☆23Updated 2 months ago
- ☆19Updated 4 months ago
- DISB is a new DNN inference serving benchmark with diverse workloads and models, as well as real-world traces.☆53Updated 6 months ago
- Official Repo for "LLM-PQ: Serving LLM on Heterogeneous Clusters with Phase-Aware Partition and Adaptive Quantization"☆28Updated 11 months ago
- ☆8Updated last year
- A Cluster-Wide Model Manager to Accelerate DNN Training via Automated Training Warmup☆34Updated 2 years ago
- ☆16Updated 2 years ago
- ☆12Updated 2 years ago
- FlexFlow Serve: Low-Latency, High-Performance LLM Serving☆17Updated this week
- An extention of TVMScript to write simple and high performance GPU kernels with tensorcore.☆51Updated 6 months ago
- ☆50Updated 8 months ago
- A GPU-accelerated DNN inference serving system that supports instant kernel preemption and biased concurrent execution in GPU scheduling.☆40Updated 2 years ago
- ASPLOS'24: Optimal Kernel Orchestration for Tensor Programs with Korch☆31Updated 6 months ago
- Artifact of ASPLOS'23 paper entitled: GRACE: A Scalable Graph-Based Approach to Accelerating Recommendation Model Inference☆17Updated last year
- PyTorch compilation tutorial covering TorchScript, torch.fx, and Slapo☆19Updated last year
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆37Updated 2 years ago
- ☆20Updated this week
- ☆14Updated 3 years ago
- ☆21Updated last year
- [ICDCS 2023] DeAR: Accelerating Distributed Deep Learning with Fine-Grained All-Reduce Pipelining☆12Updated last year