Microsoft Collective Communication Library
☆384Sep 20, 2023Updated 2 years ago
Alternatives and similar repositories for msccl
Users that are interested in msccl are comparing it to the libraries listed below
Sorting:
- Synthesizer for optimal collective communication algorithms☆124Apr 8, 2024Updated last year
- MSCCL++: A GPU-driven communication stack for scalable AI applications☆469Feb 21, 2026Updated last week
- NCCL Profiling Kit☆152Jul 1, 2024Updated last year
- Microsoft Collective Communication Library☆66Nov 23, 2024Updated last year
- TACCL: Guiding Collective Algorithm Synthesis using Communication Sketches☆80Jul 25, 2023Updated 2 years ago
- RDMA and SHARP plugins for nccl library☆223Jan 12, 2026Updated last month
- ☆84Dec 2, 2022Updated 3 years ago
- [DEPRECATED] Moved to ROCm/rocm-systems repo☆411Updated this week
- Unified Collective Communication Library☆293Feb 19, 2026Updated last week
- Optimized primitives for collective multi-GPU communication☆4,474Updated this week
- NCCL Fast Socket is a transport layer plugin to improve NCCL collective communication performance on Google Cloud.☆122Nov 15, 2023Updated 2 years ago
- NCCL Tests☆1,441Feb 9, 2026Updated 2 weeks ago
- ASTRA-sim2.0: Modeling Hierarchical Networks and Disaggregated Systems for Large-model Training at Scale☆522Jan 3, 2026Updated last month
- ☆384Apr 23, 2024Updated last year
- Tutel MoE: Optimized Mixture-of-Experts Library, Support GptOss/DeepSeek/Kimi-K2/Qwen3 using FP8/NVFP4/MXFP4☆965Dec 21, 2025Updated 2 months ago
- ☆47Dec 13, 2024Updated last year
- A fast communication-overlapping library for tensor/expert parallelism on GPUs.☆1,261Aug 28, 2025Updated 6 months ago
- Distributed Compiler based on Triton for Parallel Systems☆1,361Feb 13, 2026Updated 2 weeks ago
- This is a plugin which lets EC2 developers use libfabric as network provider while running NCCL applications.☆205Updated this week
- Automatically Discovering Fast Parallelization Strategies for Distributed Deep Neural Network Training☆1,861Feb 20, 2026Updated last week
- Byted PyTorch Distributed for Hyperscale Training of LLMs and RLs☆938Nov 27, 2025Updated 3 months ago
- ☆49Aug 27, 2024Updated last year
- A fast GPU memory copy library based on NVIDIA GPUDirect RDMA technology☆1,345Dec 17, 2025Updated 2 months ago
- Bamboo is a system for running large pipeline-parallel DNNs affordably, reliably, and efficiently using spot instances.☆55Dec 11, 2022Updated 3 years ago
- Repository for MLCommons Chakra schema and tools☆39Dec 24, 2023Updated 2 years ago
- A large-scale simulation framework for LLM inference☆539Jul 25, 2025Updated 7 months ago
- [DEPRECATED] Moved to ROCm/rocm-systems repo☆144Updated this week
- Fine-grained GPU sharing primitives☆148Jul 28, 2025Updated 7 months ago
- Thunder Research Group's Collective Communication Library☆47Jul 8, 2025Updated 7 months ago
- gossip: Efficient Communication Primitives for Multi-GPU Systems☆62Jul 1, 2022Updated 3 years ago
- ☆392Nov 4, 2022Updated 3 years ago
- A flexible and efficient deep neural network (DNN) compiler that generates high-performance executable from a DNN model description.☆1,006Sep 19, 2024Updated last year
- ☆26Feb 17, 2025Updated last year
- A low-latency & high-throughput serving engine for LLMs☆480Jan 8, 2026Updated last month
- ☆25May 26, 2021Updated 4 years ago
- A CPU+GPU Profiling library that provides access to timeline traces and hardware performance counters.☆922Updated this week
- PipeSwitch: Fast Pipelined Context Switching for Deep Learning Applications☆127May 9, 2022Updated 3 years ago
- TACOS: [T]opology-[A]ware [Co]llective Algorithm [S]ynthesizer for Distributed Machine Learning☆32Jun 13, 2025Updated 8 months ago
- An Attention Superoptimizer☆22Jan 20, 2025Updated last year