Mellanox / nccl-rdma-sharp-pluginsLinks
RDMA and SHARP plugins for nccl library
☆200Updated 2 months ago
Alternatives and similar repositories for nccl-rdma-sharp-plugins
Users that are interested in nccl-rdma-sharp-plugins are comparing it to the libraries listed below
Sorting:
- NCCL Fast Socket is a transport layer plugin to improve NCCL collective communication performance on Google Cloud.☆120Updated last year
- NCCL Profiling Kit☆141Updated last year
- ☆362Updated last year
- Microsoft Collective Communication Library☆357Updated last year
- Synthesizer for optimal collective communication algorithms☆116Updated last year
- GPUDirect Async support for IB Verbs☆129Updated 2 years ago
- example code for using DC QP for providing RDMA READ and WRITE operations to remote GPU memory☆140Updated last year
- Fine-grained GPU sharing primitives☆143Updated 3 weeks ago
- GPU-scheduler-for-deep-learning☆210Updated 4 years ago
- Unified Collective Communication Library☆263Updated this week
- PArametrized Recommendation and Ai Model benchmark is a repository for development of numerous uBenchmarks as well as end to end nets for…☆149Updated this week
- ☆47Updated 8 months ago
- This is a plugin which lets EC2 developers use libfabric as network provider while running NCCL applications.☆183Updated last week
- MSCCL++: A GPU-driven communication stack for scalable AI applications☆398Updated this week
- Ultra and Unified CCL☆483Updated this week
- Repository for MLCommons Chakra schema and tools☆120Updated 3 weeks ago
- A tool for bandwidth measurements on NVIDIA GPUs.☆511Updated 4 months ago
- Paella: Low-latency Model Serving with Virtualized GPU Scheduling☆60Updated last year
- pytorch ucc plugin☆23Updated 4 years ago
- TACCL: Guiding Collective Algorithm Synthesis using Communication Sketches☆75Updated 2 years ago
- Hooked CUDA-related dynamic libraries by using automated code generation tools.☆165Updated last year
- NVIDIA NCCL Tests for Distributed Training☆105Updated this week
- Magnum IO community repo☆95Updated 3 months ago
- Intercepting CUDA runtime calls with LD_PRELOAD☆41Updated 11 years ago
- Microsoft Collective Communication Library☆66Updated 8 months ago
- Pytorch process group third-party plugin for UCC☆21Updated last year
- PipeSwitch: Fast Pipelined Context Switching for Deep Learning Applications☆126Updated 3 years ago
- Thunder Research Group's Collective Communication Library☆40Updated last month
- An Efficient Pipelined Data Parallel Approach for Training Large Model☆77Updated 4 years ago
- Code for "Heterogenity-Aware Cluster Scheduling Policies for Deep Learning Workloads", which appeared at OSDI 2020☆128Updated last year