AI-HPC-Research-Team / AIPerfLinks
Automated machine learning as an AI-HPC benchmark
☆65Updated 3 years ago
Alternatives and similar repositories for AIPerf
Users that are interested in AIPerf are comparing it to the libraries listed below
Sorting:
- Synthesizer for optimal collective communication algorithms☆118Updated last year
- Paella: Low-latency Model Serving with Virtualized GPU Scheduling☆62Updated last year
- RDMA and SHARP plugins for nccl library☆210Updated this week
- A prefill & decode disaggregated LLM serving framework with shared GPU memory and fine-grained compute isolation.☆113Updated 5 months ago
- NCCL Examples from Official NVIDIA NCCL Developer Guide.☆19Updated 7 years ago
- NCCL Profiling Kit☆145Updated last year
- GPU-scheduler-for-deep-learning☆210Updated 4 years ago
- AI Accelerator Benchmark focuses on evaluating AI Accelerators from a practical production perspective, including the ease of use and ver…☆265Updated 2 months ago
- ☆24Updated 3 years ago
- PipeSwitch: Fast Pipelined Context Switching for Deep Learning Applications☆126Updated 3 years ago
- Fine-grained GPU sharing primitives☆146Updated 3 months ago
- Artifact of OSDI '24 paper, ”Llumnix: Dynamic Scheduling for Large Language Model Serving“☆62Updated last year
- GVProf: A Value Profiler for GPU-based Clusters☆52Updated last year
- ☆83Updated 2 years ago
- ☆154Updated last year
- DISB is a new DNN inference serving benchmark with diverse workloads and models, as well as real-world traces.☆54Updated last year
- AI and Memory Wall☆219Updated last year
- PET: Optimizing Tensor Programs with Partially Equivalent Transformations and Automated Corrections☆122Updated 3 years ago
- gossip: Efficient Communication Primitives for Multi-GPU Systems☆59Updated 3 years ago
- NCCL Fast Socket is a transport layer plugin to improve NCCL collective communication performance on Google Cloud.☆121Updated last year
- Thunder Research Group's Collective Communication Library☆42Updated 3 months ago
- TACOS: [T]opology-[A]ware [Co]llective Algorithm [S]ynthesizer for Distributed Machine Learning☆27Updated 4 months ago
- ☆46Updated 10 months ago
- An Efficient Pipelined Data Parallel Approach for Training Large Model☆76Updated 4 years ago
- ☆376Updated last year
- An unofficial cuda assembler, for all generations of SASS, hopefully :)☆83Updated 2 years ago
- ☆58Updated 5 years ago
- A home for the final text of all TVM RFCs.☆108Updated last year
- ☆53Updated 10 months ago
- High performance RDMA-based distributed feature collection component for training GNN model on EXTREMELY large graph☆55Updated 3 years ago