bertmaher / tf32_gemmLinks
Example of binding a TF32 CUTLASS GEMM kernel to PyTorch
☆12Updated last year
Alternatives and similar repositories for tf32_gemm
Users that are interested in tf32_gemm are comparing it to the libraries listed below
Sorting:
- ☆153Updated last year
- A resilient distributed training framework☆96Updated last year
- Github mirror of trition-lang/triton repo.☆119Updated this week
- extensible collectives library in triton☆91Updated 9 months ago
- Dynamic resources changes for multi-dimensional parallelism training☆30Updated 4 months ago
- ☆187Updated last year
- Synthesizer for optimal collective communication algorithms☆123Updated last year
- ☆84Updated 3 years ago
- torchcomms: a modern PyTorch communications API☆319Updated this week
- ☆100Updated last year
- NCCL Profiling Kit☆150Updated last year
- nnScaler: Compiling DNN models for Parallel Training☆123Updated 3 months ago
- A ChatGPT(GPT-3.5) & GPT-4 Workload Trace to Optimize LLM Serving Systems☆228Updated 5 months ago
- Thunder Research Group's Collective Communication Library☆46Updated 6 months ago
- This repository contains companion software for the Colfax Research paper "Categorical Foundations for CuTe Layouts".☆83Updated 3 months ago
- ☆255Updated last year
- Microsoft Collective Communication Library☆378Updated 2 years ago
- Chimera: bidirectional pipeline parallelism for efficiently training large-scale models.☆69Updated 9 months ago
- A library to analyze PyTorch traces.☆454Updated 3 weeks ago
- AMD RAD's multi-GPU Triton-based framework for seamless multi-GPU programming☆148Updated this week
- ☆77Updated 4 years ago
- ☆82Updated 7 months ago
- An experimental parallel training platform☆56Updated last year
- Microsoft Collective Communication Library☆66Updated last year
- A lightweight design for computation-communication overlap.☆207Updated 2 weeks ago
- Tritonbench is a collection of PyTorch custom operators with example inputs to measure their performance.☆308Updated this week
- MSCCL++: A GPU-driven communication stack for scalable AI applications☆449Updated this week
- ☆73Updated last year
- An interference-aware scheduler for fine-grained GPU sharing☆158Updated last month
- Shared Middle-Layer for Triton Compilation☆321Updated last month