meta-pytorch / MSLKLinks
MSLK (Meta Superintelligence Labs Kernels) is a collection of PyTorch GPU operator libraries that are designed and optimized for GenAI training and inference, such as FP8 row-wise quantization and collective communications.
☆45Updated last week
Alternatives and similar repositories for MSLK
Users that are interested in MSLK are comparing it to the libraries listed below
Sorting:
- TileFusion is an experimental C++ macro kernel template library that elevates the abstraction level in CUDA C for tile processing.☆106Updated 7 months ago
- ☆88Updated 8 months ago
- ☆53Updated 9 months ago
- Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.☆84Updated 2 weeks ago
- incubator repo for CUDA-TileIR backend☆97Updated 3 weeks ago
- We invite you to visit and follow our new repository at https://github.com/microsoft/TileFusion. TiledCUDA is a highly efficient kernel …☆192Updated last year
- DeeperGEMM: crazy optimized version☆73Updated 9 months ago
- Tutorials for NVIDIA CUPTI samples☆50Updated 3 months ago
- ☆65Updated 9 months ago
- Framework to reduce autotune overhead to zero for well known deployments.☆96Updated 4 months ago
- NVSHMEM‑Tutorial: Build a DeepEP‑like GPU Buffer☆161Updated 4 months ago
- An extention of TVMScript to write simple and high performance GPU kernels with tensorcore.☆51Updated last year
- a size profiler for cuda binary☆70Updated 3 weeks ago
- ☆20Updated last year
- Artifacts of EVT ASPLOS'24☆28Updated last year
- FractalTensor is a programming framework that introduces a novel approach to organizing data in deep neural networks (DNNs) as a list of …☆32Updated last year
- ⚡️Write HGEMM from scratch using Tensor Cores with WMMA, MMA and CuTe API, Achieve Peak⚡️ Performance.☆148Updated 9 months ago
- Tile-based language built for AI computation across all scales☆120Updated this week
- ☆104Updated last year
- AMD RAD's multi-GPU Triton-based framework for seamless multi-GPU programming☆168Updated this week
- Accepted to MLSys 2026☆70Updated last week
- A lightweight design for computation-communication overlap.☆219Updated 3 weeks ago
- ☆38Updated 6 months ago
- GPTQ inference TVM kernel☆40Updated last year
- DLSlime: Flexible & Efficient Heterogeneous Transfer Toolkit☆92Updated 2 weeks ago
- ☆60Updated last week
- High-speed GEMV kernels, at most 2.7x speedup compared to pytorch baseline.☆127Updated last year
- A Triton JIT runtime and ffi provider in C++☆31Updated 2 weeks ago
- Standalone Flash Attention v2 kernel without libtorch dependency☆114Updated last year
- A standalone GEMM kernel for fp16 activation and quantized weight, extracted from FasterTransformer☆96Updated 4 months ago