pytorch-labs / triton-cpu
An experimental CPU backend for Triton (https//github.com/openai/triton)
☆39Updated this week
Alternatives and similar repositories for triton-cpu:
Users that are interested in triton-cpu are comparing it to the libraries listed below
- TileFusion is an experimental C++ macro kernel template library that elevates the abstraction level in CUDA C for tile processing. By pro…☆68Updated this week
- Unified compiler/runtime for interfacing with PyTorch Dynamo.☆99Updated 3 weeks ago
- An experimental CPU backend for Triton☆100Updated last week
- MLIR-based partitioning system☆73Updated this week
- High-speed GEMV kernels, at most 2.7x speedup compared to pytorch baseline.☆101Updated 8 months ago
- ☆73Updated 4 months ago
- Ahead of Time (AOT) Triton Math Library☆55Updated this week
- ☆25Updated this week
- extensible collectives library in triton☆84Updated 6 months ago
- ☆49Updated last year
- ☆138Updated this week
- ☆87Updated 2 weeks ago
- ☆91Updated 11 months ago
- IREE's PyTorch Frontend, based on Torch Dynamo.☆72Updated this week
- An extension library of WMMA API (Tensor Core API)☆91Updated 8 months ago
- We invite you to visit and follow our new repository at https://github.com/microsoft/TileFusion. TiledCUDA is a highly efficient kernel …☆178Updated last month
- Shared Middle-Layer for Triton Compilation☆232Updated last week
- ☆26Updated this week
- A language and compiler for irregular tensor programs.☆139Updated 3 months ago
- ☆37Updated this week
- OpenAI Triton backend for Intel® GPUs☆169Updated this week
- MatMul Performance Benchmarks for a Single CPU Core comparing both hand engineered and codegen kernels.☆129Updated last year
- A lightweight, Pythonic, frontend for MLIR☆80Updated last year
- Standalone Flash Attention v2 kernel without libtorch dependency☆106Updated 6 months ago
- Several optimization methods of half-precision general matrix vector multiplication (HGEMV) using CUDA core.☆57Updated 6 months ago
- TPP experimentation on MLIR for linear algebra☆121Updated last week
- The missing pieces (as far as boilerplate reduction goes) of the upstream MLIR python bindings.☆84Updated this week
- Benchmark code for the "Online normalizer calculation for softmax" paper☆85Updated 6 years ago
- Test suite for probing the numerical behavior of NVIDIA tensor cores☆37Updated 7 months ago
- High-Performance SGEMM on CUDA devices☆86Updated 2 months ago