manishucsd / py-codegenLinks
☆16Updated last year
Alternatives and similar repositories for py-codegen
Users that are interested in py-codegen are comparing it to the libraries listed below
Sorting:
- TileFusion is an experimental C++ macro kernel template library that elevates the abstraction level in CUDA C for tile processing.☆105Updated 5 months ago
- ☆52Updated 7 months ago
- ☆50Updated last year
- MLIR-based partitioning system☆151Updated this week
- extensible collectives library in triton☆91Updated 8 months ago
- High-speed GEMV kernels, at most 2.7x speedup compared to pytorch baseline.☆123Updated last year
- A library of GPU kernels for sparse matrix operations.☆277Updated 5 years ago
- Ahead of Time (AOT) Triton Math Library☆84Updated last week
- A lightweight, Pythonic, frontend for MLIR☆80Updated 2 years ago
- GEMM and Winograd based convolutions using CUTLASS☆28Updated 5 years ago
- AMD RAD's multi-GPU Triton-based framework for seamless multi-GPU programming☆133Updated this week
- An experimental CPU backend for Triton (https//github.com/openai/triton)☆47Updated 4 months ago
- cuASR: CUDA Algebra for Semirings☆42Updated 3 years ago
- This repository contains companion software for the Colfax Research paper "Categorical Foundations for CuTe Layouts".☆83Updated 2 months ago
- SparseTIR: Sparse Tensor Compiler for Deep Learning☆140Updated 2 years ago
- ☆28Updated 11 months ago
- ☆97Updated last year
- An extension library of WMMA API (Tensor Core API)☆109Updated last year
- Training neural networks in TensorFlow 2.0 with 5x less memory☆137Updated 3 years ago
- JaxPP is a library for JAX that enables flexible MPMD pipeline parallelism for large-scale LLM training☆60Updated last month
- ☆23Updated 3 months ago
- CUDA templates for tile-sparse matrix multiplication based on CUTLASS.☆50Updated 7 years ago
- High-Performance SGEMM on CUDA devices☆113Updated 10 months ago
- Github mirror of trition-lang/triton repo.☆105Updated this week
- ☆253Updated last year
- ☆110Updated last year
- Matrix Multiply-Accumulate with CUDA and WMMA( Tensor Core)☆146Updated 5 years ago
- Framework to reduce autotune overhead to zero for well known deployments.☆90Updated 3 months ago
- A bunch of kernels that might make stuff slower 😉☆65Updated 2 weeks ago
- Several optimization methods of half-precision general matrix vector multiplication (HGEMV) using CUDA core.☆70Updated last year