nod-ai / techtalksLinks
☆15Updated 2 months ago
Alternatives and similar repositories for techtalks
Users that are interested in techtalks are comparing it to the libraries listed below
Sorting:
- CUDA Matrix Multiplication Optimization☆256Updated last year
- ☆259Updated last year
- Github mirror of trition-lang/triton repo.☆128Updated this week
- extensible collectives library in triton☆95Updated 10 months ago
- ☆189Updated last year
- MLIR-based partitioning system☆164Updated last week
- Collection of kernels written in Triton language☆178Updated 2 weeks ago
- AMD RAD's multi-GPU Triton-based framework for seamless multi-GPU programming☆168Updated this week
- High-speed GEMV kernels, at most 2.7x speedup compared to pytorch baseline.☆127Updated last year
- ☆104Updated last year
- Shared Middle-Layer for Triton Compilation☆326Updated 2 months ago
- ☆175Updated 9 months ago
- A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")☆380Updated this week
- ☆286Updated last week
- This repository contains companion software for the Colfax Research paper "Categorical Foundations for CuTe Layouts".☆103Updated 4 months ago
- Stores documents and resources used by the OpenXLA developer community☆133Updated last year
- Nsight Python is a Python kernel profiling interface based on NVIDIA Nsight Tools☆111Updated this week
- Matrix Multiply-Accumulate with CUDA and WMMA( Tensor Core)☆145Updated 5 years ago
- Fastest kernels written from scratch☆532Updated 4 months ago
- An extension library of WMMA API (Tensor Core API)☆109Updated last year
- A library of GPU kernels for sparse matrix operations.☆283Updated 5 years ago
- Tritonbench is a collection of PyTorch custom operators with example inputs to measure their performance.☆324Updated this week
- An experimental CPU backend for Triton☆174Updated 3 months ago
- SparseTIR: Sparse Tensor Compiler for Deep Learning☆142Updated 2 years ago
- TileFusion is an experimental C++ macro kernel template library that elevates the abstraction level in CUDA C for tile processing.☆106Updated 7 months ago
- Cataloging released Triton kernels.☆292Updated 5 months ago
- High-Performance FP32 GEMM on CUDA devices☆117Updated last year
- ☆222Updated last year
- A curated collection of resources, tutorials, and best practices for learning and mastering NVIDIA CUTLASS☆251Updated 9 months ago
- Fast low-bit matmul kernels in Triton☆427Updated last week