tile-ai / tilelang-ascendLinks
Ascend TileLang adapter
☆146Updated this week
Alternatives and similar repositories for tilelang-ascend
Users that are interested in tilelang-ascend are comparing it to the libraries listed below
Sorting:
- Examples of CUDA implementations by Cutlass CuTe☆249Updated 4 months ago
- ☆152Updated 10 months ago
- ☆111Updated 6 months ago
- ☆144Updated last week
- ☆102Updated last year
- A Easy-to-understand TensorOp Matmul Tutorial☆393Updated last month
- ☆156Updated 10 months ago
- ☆112Updated 7 months ago
- Development repository for the Triton-Linalg conversion☆204Updated 9 months ago
- ☆139Updated last year
- Yinghan's Code Sample☆356Updated 3 years ago
- A benchmark suited especially for deep learning operators☆42Updated 2 years ago
- 使用 cutlass 仓库在 ada 架构上实现 fp8 的 flash attention☆78Updated last year
- Summary of the Specs of Commonly Used GPUs for Training and Inference of LLM☆64Updated 3 months ago
- A lightweight design for computation-communication overlap.☆187Updated last month
- ☆154Updated 6 months ago
- ☆143Updated last year
- High performance Transformer implementation in C++.☆142Updated 10 months ago
- ☆70Updated 10 months ago
- A collection of memory efficient attention operators implemented in the Triton language.☆284Updated last year
- FlagGems is an operator library for large language models implemented in the Triton Language.☆763Updated this week
- Implement Flash Attention using Cute.☆96Updated 11 months ago
- flash attention tutorial written in python, triton, cuda, cutlass☆448Updated 6 months ago
- NVSHMEM‑Tutorial: Build a DeepEP‑like GPU Buffer☆143Updated 2 months ago
- SGLang kernel library for NPU☆73Updated this week
- ☆59Updated 4 months ago
- nnScaler: Compiling DNN models for Parallel Training☆119Updated last month
- [USENIX ATC '24] Accelerating the Training of Large Language Models using Efficient Activation Rematerialization and Optimal Hybrid Paral…☆66Updated last year
- Tile-based language built for AI computation across all scales☆80Updated last week
- Allow torch tensor memory to be released and resumed later☆167Updated last week