AlibabaResearch / mononnLinks
☆32Updated last year
Alternatives and similar repositories for mononn
Users that are interested in mononn are comparing it to the libraries listed below
Sorting:
- MAGIS: Memory Optimization via Coordinated Graph Transformation and Scheduling for DNN (ASPLOS'24)☆56Updated last year
- Magicube is a high-performance library for quantized sparse matrix operations (SpMM and SDDMM) of deep learning on Tensor Cores.☆91Updated 3 years ago
- ☆84Updated 3 years ago
- ☆83Updated 7 months ago
- ASPLOS'24: Optimal Kernel Orchestration for Tensor Programs with Korch☆40Updated 9 months ago
- ☆110Updated last year
- An extention of TVMScript to write simple and high performance GPU kernels with tensorcore.☆51Updated last year
- Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.☆81Updated last week
- Automatic Mapping Generation, Verification, and Exploration for ISA-based Spatial Accelerators☆120Updated 3 years ago
- OSDI 2023 Welder, deeplearning compiler☆31Updated 2 years ago
- Artifacts of EVT ASPLOS'24☆28Updated last year
- A lightweight design for computation-communication overlap.☆209Updated 3 weeks ago
- ☆92Updated 9 months ago
- ☆41Updated 2 months ago
- Horizontal Fusion☆24Updated 4 years ago
- play gemm with tvm☆92Updated 2 years ago
- DietCode Code Release☆65Updated 3 years ago
- ☆17Updated 10 months ago
- Tacker: Tensor-CUDA Core Kernel Fusion for Improving the GPU Utilization while Ensuring QoS☆33Updated 11 months ago
- An Optimizing Compiler for Recommendation Model Inference☆26Updated 7 months ago
- Dissecting NVIDIA GPU Architecture☆116Updated 3 years ago
- ☆165Updated 8 months ago
- NVSHMEM‑Tutorial: Build a DeepEP‑like GPU Buffer☆152Updated 3 months ago
- We invite you to visit and follow our new repository at https://github.com/microsoft/TileFusion. TiledCUDA is a highly efficient kernel …☆191Updated 11 months ago
- ☆18Updated last year
- Open ABI and FFI for Machine Learning Systems☆293Updated this week
- WaferLLM: Large Language Model Inference at Wafer Scale☆83Updated last week
- ☆48Updated last year
- ☆32Updated 3 years ago
- [HPCA 2026] A GPU-optimized system for efficient long-context LLMs decoding with low-bit KV cache.☆76Updated 3 weeks ago