AI Accelerator Benchmark focuses on evaluating AI Accelerators from a practical production perspective, including the ease of use and versatility of software and hardware.
☆308Feb 25, 2026Updated 3 weeks ago
Alternatives and similar repositories for xpu-perf
Users that are interested in xpu-perf are comparing it to the libraries listed below
Sorting:
- A model compilation solution for various hardware☆467Aug 20, 2025Updated 7 months ago
- optimized BERT transformer inference on NVIDIA GPU. https://arxiv.org/abs/2210.03052☆477Mar 15, 2024Updated 2 years ago
- ☆23Dec 8, 2022Updated 3 years ago
- Byted PyTorch Distributed for Hyperscale Training of LLMs and RLs☆1,000Mar 3, 2026Updated 2 weeks ago
- Running BERT without Padding☆480Mar 18, 2022Updated 4 years ago
- A torch compile backend for multi-targets☆46Mar 11, 2026Updated last week
- Perplexity GPU Kernels☆566Nov 7, 2025Updated 4 months ago
- DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling☆21Updated this week
- A fast communication-overlapping library for tensor/expert parallelism on GPUs.☆1,273Aug 28, 2025Updated 6 months ago
- An external memory allocator example for PyTorch.☆16Aug 10, 2025Updated 7 months ago
- Benchmark tests supporting the TiledCUDA library.☆18Nov 19, 2024Updated last year
- A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on H…☆3,211Updated this week
- A flexible and efficient deep neural network (DNN) compiler that generates high-performance executable from a DNN model description.☆1,003Sep 19, 2024Updated last year
- HeteroHalide: From Image Processing DSL to Efficient FPGA Acceleration☆15Sep 14, 2020Updated 5 years ago
- A primitive library for neural network☆1,367Nov 24, 2024Updated last year
- A Easy-to-understand TensorOp Matmul Tutorial☆409Mar 5, 2026Updated 2 weeks ago
- A standalone GEMM kernel for fp16 activation and quantized weight, extracted from FasterTransformer☆95Feb 20, 2026Updated last month
- ☆176Aug 9, 2023Updated 2 years ago
- TileFusion is an experimental C++ macro kernel template library that elevates the abstraction level in CUDA C for tile processing.☆106Jun 28, 2025Updated 8 months ago
- Distributed Compiler based on Triton for Parallel Systems☆1,386Mar 11, 2026Updated last week
- FlagGems is an operator library for large language models implemented in the Triton Language.☆917Updated this week
- ⚡️Write HGEMM from scratch using Tensor Cores with WMMA, MMA and CuTe API, Achieve Peak ⚡️ Performance.☆149May 10, 2025Updated 10 months ago
- FlashInfer: Kernel Library for LLM Serving☆5,145Updated this week
- PTX-EMU is a simple emulator for CUDA program.☆38Apr 25, 2025Updated 10 months ago
- study of cutlass☆22Nov 10, 2024Updated last year
- ☆65Apr 26, 2025Updated 10 months ago
- An unofficial cuda assembler, for all generations of SASS, hopefully :)☆573Apr 20, 2023Updated 2 years ago
- NCCL Tests☆1,459Mar 11, 2026Updated last week
- ☆207May 5, 2025Updated 10 months ago
- A list of awesome compiler projects and papers for tensor computation and deep learning.☆2,733Oct 19, 2024Updated last year
- Representation and Reference Lowering of ONNX Models in MLIR Compiler Infrastructure☆983Mar 13, 2026Updated last week
- ☆150Jan 9, 2025Updated last year
- High performance RDMA-based distributed feature collection component for training GNN model on EXTREMELY large graph☆55Jul 3, 2022Updated 3 years ago
- CUDA Kernel Benchmarking Library☆831Updated this week
- MegCC是一个运行时超轻量,高效,移植简单的深度学习模型编译器☆484Oct 23, 2024Updated last year
- A prefill & decode disaggregated LLM serving framework with shared GPU memory and fine-grained compute isolation.☆123Dec 25, 2025Updated 2 months ago
- An MLIR-based compiler from C/C++ to AMD-Xilinx Versal AIE☆17Aug 5, 2022Updated 3 years ago
- Automatically Discovering Fast Parallelization Strategies for Distributed Deep Neural Network Training☆1,864Mar 12, 2026Updated last week
- ☆97Mar 26, 2025Updated 11 months ago