intel / intel-xpu-backend-for-tritonView external linksLinks
OpenAI Triton backend for Intel® GPUs
☆226Feb 9, 2026Updated last week
Alternatives and similar repositories for intel-xpu-backend-for-triton
Users that are interested in intel-xpu-backend-for-triton are comparing it to the libraries listed below
Sorting:
- ☆61Dec 18, 2024Updated last year
- Shared Middle-Layer for Triton Compilation☆326Dec 5, 2025Updated 2 months ago
- SYCL* Templates for Linear Algebra (SYCL*TLA) - SYCL based CUTLASS implementation for Intel GPUs☆66Updated this week
- An experimental CPU backend for Triton (https//github.com/openai/triton)☆49Aug 18, 2025Updated 5 months ago
- An experimental CPU backend for Triton☆175Nov 10, 2025Updated 3 months ago
- Intel® Extension for MLIR. A staging ground for MLIR dialects and tools for Intel devices using the MLIR toolchain.☆147Updated this week
- Development repository for the Triton-Linalg conversion☆214Feb 7, 2025Updated last year
- TPP experimentation on MLIR for linear algebra☆144Feb 2, 2026Updated 2 weeks ago
- ☆75Updated this week
- ☆288Updated this week
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆64Jun 30, 2025Updated 7 months ago
- My study note for mlsys☆14Nov 4, 2024Updated last year
- Profiling Tools Interfaces for GPU (PTI for GPU) is a set of Getting Started Documentation and Tools Library to start performance analysi…☆258Jan 21, 2026Updated 3 weeks ago
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆2,009Updated this week
- ☆282Updated this week
- oneAPI Collective Communications Library (oneCCL)☆254Feb 4, 2026Updated last week
- FlagGems is an operator library for large language models implemented in the Triton Language.☆898Updated this week
- oneAPI Level Zero Specification Headers and Loader☆307Feb 5, 2026Updated last week
- The Torch-MLIR project aims to provide first class support from the PyTorch ecosystem to the MLIR ecosystem.☆1,745Updated this week
- Representation and Reference Lowering of ONNX Models in MLIR Compiler Infrastructure☆973Feb 6, 2026Updated last week
- ☆687Updated this week
- Tritonbench is a collection of PyTorch custom operators with example inputs to measure their performance.☆326Updated this week
- Generate Linux Perf event tables for Apple Silicon☆17Dec 16, 2025Updated 2 months ago
- Intel staging area for llvm.org contribution. Home for Intel LLVM-based projects.☆1,429Updated this week
- jax-triton contains integrations between JAX and OpenAI Triton☆439Feb 9, 2026Updated last week
- Intel® NPU Acceleration Library☆705Apr 24, 2025Updated 9 months ago
- Automatic differentiation for Triton Kernels☆29Aug 12, 2025Updated 6 months ago
- C/C++ frontend for MLIR. Also features polyhedral optimizations, parallel optimizations, and more!☆601Jun 19, 2025Updated 7 months ago
- ☆15Dec 17, 2025Updated last month
- Framework to reduce autotune overhead to zero for well known deployments.☆96Sep 19, 2025Updated 4 months ago
- IREE's PyTorch Frontend, based on Torch Dynamo.☆105Updated this week
- ☆59Feb 5, 2026Updated last week
- TileFusion is an experimental C++ macro kernel template library that elevates the abstraction level in CUDA C for tile processing.☆106Jun 28, 2025Updated 7 months ago
- Intel® Graphics Compute Runtime for oneAPI Level Zero and OpenCL™ Driver☆1,340Updated this week
- TORCH_TRACE parser for PT2☆76Feb 5, 2026Updated last week
- ☆21Mar 3, 2025Updated 11 months ago
- Collection of kernels written in Triton language☆178Jan 27, 2026Updated 2 weeks ago
- Reference models for Intel(R) Gaudi(R) AI Accelerator☆170Jan 8, 2026Updated last month
- An MLIR-based compiler framework bridges DSLs (domain-specific languages) to DSAs (domain-specific architectures).☆695Updated this week