Intel® NPU Acceleration Library
☆708Apr 24, 2025Updated 10 months ago
Alternatives and similar repositories for intel-npu-acceleration-library
Users that are interested in intel-npu-acceleration-library are comparing it to the libraries listed below
Sorting:
- Intel® NPU (Neural Processing Unit) Driver☆379Feb 19, 2026Updated last week
- OpenVINO Intel NPU Compiler☆83Feb 23, 2026Updated last week
- Library for modelling performance costs of different Neural Network workloads on NPU devices☆34Feb 11, 2026Updated 3 weeks ago
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆449Updated this week
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆2,012Feb 13, 2026Updated 2 weeks ago
- OpenAI Triton backend for Intel® GPUs☆230Updated this week
- OpenVINO LLM Benchmark☆11Dec 7, 2023Updated 2 years ago
- Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V,…☆8,705Jan 28, 2026Updated last month
- OpenVINO™ is an open source toolkit for optimizing and deploying AI inference☆9,790Updated this week
- ☆17Feb 3, 2026Updated last month
- Fork of LLVM to support AMD AIEngine processors☆189Updated this week
- ☆61Dec 18, 2024Updated last year
- ☆20Nov 27, 2025Updated 3 months ago
- Tenstorrent MLIR compiler☆249Updated this week
- ☆692Updated this week
- SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, …☆2,590Updated this week
- portDNN is a library implementing neural network algorithms written using SYCL☆114May 21, 2024Updated last year
- ☆240Apr 8, 2024Updated last year
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆542Updated this week
- ⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Pl…☆2,175Oct 8, 2024Updated last year
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆779Feb 9, 2026Updated 3 weeks ago
- Intel® Graphics Compute Runtime for oneAPI Level Zero and OpenCL™ Driver☆1,346Updated this week
- Chisel implementation of Neural Processing Unit for System on the Chip☆26Jan 19, 2026Updated last month
- OpenVINO Tokenizers extension☆49Updated this week
- Olive: Simplify ML Model Finetuning, Conversion, Quantization, and Optimization for CPUs, GPUs and NPUs.☆2,255Updated this week
- ☆57Nov 18, 2025Updated 3 months ago
- ☆282Updated this week
- oneAPI Deep Neural Network Library (oneDNN)☆3,958Updated this week
- oneAPI Level Zero Specification Headers and Loader☆311Feb 24, 2026Updated last week
- RISC-V vector and tensor compute extensions for Vortex GPGPU acceleration for ML workloads. Optimized for transformer models, CNNs, and g…☆21Apr 25, 2025Updated 10 months ago
- Matrix multiplication on the NPU inside RK3588☆17Jun 27, 2024Updated last year
- Generate Linux Perf event tables for Apple Silicon☆17Dec 16, 2025Updated 2 months ago
- The Riallto Open Source Project from AMD☆84Apr 10, 2025Updated 10 months ago
- ⚠️DirectML is in maintenance mode ⚠️ DirectML is a high-performance, hardware-accelerated DirectX 12 library for machine learning. Direct…☆2,548Feb 20, 2026Updated last week
- 🎯An accuracy-first, highly efficient quantization toolkit for LLMs, designed to minimize quality degradation across Weight-Only Quantiza…☆853Updated this week
- Development repository for the Triton language and compiler☆18,501Updated this week
- ☆13May 11, 2023Updated 2 years ago
- A suite of tools for pretty printing, diffing, and exploring abstract syntax trees.☆15Updated this week
- Tensor library for machine learning☆14,152Updated this week