Intel® NPU Acceleration Library
☆710Apr 24, 2025Updated last year
Alternatives and similar repositories for intel-npu-acceleration-library
Users that are interested in intel-npu-acceleration-library are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Intel® NPU (Neural Processing Unit) Driver☆412Apr 22, 2026Updated last week
- OpenVINO Intel NPU Compiler☆88Apr 27, 2026Updated last week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆498Updated this week
- Library for modelling performance costs of different Neural Network workloads on NPU devices☆35Mar 24, 2026Updated last month
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆2,011Mar 30, 2026Updated last month
- Deploy on Railway without the complexity - Free Credits Offer • AdConnect your repo and Railway handles the rest with instant previews. Quickly provision container image services, databases, and storage volumes.
- OpenAI Triton backend for Intel® GPUs☆249Updated this week
- OpenVINO LLM Benchmark☆11Dec 7, 2023Updated 2 years ago
- Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V,…☆8,795Jan 28, 2026Updated 3 months ago
- Fork of LLVM to support AMD AIEngine processors☆196Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆580Updated this week
- ☆61Dec 18, 2024Updated last year
- ☆17Apr 27, 2026Updated last week
- OpenVINO Tokenizers extension☆50Updated this week
- oneAPI Level Zero Specification Headers and Loader☆315Updated this week
- Wordpress hosting with auto-scaling - Free Trial Offer • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- ☆701Updated this week
- ⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Pl…☆2,178Oct 8, 2024Updated last year
- Tenstorrent MLIR compiler☆265Updated this week
- SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, …☆2,628Updated this week
- Intel® Graphics Compute Runtime for oneAPI Level Zero and OpenCL™ Driver☆1,377Updated this week
- ☆284Updated this week
- ☆152Apr 23, 2026Updated last week
- portDNN is a library implementing neural network algorithms written using SYCL☆114May 21, 2024Updated last year
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆809Apr 17, 2026Updated 2 weeks ago
- Wordpress hosting with auto-scaling - Free Trial Offer • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- ☆249Apr 8, 2024Updated 2 years ago
- oneAPI Deep Neural Network Library (oneDNN)☆3,985Updated this week
- An innovative library for efficient LLM inference via low-bit quantization☆352Aug 30, 2024Updated last year
- A Gradio Web UI for running local LLM on Intel GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) using IPEX-LLM.☆17Apr 26, 2026Updated last week
- 📚 Jupyter notebook tutorials for OpenVINO™☆3,119Updated this week
- ☆25Sep 19, 2025Updated 7 months ago
- ☆20Nov 27, 2025Updated 5 months ago
- Generative AI extensions for onnxruntime☆1,018Updated this week
- AI PC starter app for doing AI image creation, image stylizing, and chatbot on a PC powered by an Intel® Arc™ GPU.☆850Updated this week
- Deploy on Railway without the complexity - Free Credits Offer • AdConnect your repo and Railway handles the rest with instant previews. Quickly provision container image services, databases, and storage volumes.
- GIMP AI plugins with OpenVINO Backend☆768Mar 17, 2026Updated last month
- An MLIR-based toolchain for AMD AI Engine-enabled devices.☆630Updated this week
- A SOTA quantization algorithm for high-accuracy low-bit LLM inference, seamlessly optimized for CPU/XPU/CUDA, with multi-datatype support…☆1,068Updated this week
- Development repository for the Triton language and compiler☆19,087Updated this week
- ☆89Updated this week
- ⚠️DirectML is in maintenance mode ⚠️ DirectML is a high-performance, hardware-accelerated DirectX 12 library for machine learning. Direct…☆2,551Apr 27, 2026Updated last week
- ☆55Apr 27, 2026Updated last week