Intel® NPU Acceleration Library
☆709Apr 24, 2025Updated 11 months ago
Alternatives and similar repositories for intel-npu-acceleration-library
Users that are interested in intel-npu-acceleration-library are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Intel® NPU (Neural Processing Unit) Driver☆402Apr 3, 2026Updated last week
- OpenVINO Intel NPU Compiler☆86Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆486Updated this week
- Library for modelling performance costs of different Neural Network workloads on NPU devices☆34Mar 24, 2026Updated 3 weeks ago
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆2,010Mar 30, 2026Updated 2 weeks ago
- Managed Database hosting by DigitalOcean • AdPostgreSQL, MySQL, MongoDB, Kafka, Valkey, and OpenSearch available. Automatically scale up storage and focus on building your apps.
- OpenAI Triton backend for Intel® GPUs☆241Updated this week
- Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V,…☆8,763Jan 28, 2026Updated 2 months ago
- OpenVINO™ is an open source toolkit for optimizing and deploying AI inference☆10,054Updated this week
- Fork of LLVM to support AMD AIEngine processors☆193Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆565Updated this week
- ☆61Dec 18, 2024Updated last year
- OpenVINO Tokenizers extension☆50Apr 2, 2026Updated last week
- Chisel implementation of Neural Processing Unit for System on the Chip☆27Jan 19, 2026Updated 2 months ago
- oneAPI Level Zero Specification Headers and Loader☆313Updated this week
- GPUs on demand by Runpod - Special Offer Available • AdRun AI, ML, and HPC workloads on powerful cloud GPUs—without limits or wasted spend. Deploy GPUs in under a minute and pay by the second.
- ☆697Updated this week
- ⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Pl…☆2,178Oct 8, 2024Updated last year
- Tenstorrent MLIR compiler☆261Updated this week
- SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, …☆2,612Updated this week
- ☆152Mar 31, 2026Updated 2 weeks ago
- portDNN is a library implementing neural network algorithms written using SYCL☆113May 21, 2024Updated last year
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆801Mar 27, 2026Updated 2 weeks ago
- ☆247Apr 8, 2024Updated 2 years ago
- ☆13May 11, 2023Updated 2 years ago
- Managed Database hosting by DigitalOcean • AdPostgreSQL, MySQL, MongoDB, Kafka, Valkey, and OpenSearch available. Automatically scale up storage and focus on building your apps.
- oneAPI Deep Neural Network Library (oneDNN)☆3,980Updated this week
- Olive: Simplify ML Model Finetuning, Conversion, Quantization, and Optimization for CPUs, GPUs and NPUs.☆2,289Updated this week
- An innovative library for efficient LLM inference via low-bit quantization☆352Aug 30, 2024Updated last year
- A Gradio Web UI for running local LLM on Intel GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) using IPEX-LLM.☆18Apr 5, 2026Updated last week
- 📚 Jupyter notebook tutorials for OpenVINO™☆3,096Updated this week
- ☆25Sep 19, 2025Updated 6 months ago
- ☆20Nov 27, 2025Updated 4 months ago
- AI PC starter app for doing AI image creation, image stylizing, and chatbot on a PC powered by an Intel® Arc™ GPU.☆823Updated this week
- Generative AI extensions for onnxruntime☆1,004Updated this week
- Deploy open-source AI quickly and easily - Bonus Offer • AdRunpod Hub is built for open source. One-click deployment and autoscaling endpoints without provisioning your own infrastructure.
- GIMP AI plugins with OpenVINO Backend☆758Mar 17, 2026Updated 3 weeks ago
- SOTA rounding-based quantization for high-accuracy low-bit LLM inference, seamlessly optimized for CPU/XPU/CUDA, with multi-datatype supp…☆957Updated this week
- An MLIR-based toolchain for AMD AI Engine-enabled devices.☆619Updated this week
- ☆83Updated this week
- ☆55Updated this week
- Development repository for the Triton language and compiler☆18,902Updated this week
- ⚠️DirectML is in maintenance mode ⚠️ DirectML is a high-performance, hardware-accelerated DirectX 12 library for machine learning. Direct…☆2,551Feb 20, 2026Updated last month