Intel® NPU Acceleration Library
☆709Apr 24, 2025Updated 11 months ago
Alternatives and similar repositories for intel-npu-acceleration-library
Users that are interested in intel-npu-acceleration-library are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Intel® NPU (Neural Processing Unit) Driver☆390Mar 12, 2026Updated last week
- OpenVINO Intel NPU Compiler☆84Mar 16, 2026Updated last week
- Library for modelling performance costs of different Neural Network workloads on NPU devices☆34Updated this week
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆2,014Mar 13, 2026Updated last week
- OpenAI Triton backend for Intel® GPUs☆236Updated this week
- OpenVINO LLM Benchmark☆11Dec 7, 2023Updated 2 years ago
- Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V,…☆8,740Jan 28, 2026Updated last month
- OpenVINO™ is an open source toolkit for optimizing and deploying AI inference☆9,917Updated this week
- Fork of LLVM to support AMD AIEngine processors☆190Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆553Updated this week
- ☆60Dec 18, 2024Updated last year
- OpenVINO Tokenizers extension☆49Updated this week
- Chisel implementation of Neural Processing Unit for System on the Chip☆26Jan 19, 2026Updated 2 months ago
- oneAPI Level Zero Specification Headers and Loader☆312Feb 24, 2026Updated last month
- ☆693Updated this week
- ⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Pl…☆2,178Oct 8, 2024Updated last year
- Tenstorrent MLIR compiler☆250Mar 18, 2026Updated last week
- SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, …☆2,598Updated this week
- ☆282Mar 18, 2026Updated last week
- ☆152Mar 16, 2026Updated last week
- portDNN is a library implementing neural network algorithms written using SYCL☆113May 21, 2024Updated last year
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆793Feb 9, 2026Updated last month
- ☆243Apr 8, 2024Updated last year
- ☆13May 11, 2023Updated 2 years ago
- Olive: Simplify ML Model Finetuning, Conversion, Quantization, and Optimization for CPUs, GPUs and NPUs.☆2,269Updated this week
- An innovative library for efficient LLM inference via low-bit quantization☆352Aug 30, 2024Updated last year
- A Gradio Web UI for running local LLM on Intel GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) using IPEX-LLM.☆18Mar 15, 2026Updated last week
- 📚 Jupyter notebook tutorials for OpenVINO™☆3,066Updated this week
- AI PC starter app for doing AI image creation, image stylizing, and chatbot on a PC powered by an Intel® Arc™ GPU.☆780Mar 12, 2026Updated last week
- ☆25Sep 19, 2025Updated 6 months ago
- AI Plugins for Windows on Snapdragon☆31May 9, 2025Updated 10 months ago
- ☆20Nov 27, 2025Updated 3 months ago
- Generative AI extensions for onnxruntime☆985Updated this week
- GIMP AI plugins with OpenVINO Backend☆747Mar 17, 2026Updated last week
- 🎯An accuracy-first, highly efficient quantization toolkit for LLMs, designed to minimize quality degradation across Weight-Only Quantiza…☆914Updated this week
- ☆80Mar 18, 2026Updated last week
- Development repository for the Triton language and compiler☆18,708Updated this week
- ☆54Updated this week
- Displaying Intel NPU usage on Linux.☆28Aug 19, 2025Updated 7 months ago