intel / intel-npu-acceleration-library
Intel® NPU Acceleration Library
☆667Updated 3 months ago
Alternatives and similar repositories for intel-npu-acceleration-library:
Users that are interested in intel-npu-acceleration-library are comparing it to the libraries listed below
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆260Updated this week
- ☆505Updated 3 weeks ago
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆1,836Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆460Updated this week
- Intel® NPU (Neural Processing Unit) Driver☆244Updated 3 weeks ago
- OpenVINO Intel NPU Compiler☆49Updated this week
- Generative AI extensions for onnxruntime☆693Updated this week
- SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX R…☆2,380Updated this week
- An innovative library for efficient LLM inference via low-bit quantization☆350Updated 7 months ago
- OpenAI Triton backend for Intel® GPUs☆179Updated last week
- Low-bit LLM inference on CPU with lookup table☆735Updated 3 months ago
- BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.☆594Updated 2 months ago
- Local LLM Server with NPU Acceleration☆156Updated last week
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆543Updated this week
- Tools for easier OpenVINO development/debugging☆9Updated last month
- An MLIR-based toolchain for AMD AI Engine-enabled devices.☆363Updated this week
- PyTorch emulation library for Microscaling (MX)-compatible data formats☆217Updated this week
- ⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Pl…☆2,170Updated 6 months ago
- OpenVINO Tokenizers extension☆32Updated this week
- Tenstorrent TT-BUDA Repository☆312Updated 3 weeks ago
- ☆251Updated this week
- ☆415Updated this week
- nvidia-modelopt is a unified library of state-of-the-art model optimization techniques like quantization, pruning, distillation, speculat…☆870Updated this week
- Intel® Extension for TensorFlow*☆336Updated last month
- Repository for OpenVINO's extra modules☆119Updated 2 weeks ago
- HIPIFY: Convert CUDA to Portable C++ Code☆571Updated last week
- LiteRT is the new name for TensorFlow Lite (TFLite). While the name is new, it's still the same trusted, high-performance runtime for on-…☆365Updated this week
- Fork of LLVM to support AMD AIEngine processors☆134Updated this week
- A curated list of OpenVINO based AI projects☆126Updated 4 months ago
- AI Tensor Engine for ROCm☆180Updated this week