slyalin / openvino_devtoolsLinks
Tools for easier OpenVINO development/debugging
☆10Updated 6 months ago
Alternatives and similar repositories for openvino_devtools
Users that are interested in openvino_devtools are comparing it to the libraries listed below
Sorting:
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆428Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆532Updated this week
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,123Updated this week
- OpenVINO Tokenizers extension☆48Updated this week
- OpenAI Triton backend for Intel® GPUs☆226Updated this week
- OpenVINO Intel NPU Compiler☆81Updated last week
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆2,010Updated this week
- ☆152Updated last month
- ☆437Updated 4 months ago
- SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, …☆2,581Updated this week
- Repository for OpenVINO's extra modules☆163Updated this week
- ☆38Updated 2 months ago
- Intel® Tensor Processing Primitives extension for Pytorch*☆18Updated 3 weeks ago
- A parser, editor and profiler tool for ONNX models.☆480Updated 3 months ago
- A scalable inference server for models optimized with OpenVINO™☆823Updated this week
- A unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresse…☆1,964Updated this week
- Common utilities for ONNX converters☆294Updated last month
- SYCL* Templates for Linear Algebra (SYCL*TLA) - SYCL based CUTLASS implementation for Intel GPUs☆66Updated this week
- Experimental projects related to TensorRT☆118Updated last week
- Universal cross-platform tokenizers binding to HF and sentencepiece☆451Updated 2 weeks ago
- Intel® NPU Acceleration Library☆703Updated 9 months ago
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆64Updated 7 months ago
- Intel® AI Reference Models: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors and Inte…☆728Updated this week
- Common source, scripts and utilities for creating Triton backends.☆366Updated last week
- ONNX Optimizer☆795Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆85Updated this week
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆420Updated last week
- cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it☆681Updated last week
- [DEPRECATED] Moved to ROCm/rocm-libraries repo. NOTE: develop branch is maintained as a read-only mirror☆518Updated this week
- llm-export can export llm model to onnx.☆343Updated 3 months ago