openvinotoolkit / openvinoLinks
OpenVINO™ is an open source toolkit for optimizing and deploying AI inference
☆9,159Updated this week
Alternatives and similar repositories for openvino
Users that are interested in openvino are comparing it to the libraries listed below
Sorting:
- Pre-trained Deep Learning models and demos (high quality and extremely fast)☆4,311Updated 3 weeks ago
- ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator☆18,324Updated this week
- NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source compone…☆12,347Updated this week
- 📚 Jupyter notebook tutorials for OpenVINO™☆2,936Updated last week
- Simplify your onnx model☆4,222Updated 2 months ago
- Open standard for machine learning interoperability☆19,825Updated this week
- oneAPI Deep Neural Network Library (oneDNN)☆3,913Updated this week
- ⚠️DirectML is in maintenance mode ⚠️ DirectML is a high-performance, hardware-accelerated DirectX 12 library for machine learning. Direct…☆2,524Updated last month
- Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX☆2,490Updated last month
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,098Updated this week
- Visualizer for neural network, deep learning and machine learning models☆31,738Updated this week
- ONNX-TensorRT: TensorRT backend for ONNX☆3,165Updated 2 months ago
- Open deep learning compiler stack for cpu, gpu and specialized accelerators☆12,800Updated this week
- PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT☆2,878Updated this week
- AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (N…☆4,691Updated 2 weeks ago
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆1,988Updated this week
- The Triton Inference Server provides an optimized cloud and edge inferencing solution.☆9,972Updated last week
- Development repository for the Triton language and compiler☆17,467Updated this week
- CV-CUDA™ is an open-source, GPU accelerated library for cloud-scale image processing and computer vision.☆2,598Updated 5 months ago
- SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX R…☆2,521Updated last week
- An easy to use PyTorch to TensorRT converter☆4,826Updated last year
- A scalable inference server for models optimized with OpenVINO™☆788Updated this week
- A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep lear…☆5,546Updated this week
- AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.☆2,492Updated this week
- TNN: developed by Tencent Youtu Lab and Guangying Lab, a uniform deep learning inference framework for mobile、desktop and server. TNN is …☆4,589Updated 6 months ago
- Examples for using ONNX Runtime for machine learning inferencing.☆1,529Updated last week
- Transformer related optimization, including BERT, GPT☆6,344Updated last year
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆2,158Updated this week
- A machine learning compiler for GPUs, CPUs, and ML accelerators☆3,650Updated this week
- Annotate better with CVAT, the industry-leading data engine for machine learning. Used and trusted by teams at any scale, for data of any…☆14,694Updated last week