openvinotoolkit / openvino
OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
☆7,232Updated this week
Related projects ⓘ
Alternatives and complementary repositories for openvino
- ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator☆14,642Updated this week
- Pre-trained Deep Learning models and demos (high quality and extremely fast)☆4,101Updated this week
- Open standard for machine learning interoperability☆17,893Updated this week
- NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source compone…☆10,764Updated this week
- oneAPI Deep Neural Network Library (oneDNN)☆3,619Updated this week
- The Triton Inference Server provides an optimized cloud and edge inferencing solution.☆8,296Updated this week
- Open deep learning compiler stack for cpu, gpu and specialized accelerators☆11,761Updated this week
- MNN is a blazing fast, lightweight deep learning framework, battle-tested by business-critical use cases in Alibaba☆8,712Updated last week
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆939Updated this week
- A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep lear…☆5,139Updated this week
- Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX☆2,320Updated 2 months ago
- Simplify your onnx model☆3,849Updated 2 months ago
- Development repository for the Triton language and compiler☆13,311Updated this week
- PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT☆2,585Updated this week
- ONNX-TensorRT: TensorRT backend for ONNX☆2,948Updated this week
- TNN: developed by Tencent Youtu Lab and Guangying Lab, a uniform deep learning inference framework for mobile、desktop and server. TNN is …☆4,407Updated 2 weeks ago
- Google Brain AutoML☆6,245Updated 7 months ago
- MACE is a deep learning inference framework optimized for mobile heterogeneous computing platforms.☆4,932Updated 4 months ago
- SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX R…☆2,218Updated this week
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆1,865Updated this week
- CV-CUDA™ is an open-source, GPU accelerated library for cloud-scale image processing and computer vision.☆2,373Updated last month
- Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more☆30,426Updated this week
- Repository for OpenCV's extra modules☆9,420Updated this week
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆1,609Updated this week
- a language for fast, portable data-parallel computation☆5,890Updated this week
- Visualizer for neural network, deep learning and machine learning models☆28,075Updated this week
- ncnn is a high-performance neural network inference framework optimized for the mobile platform☆20,436Updated this week
- The Compute Library is a set of computer vision and machine learning functions optimised for both Arm CPUs and GPUs using SIMD technologi…☆2,845Updated last month
- Tensorflow Backend for ONNX☆1,285Updated 7 months ago
- Tutorials for creating and using ONNX models☆3,372Updated 3 months ago