openvinotoolkit / openvinoLinks
OpenVINO™ is an open source toolkit for optimizing and deploying AI inference
☆9,365Updated this week
Alternatives and similar repositories for openvino
Users that are interested in openvino are comparing it to the libraries listed below
Sorting:
- Pre-trained Deep Learning models and demos (high quality and extremely fast)☆4,331Updated last week
- The Triton Inference Server provides an optimized cloud and edge inferencing solution.☆10,131Updated this week
- NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source compone…☆12,473Updated last week
- ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator☆18,720Updated this week
- Open standard for machine learning interoperability☆20,031Updated this week
- oneAPI Deep Neural Network Library (oneDNN)☆3,940Updated this week
- Simplify your onnx model☆4,253Updated 3 months ago
- Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX☆2,500Updated 3 months ago
- A scalable inference server for models optimized with OpenVINO™☆804Updated this week
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆1,997Updated last week
- PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT☆2,907Updated this week
- MindSpore is a new open source deep learning training/inference framework that could be used for mobile, edge and cloud scenarios.☆4,651Updated last year
- Tutorials for creating and using ONNX models☆3,636Updated last year
- ONNX-TensorRT: TensorRT backend for ONNX☆3,175Updated last month
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,111Updated this week
- Open Machine Learning Compiler Framework☆12,939Updated this week
- Visualizer for neural network, deep learning and machine learning models☆32,046Updated this week
- SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, …☆2,544Updated this week
- Transformer related optimization, including BERT, GPT☆6,370Updated last year
- A collection of pre-trained, state-of-the-art models in the ONNX format☆9,283Updated 3 months ago
- ncnn is a high-performance neural network inference framework optimized for the mobile platform☆22,438Updated this week
- CV-CUDA™ is an open-source, GPU accelerated library for cloud-scale image processing and computer vision.☆2,622Updated last month
- A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep lear…☆5,577Updated this week
- A C++ standalone library for machine learning☆5,428Updated 2 weeks ago
- TNN: developed by Tencent Youtu Lab and Guangying Lab, a uniform deep learning inference framework for mobile、desktop and server. TNN is …☆4,601Updated 7 months ago
- Serve, optimize and scale PyTorch models in production☆4,358Updated 4 months ago
- A machine learning compiler for GPUs, CPUs, and ML accelerators☆3,819Updated this week
- 🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization…☆3,215Updated 2 weeks ago
- Build and run containers leveraging NVIDIA GPUs☆3,941Updated this week
- An easy to use PyTorch to TensorRT converter☆4,839Updated last year