openvinotoolkit / openvinoLinks
OpenVINO™ is an open source toolkit for optimizing and deploying AI inference
☆8,786Updated this week
Alternatives and similar repositories for openvino
Users that are interested in openvino are comparing it to the libraries listed below
Sorting:
- Pre-trained Deep Learning models and demos (high quality and extremely fast)☆4,282Updated 3 weeks ago
- NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source compone…☆12,125Updated this week
- ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator☆17,786Updated this week
- 📚 Jupyter notebook tutorials for OpenVINO™☆2,893Updated this week
- oneAPI Deep Neural Network Library (oneDNN)☆3,879Updated this week
- A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep lear…☆5,498Updated this week
- Open deep learning compiler stack for cpu, gpu and specialized accelerators☆12,584Updated this week
- Open standard for machine learning interoperability☆19,554Updated this week
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,075Updated this week
- Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX☆2,469Updated last month
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆1,953Updated last week
- PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT☆2,848Updated this week
- Simplify your onnx model☆4,165Updated 2 weeks ago
- ONNX-TensorRT: TensorRT backend for ONNX☆3,149Updated last month
- SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX R…☆2,491Updated this week
- A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch☆8,795Updated this week
- The Triton Inference Server provides an optimized cloud and edge inferencing solution.☆9,755Updated this week
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆2,105Updated this week
- CV-CUDA™ is an open-source, GPU accelerated library for cloud-scale image processing and computer vision.☆2,578Updated 3 months ago
- Transformer related optimization, including BERT, GPT☆6,295Updated last year
- A library for efficient similarity search and clustering of dense vectors.☆36,965Updated this week
- A scalable inference server for models optimized with OpenVINO™☆757Updated this week
- TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizati…☆11,531Updated this week
- A collection of pre-trained, state-of-the-art models in the ONNX format☆8,968Updated 2 months ago
- Olive: Simplify ML Model Finetuning, Conversion, Quantization, and Optimization for CPUs, GPUs and NPUs.☆2,101Updated this week
- Tutorials for creating and using ONNX models☆3,592Updated last year
- An easy to use PyTorch to TensorRT converter☆4,805Updated last year
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆331Updated this week
- AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.☆2,443Updated this week
- Compiler for Neural Network hardware accelerators☆3,311Updated last year