openvinotoolkit / openvinoLinks
OpenVINO™ is an open source toolkit for optimizing and deploying AI inference
☆9,460Updated this week
Alternatives and similar repositories for openvino
Users that are interested in openvino are comparing it to the libraries listed below
Sorting:
- Open standard for machine learning interoperability☆20,114Updated this week
- NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source compone…☆12,546Updated last month
- 📚 Jupyter notebook tutorials for OpenVINO™☆2,997Updated this week
- ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator☆18,909Updated this week
- Pre-trained Deep Learning models and demos (high quality and extremely fast)☆4,340Updated 3 weeks ago
- oneAPI Deep Neural Network Library (oneDNN)☆3,953Updated this week
- Open Machine Learning Compiler Framework☆13,005Updated this week
- The Triton Inference Server provides an optimized cloud and edge inferencing solution.☆10,189Updated last week
- Tutorials for creating and using ONNX models☆3,643Updated last year
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆2,217Updated this week
- ONNX-TensorRT: TensorRT backend for ONNX☆3,175Updated 2 months ago
- A collection of pre-trained, state-of-the-art models in the ONNX format☆9,324Updated 3 months ago
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆2,001Updated 2 weeks ago
- PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT☆2,917Updated this week
- Simplify your onnx model☆4,273Updated this week
- Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX☆2,507Updated 3 months ago
- CV-CUDA™ is an open-source, GPU accelerated library for cloud-scale image processing and computer vision.☆2,630Updated last month
- A machine learning compiler for GPUs, CPUs, and ML accelerators☆3,880Updated this week
- SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, …☆2,561Updated this week
- AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (N…☆4,694Updated 3 weeks ago
- A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep lear…☆5,590Updated this week
- Compiler for Neural Network hardware accelerators☆3,323Updated last year
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆410Updated this week
- CUDA Templates and Python DSLs for High-Performance Linear Algebra☆9,076Updated this week
- Serve, optimize and scale PyTorch models in production☆4,357Updated 5 months ago
- The Compute Library is a set of computer vision and machine learning functions optimised for both Arm CPUs and GPUs using SIMD technologi…☆3,098Updated last week
- Transformer related optimization, including BERT, GPT☆6,378Updated last year
- A retargetable MLIR-based machine learning compiler and runtime toolkit.☆3,547Updated this week
- 🛠A lite C++ AI toolkit: 100+ models with MNN, ORT and TRT, including Det, Seg, Stable-Diffusion, Face-Fusion, etc.🎉☆4,342Updated 3 weeks ago
- TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizat…☆12,588Updated this week