microsoft / onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
☆15,493Updated this week
Alternatives and similar repositories for onnxruntime:
Users that are interested in onnxruntime are comparing it to the libraries listed below
- Open standard for machine learning interoperability☆18,369Updated this week
- Open deep learning compiler stack for cpu, gpu and specialized accelerators☆11,990Updated this week
- The Triton Inference Server provides an optimized cloud and edge inferencing solution.☆8,684Updated this week
- Development repository for the Triton language and compiler☆14,294Updated this week
- A collection of pre-trained, state-of-the-art models in the ONNX format☆8,229Updated 9 months ago
- NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source compone…☆11,134Updated last week
- Serve, optimize and scale PyTorch models in production☆4,281Updated this week
- 🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (i…☆8,272Updated this week
- Visualizer for neural network, deep learning and machine learning models☆29,290Updated this week
- Tutorials for creating and using ONNX models☆3,446Updated 6 months ago
- Simplify your onnx model☆3,970Updated 5 months ago
- Ongoing research training transformer models at scale☆11,269Updated this week
- Pretrain, finetune ANY AI model of ANY size on multiple GPUs, TPUs with zero code changes.☆28,927Updated this week
- Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.☆14,363Updated last week
- 🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization…☆2,725Updated this week
- oneAPI Deep Neural Network Library (oneDNN)☆3,716Updated this week
- OpenVINO™ is an open source toolkit for optimizing and deploying AI inference☆7,735Updated this week
- Transformer related optimization, including BERT, GPT☆6,009Updated 10 months ago
- Examples for using ONNX Runtime for machine learning inferencing.☆1,281Updated 2 weeks ago
- 💥 Fast State-of-the-Art Tokenizers optimized for Research and Production☆9,349Updated last week
- Fast and memory-efficient exact attention☆15,355Updated this week
- Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX☆2,364Updated this week
- A toolkit for making real world machine learning and data analysis applications in C++☆13,753Updated this week
- A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Auto…☆13,047Updated this week
- ONNX-TensorRT: TensorRT backend for ONNX☆3,011Updated last week
- ncnn is a high-performance neural network inference framework optimized for the mobile platform☆20,904Updated this week
- A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch☆8,525Updated this week
- A retargetable MLIR-based machine learning compiler and runtime toolkit.☆2,965Updated this week
- 🐍 Geometric Computer Vision Library for Spatial AI☆10,193Updated this week
- 🤗 Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch and FLAX.☆27,402Updated this week