Accelerate PyTorch models with ONNX Runtime
☆370Feb 5, 2026Updated last month
Alternatives and similar repositories for ort
Users that are interested in ort are comparing it to the libraries listed below
Sorting:
- Examples for using ONNX Runtime for model training.☆364Oct 23, 2024Updated last year
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.☆1,077Apr 17, 2024Updated last year
- High performance model preprocessing library on PyTorch☆648Mar 29, 2024Updated last year
- PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT☆2,958Updated this week
- Smoothly deprecate and redirect Python functions/classes with smart warnings and auto-routing—keep your codebase clean while maintaining …☆55Updated this week
- Simple repository contribution statistics☆15Mar 6, 2026Updated 2 weeks ago
- PyTorch extensions for high performance and large scale training.☆3,403Apr 26, 2025Updated 10 months ago
- A PyTorch repo for data loading and utilities to be shared by the PyTorch domain libraries.☆1,250Mar 14, 2026Updated last week
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆421Updated this week
- functorch is JAX-like composable function transforms for PyTorch.☆1,437Aug 21, 2025Updated 7 months ago
- 🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization…☆3,325Mar 13, 2026Updated last week
- Torch Distributed Experimental☆117Aug 5, 2024Updated last year
- A CPU+GPU Profiling library that provides access to timeline traces and hardware performance counters.☆932Updated this week
- torch::deploy (multipy for non-torch uses) is a system that lets you get around the GIL problem by running multiple Python interpreters i…☆179Dec 16, 2025Updated 3 months ago
- Tutorial on how to convert machine learned models into ONNX☆14Mar 11, 2023Updated 3 years ago
- ☆15Aug 3, 2021Updated 4 years ago
- ONNX Optimizer☆800Mar 2, 2026Updated 2 weeks ago
- Triton Server Component for lightning.ai☆14Feb 15, 2023Updated 3 years ago
- Supplementary components to accelerate research and development in PyTorch☆280Mar 12, 2026Updated last week
- Slicing a PyTorch Tensor Into Parallel Shards☆300Jun 7, 2025Updated 9 months ago
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆432Updated this week
- Olive: Simplify ML Model Finetuning, Conversion, Quantization, and Optimization for CPUs, GPUs and NPUs.☆2,269Updated this week
- Common utilities for ONNX converters☆296Dec 16, 2025Updated 3 months ago
- Serve, optimize and scale PyTorch models in production☆4,362Aug 6, 2025Updated 7 months ago
- Library for 8-bit optimizers and quantization routines.☆780Aug 18, 2022Updated 3 years ago
- Flexible components pairing 🤗 Transformers with Pytorch Lightning☆612Nov 21, 2022Updated 3 years ago
- ONNX Command-Line Toolbox☆36Oct 11, 2024Updated last year
- A GPU performance profiling tool for PyTorch models☆510Jul 13, 2021Updated 4 years ago
- Parallelformers: An Efficient Model Parallelization Toolkit for Deployment☆791Apr 24, 2023Updated 2 years ago
- hydra-pl-wandb-sample-project is a NN experiment management code using hydra, pytorch-lightinig, and wandb.☆11Nov 22, 2021Updated 4 years ago
- AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (N…☆4,709Updated this week
- Type annotations and dynamic checking for a tensor's shape, dtype, names, etc.☆1,475May 2, 2025Updated 10 months ago
- Prototype routines for GPU quantization written using PyTorch.☆21Feb 8, 2026Updated last month
- A collection of Models, Datasets, DataModules, Callbacks, Metrics, Losses and Loggers to better integrate pytorch-lightning with transfor…☆47May 29, 2023Updated 2 years ago
- pytest plugin for a better developer experience when working with the PyTorch test suite☆44Dec 13, 2021Updated 4 years ago
- The Triton backend for the ONNX Runtime.☆173Mar 10, 2026Updated last week
- A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on H…☆3,211Updated this week
- Toolbox of models, callbacks, and datasets for AI/ML researchers.☆1,752Jan 20, 2026Updated 2 months ago
- Prune a model while finetuning or training.☆406Jun 21, 2022Updated 3 years ago