microsoft / onnxruntime-training-examplesLinks
Examples for using ONNX Runtime for model training.
☆338Updated 7 months ago
Alternatives and similar repositories for onnxruntime-training-examples
Users that are interested in onnxruntime-training-examples are comparing it to the libraries listed below
Sorting:
- onnxruntime-extensions: A specialized pre- and post- processing library for ONNX Runtime☆388Updated last week
- Common utilities for ONNX converters☆270Updated 5 months ago
- Accelerate PyTorch models with ONNX Runtime☆361Updated 3 months ago
- ONNX Optimizer☆715Updated this week
- The Triton backend for the ONNX Runtime.☆148Updated 2 weeks ago
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆354Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆466Updated this week
- The Triton backend for the PyTorch TorchScript models.☆149Updated 2 weeks ago
- Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Serv…☆476Updated last month
- Dockerfiles and scripts for ONNX container images☆137Updated 2 years ago
- Generative AI extensions for onnxruntime☆722Updated this week
- Transform ONNX model to PyTorch representation☆336Updated 6 months ago
- ONNXMLTools enables conversion of models to ONNX☆1,086Updated this week
- Common source, scripts and utilities for creating Triton backends.☆323Updated 2 weeks ago
- Prune a model while finetuning or training.☆402Updated 2 years ago
- PyTriton is a Flask/FastAPI-like interface that simplifies Triton's deployment in Python environments.☆791Updated 3 months ago
- Convert tf.keras/Keras models to ONNX☆379Updated 3 years ago
- ☆411Updated last year
- Implementation of a Transformer, but completely in Triton☆265Updated 3 years ago
- Examples for using ONNX Runtime for machine learning inferencing.☆1,385Updated last week
- Dynamic batching library for Deep Learning inference. Tutorials for LLM, GPT scenarios.☆97Updated 9 months ago
- The Triton backend that allows running GPU-accelerated data pre-processing pipelines implemented in DALI's python API.☆133Updated last week
- Model compression for ONNX☆95Updated 6 months ago
- Triton backend that enables pre-process, post-processing and other logic to be implemented in Python.☆615Updated 2 weeks ago
- PyTorch RFCs (experimental)☆131Updated this week
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,034Updated this week
- Triton Python, C++ and Java client libraries, and GRPC-generated client examples for go, java and scala.☆623Updated 2 weeks ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆262Updated 7 months ago
- Library for 8-bit optimizers and quantization routines.☆716Updated 2 years ago
- A code generator from ONNX to PyTorch code☆137Updated 2 years ago