triton-inference-server / developer_toolsLinks
☆18Updated 2 weeks ago
Alternatives and similar repositories for developer_tools
Users that are interested in developer_tools are comparing it to the libraries listed below
Sorting:
- The Triton backend that allows running GPU-accelerated data pre-processing pipelines implemented in DALI's python API.☆135Updated 3 weeks ago
- The Triton backend for TensorRT.☆77Updated last week
- The Triton backend for the ONNX Runtime.☆153Updated last week
- Model compression for ONNX☆96Updated 7 months ago
- Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.☆205Updated 2 months ago
- Common source, scripts and utilities shared across all Triton repositories.☆74Updated last week
- Triton CLI is an open source command line interface that enables users to create, deploy, and profile models served by the Triton Inferen…☆64Updated 2 weeks ago
- This repository provides optical character detection and recognition solution optimized on Nvidia devices.☆76Updated last month
- A nvImageCodec library of GPU- and CPU- accelerated codecs featuring a unified interface☆109Updated 3 months ago
- The Triton backend for TensorFlow.☆52Updated last week
- C++ implementations for various tokenizers (sentencepiece, tiktoken etc).☆31Updated this week
- FIL backend for the Triton Inference Server☆81Updated 2 weeks ago
- WIP. Veloce is a low-code Ray-based parallelization library that makes machine learning computation novel, efficient, and heterogeneous.☆18Updated 2 years ago
- ☆67Updated 2 years ago
- OpenVINO backend for Triton.☆32Updated last week
- ☆35Updated last year
- ☆31Updated 2 years ago
- A faster implementation of OpenCV-CUDA that uses OpenCV objects, and more!☆51Updated this week
- RidgeRun Inference Framework☆27Updated 2 years ago
- Count number of parameters / MACs / FLOPS for ONNX models.☆93Updated 8 months ago
- An easy way to run, test, benchmark and tune OpenCL kernel files☆23Updated last year
- ONNX Command-Line Toolbox☆35Updated 8 months ago
- New operators for the ReferenceEvaluator, new kernels for onnxruntime, CPU, CUDA☆32Updated 3 months ago
- Demonstration of the use of TensorRT and TRITON☆16Updated 4 years ago
- Fairring (FAIR + Herring) is a plug-in for PyTorch that provides a process group for distributed training that outperforms NCCL at large …☆65Updated 3 years ago
- Common utilities for ONNX converters☆272Updated 6 months ago
- TensorFlow and TVM integration☆37Updated 5 years ago
- Common source, scripts and utilities for creating Triton backends.☆328Updated last week
- A Ray-based data loader with per-epoch shuffling and configurable pipelining, for shuffling and loading training data for distributed tra…☆18Updated 2 years ago
- The core library and APIs implementing the Triton Inference Server.☆137Updated last week