triton-inference-server / fil_backendLinks
FIL backend for the Triton Inference Server
☆81Updated last month
Alternatives and similar repositories for fil_backend
Users that are interested in fil_backend are comparing it to the libraries listed below
Sorting:
- Merlin Systems provides tools for combining recommendation models with other elements of production recommender systems (like feature sto…☆94Updated last year
- Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.☆206Updated 2 months ago
- Distributed XGBoost on Ray☆149Updated last year
- The Triton backend for the ONNX Runtime.☆155Updated this week
- Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Serv…☆479Updated last month
- Provide Python access to the NVML library for GPU diagnostics☆241Updated 7 months ago
- ☆55Updated last year
- The Triton backend for the PyTorch TorchScript models.☆154Updated 2 weeks ago
- Utilities for Dask and CUDA interactions☆311Updated this week
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆370Updated last week
- The core library and APIs implementing the Triton Inference Server.☆138Updated this week
- RAPIDS GPU-BDB☆108Updated last year
- Home for OctoML PyTorch Profiler☆113Updated 2 years ago
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mind…☆158Updated 3 weeks ago
- Python bindings for UCX☆137Updated last week
- Introduction to Ray Core Design Patterns and APIs.☆71Updated last year
- Productionize machine learning predictions, with ONNX or without☆65Updated last year
- Plugin for deploying MLflow models to TorchServe☆110Updated 2 years ago
- The deepr module provide abstractions (layers, readers, prepro, metrics, config) to help build tensorflow models on top of tf estimators☆52Updated last year
- ClearML - Model-Serving Orchestration and Repository Solution☆151Updated 6 months ago
- Unified specification for defining and executing ML workflows, making reproducibility, consistency, and governance easier across the ML p…☆93Updated last year
- WIP. Veloce is a low-code Ray-based parallelization library that makes machine learning computation novel, efficient, and heterogeneous.☆18Updated 2 years ago
- Ray - A curated list of resources: https://github.com/ray-project/ray☆65Updated 2 weeks ago
- Common source, scripts and utilities for creating Triton backends.☆331Updated this week
- RAPIDS Deployment Documentation☆14Updated last week
- Dockerfile templates for creating RAPIDS Docker Images☆78Updated this week
- MLOps Python Library☆119Updated 3 years ago
- A Ray-based data loader with per-epoch shuffling and configurable pipelining, for shuffling and loading training data for distributed tra…☆18Updated 2 years ago
- ☆78Updated 4 years ago
- Ray-based Apache Beam runner☆42Updated last year