triton-inference-server / fil_backendLinks
FIL backend for the Triton Inference Server
☆83Updated 2 weeks ago
Alternatives and similar repositories for fil_backend
Users that are interested in fil_backend are comparing it to the libraries listed below
Sorting:
- Distributed XGBoost on Ray☆152Updated last year
- Merlin Systems provides tools for combining recommendation models with other elements of production recommender systems (like feature sto…☆94Updated last year
- Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.☆213Updated 7 months ago
- The Triton backend for the ONNX Runtime.☆167Updated last week
- Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Serv…☆497Updated this week
- RAPIDS GPU-BDB☆108Updated last year
- The Triton backend for the PyTorch TorchScript models.☆165Updated this week
- Python bindings for UCX☆140Updated 2 months ago
- Provide Python access to the NVML library for GPU diagnostics☆250Updated 2 months ago
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆400Updated this week
- Productionize machine learning predictions, with ONNX or without☆66Updated last year
- LightGBM on Ray☆51Updated last year
- The Triton backend for TensorFlow.☆54Updated 2 weeks ago
- The deepr module provide abstractions (layers, readers, prepro, metrics, config) to help build tensorflow models on top of tf estimators☆53Updated 2 years ago
- Ray - A curated list of resources: https://github.com/ray-project/ray☆72Updated last month
- Spark RAPIDS MLlib – accelerate Apache Spark MLlib with GPUs☆83Updated last week
- Triton backend that enables pre-process, post-processing and other logic to be implemented in Python.☆655Updated last week
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mind…☆161Updated 2 months ago
- Common source, scripts and utilities for creating Triton backends.☆358Updated last week
- ☆56Updated 2 years ago
- The core library and APIs implementing the Triton Inference Server.☆155Updated last week
- RayDP provides simple APIs for running Spark on Ray and integrating Spark with AI libraries.☆354Updated last week
- Introduction to Ray Core Design Patterns and APIs.☆74Updated last year
- ClearML - Model-Serving Orchestration and Repository Solution☆158Updated 2 weeks ago
- ☆78Updated 4 years ago
- Dockerfile templates for creating RAPIDS Docker Images☆83Updated this week
- Plugin for deploying MLflow models to TorchServe☆110Updated 2 years ago
- Home for OctoML PyTorch Profiler☆114Updated 2 years ago
- Python client for RedisAI☆88Updated 2 years ago
- WIP. Veloce is a low-code Ray-based parallelization library that makes machine learning computation novel, efficient, and heterogeneous.☆17Updated 3 years ago