Curt-Park / mnist-fastapi-aio-tritonLinks
Simple example of FastAPI + gRPC AsyncIO + Triton
☆67Updated 3 years ago
Alternatives and similar repositories for mnist-fastapi-aio-triton
Users that are interested in mnist-fastapi-aio-triton are comparing it to the libraries listed below
Sorting:
- Simple example of FastAPI + Celery + Triton for benchmarking☆64Updated 3 years ago
- The Triton backend for TensorRT.☆78Updated last month
- Tiny configuration for Triton Inference Server☆45Updated 9 months ago
- ☆32Updated 2 years ago
- Deploy stable diffusion model with onnx/tenorrt + tritonserver☆127Updated 2 years ago
- The Triton backend that allows running GPU-accelerated data pre-processing pipelines implemented in DALI's python API.☆138Updated last month
- Triton backend for https://github.com/OpenNMT/CTranslate2☆35Updated 2 years ago
- ☆298Updated last week
- The Triton backend for the ONNX Runtime.☆162Updated this week
- Integrating SSE with NVIDIA Triton Inference Server using a Python backend and Zephyr model. There is very less documentation how to use …☆10Updated last year
- ☆33Updated 3 years ago
- Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.☆211Updated 5 months ago
- Converting weights of Pytorch models to ONNX & TensorRT engines☆50Updated 2 years ago
- showing various ways to serve Keras based stable diffusion