MarvinSt / ray-docker-composeLinks
☆27Updated 2 years ago
Alternatives and similar repositories for ray-docker-compose
Users that are interested in ray-docker-compose are comparing it to the libraries listed below
Sorting:
- Simple dependency injection framework for Python☆21Updated last year
- MLFlow Deployment Plugin for Ray Serve☆46Updated 3 years ago
- Benchmark for machine learning model online serving (LLM, embedding, Stable-Diffusion, Whisper)☆28Updated 2 years ago
- Distributed XGBoost on Ray☆152Updated last year
- Plugin for deploying MLflow models to TorchServe☆110Updated 2 years ago
- FIL backend for the Triton Inference Server☆87Updated this week
- The Triton backend for the PyTorch TorchScript models.☆173Updated this week
- ClearML - Model-Serving Orchestration and Repository Solution☆161Updated last month
- Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.☆217Updated this week
- Dataset registry DVC project☆85Updated last year
- Python client for RedisAI☆89Updated 2 years ago
- Some microbenchmarks and design docs before commencement☆12Updated 5 years ago
- Ray provider for Apache Airflow☆47Updated 2 years ago
- Triton backend for managing the model state tensors automatically in sequence batcher☆17Updated last year
- In-depth code associated with my Medium blog post, "How to Load PyTorch Models 340 Times Faster with Ray"☆28Updated 3 years ago
- MLOps Python Library☆121Updated 3 years ago
- A curated list of awesome Milvus projects and resources.☆34Updated 2 years ago
- ☆17Updated 3 years ago
- BentoML Example Projects 🎨☆142Updated last year
- ☆56Updated last year
- Serverless Python with Ray☆59Updated 3 years ago
- A lightweight wrapper for PyTorch that provides a simple declarative API for context switching between devices, distributed modes, mixed-…☆66Updated 2 years ago
- Intel® End-to-End AI Optimization Kit☆31Updated last year
- Python 3 Bindings for the NVIDIA Management Library☆142Updated last year
- Dynamic batching library for Deep Learning inference. Tutorials for LLM, GPT scenarios.☆106Updated last year
- Distributed ML Optimizer☆35Updated 4 years ago
- Article about deploying machine learning models using grpc, pytorch and asyncio☆30Updated 3 years ago
- Pretrain, finetune and serve LLMs on Intel platforms with Ray☆130Updated 4 months ago
- Scaling Python Machine Learning☆53Updated 2 years ago
- Cortex-compatible model server for Python and TensorFlow☆18Updated 3 years ago