clearml / clearml-fractional-gpuLinks
ClearML Fractional GPU - Run multiple containers on the same GPU with driver level memory limitation ✨ and compute time-slicing
☆82Updated last year
Alternatives and similar repositories for clearml-fractional-gpu
Users that are interested in clearml-fractional-gpu are comparing it to the libraries listed below
Sorting:
- A top-like tool for monitoring GPUs in a cluster☆85Updated last year
 - GPU environment and cluster management with LLM support☆652Updated last year
 - Self-host LLMs with vLLM and BentoML☆153Updated last week
 - Module, Model, and Tensor Serialization/Deserialization☆272Updated 2 months ago
 - ☆262Updated 2 weeks ago
 - IBM development fork of https://github.com/huggingface/text-generation-inference☆61Updated last month
 - Inference server benchmarking tool☆121Updated last month
 - Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.☆213Updated 6 months ago
 - Where GPUs get cooked 👩🍳🔥☆294Updated last month
 - ☆64Updated 7 months ago
 - ☆40Updated this week
 - The backend behind the LLM-Perf Leaderboard☆11Updated last year
 - 🕹️ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.☆139Updated last year
 - The Triton backend for the PyTorch TorchScript models.☆163Updated this week
 - Benchmark suite for LLMs from Fireworks.ai☆83Updated 2 weeks ago
 - vLLM adapter for a TGIS-compatible gRPC server.☆42Updated this week
 - Ray - A curated list of resources: https://github.com/ray-project/ray☆71Updated 2 weeks ago
 - A tool to configure, launch and manage your machine learning experiments.☆203Updated this week
 - Pretrain, finetune and serve LLMs on Intel platforms with Ray☆132Updated last month
 - ⚡️ A fast and flexible PyTorch inference server that runs locally, on any cloud or AI HW.☆145Updated last year
 - A collection of reproducible inference engine benchmarks☆37Updated 6 months ago
 - ☆15Updated last month
 - ClearML - Model-Serving Orchestration and Repository Solution☆156Updated last month
 - Distributed Model Serving Framework☆177Updated last month
 - MLFlow Deployment Plugin for Ray Serve☆46Updated 3 years ago
 - Machine Learning Serving focused on GenAI with simplicity as the top priority.☆58Updated 3 weeks ago
 - Unified storage framework for the entire machine learning lifecycle☆155Updated last year
 - ☆24Updated this week
 - A high-throughput and memory-efficient inference and serving engine for LLMs☆266Updated last year
 - Google TPU optimizations for transformers models☆121Updated 9 months ago