nginyc / rafikiLinks
Rafiki is a distributed system that supports training and deployment of machine learning models using AutoML, built with ease-of-use in mind.
☆35Updated 2 years ago
Alternatives and similar repositories for rafiki
Users that are interested in rafiki are comparing it to the libraries listed below
Sorting:
- A Cluster-Wide Model Manager to Accelerate DNN Training via Automated Training Warmup☆35Updated 2 years ago
- this is the release repository of superneurons☆52Updated 4 years ago
- ☆47Updated 2 years ago
- PipeSwitch: Fast Pipelined Context Switching for Deep Learning Applications☆126Updated 3 years ago
- Model-less Inference Serving☆88Updated last year
- FTPipe and related pipeline model parallelism research.☆41Updated 2 years ago
- ☆16Updated 2 years ago
- PipeTransformer: Automated Elastic Pipelining for Distributed Training of Large-scale Models. ICML 2021☆56Updated 3 years ago
- ☆82Updated last week
- ☆94Updated 2 years ago
- ☆43Updated last year
- Fine-grained GPU sharing primitives☆141Updated 5 years ago
- ☆22Updated 4 years ago
- DISB is a new DNN inference serving benchmark with diverse workloads and models, as well as real-world traces.☆52Updated 10 months ago
- Machine learning on serverless platform☆9Updated 2 years ago
- Deadline-based hyperparameter tuning on RayTune.☆31Updated 5 years ago
- BytePS examples (Vision, NLP, GAN, etc)☆19Updated 2 years ago
- sensAI: ConvNets Decomposition via Class Parallelism for Fast Inference on Live Data☆64Updated 11 months ago
- Repository for SysML19 Artifacts Evaluation☆54Updated 6 years ago
- A basic Docker-based installation of TVM☆11Updated 3 years ago
- ML Input Data Processing as a Service. This repository contains the source code for Cachew (built on top of TensorFlow).☆38Updated 9 months ago
- ddl-benchmarks: Benchmarks for Distributed Deep Learning☆37Updated 5 years ago
- ☆24Updated 2 years ago
- Boost hardware utilization for ML training workloads via Inter-model Horizontal Fusion☆32Updated last year
- Artifact for "Shockwave: Fair and Efficient Cluster Scheduling for Dynamic Adaptation in Machine Learning" [NSDI '23]☆44Updated 2 years ago
- Dynamic resources changes for multi-dimensional parallelism training☆25Updated 7 months ago
- CS294-162; Machine Learning Systems Seminar☆31Updated 2 years ago
- An experimental parallel training platform☆54Updated last year
- An analytical performance modeling tool for deep neural networks.☆89Updated 4 years ago
- ☆14Updated 3 years ago