eth-easl / modynLinks
Modyn is a research-platform for training ML models on growing datasets.
☆50Updated 6 months ago
Alternatives and similar repositories for modyn
Users that are interested in modyn are comparing it to the libraries listed below
Sorting:
- VQPy: An object-oriented approach to modern video analytics☆41Updated last year
- LLM Serving Performance Evaluation Harness☆82Updated 9 months ago
- Stateful LLM Serving☆89Updated 9 months ago
- A resilient distributed training framework☆96Updated last year
- [OSDI'24] Serving LLM-based Applications Efficiently with Semantic Variable☆200Updated last year
- ML Input Data Processing as a Service. This repository contains the source code for Cachew (built on top of TensorFlow).☆40Updated last year
- Model-less Inference Serving☆92Updated 2 years ago
- ☆31Updated 3 years ago
- ☆48Updated last year
- Multi-Instance-GPU profiling tool☆58Updated 2 years ago
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mind…☆161Updated 2 months ago
- SpotServe: Serving Generative Large Language Models on Preemptible Instances☆133Updated last year
- ☆73Updated 11 months ago
- ☆79Updated last month
- ☆69Updated 2 months ago
- A framework for generating realistic LLM serving workloads☆88Updated 2 months ago
- Paper-reading notes for Berkeley OS prelim exam.☆14Updated last year
- A minimal implementation of vllm.☆62Updated last year
- Artifact for "Apparate: Rethinking Early Exits to Tame Latency-Throughput Tensions in ML Serving" [SOSP '24]☆25Updated last year
- Surrogate-based Hyperparameter Tuning System☆27Updated 2 years ago
- ByteCheckpoint: An Unified Checkpointing Library for LFMs☆256Updated this week
- Measure and optimize the energy consumption of your AI applications!☆318Updated 2 weeks ago
- Microsoft Collective Communication Library☆66Updated last year
- fmchisel: Efficient Compression and Training Algorithms for Foundation Models☆76Updated last month
- [NeurIPS 2024] Efficient LLM Scheduling by Learning to Rank☆66Updated last year
- PyTorch implementation of paper "Response Length Perception and Sequence Scheduling: An LLM-Empowered LLM Inference Pipeline".☆94Updated 2 years ago
- A library to analyze PyTorch traces.☆443Updated 3 weeks ago
- Deadline-based hyperparameter tuning on RayTune.☆31Updated 5 years ago
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆43Updated 3 years ago
- [ICLR'25] Fast Inference of MoE Models with CPU-GPU Orchestration☆246Updated last year