stanford-mast / viva-vldb23-artifactLinks
☆12Updated 2 years ago
Alternatives and similar repositories for viva-vldb23-artifact
Users that are interested in viva-vldb23-artifact are comparing it to the libraries listed below
Sorting:
- ML Input Data Processing as a Service. This repository contains the source code for Cachew (built on top of TensorFlow).☆39Updated last year
- ☆47Updated last year
- ☆13Updated last year
- AskIt: Unified programming interface for programming with LLMs (GPT-3.5, GPT-4, Gemini, Claude, Cohere, Llama 2)☆79Updated 8 months ago
- LLM Serving Performance Evaluation Harness☆79Updated 7 months ago
- Tutorial to get started with SkyPilot!☆58Updated last year
- ☆24Updated 2 years ago
- ☆44Updated 4 years ago
- Compression for Foundation Models☆35Updated 2 months ago
- Home for OctoML PyTorch Profiler☆114Updated 2 years ago
- SFS: A Smart OS Scheduler for Serverless Function Workloads (SC'22)☆13Updated 2 years ago
- ☆24Updated 5 months ago
- The source code of INFless,a native serverless platform for AI inference.☆40Updated 2 years ago
- ☆42Updated 2 years ago
- ☆25Updated 2 years ago
- SpotServe: Serving Generative Large Language Models on Preemptible Instances☆129Updated last year
- LLM-Inference-Bench☆52Updated 2 months ago
- Deadline-based hyperparameter tuning on RayTune.☆31Updated 5 years ago
- ☆71Updated last year
- ACT An Architectural Carbon Modeling Tool for Designing Sustainable Computer Systems☆43Updated 2 months ago
- ☆38Updated 4 years ago
- A curated list for Efficient Large Language Models☆11Updated last year
- Artifact for "Apparate: Rethinking Early Exits to Tame Latency-Throughput Tensions in ML Serving" [SOSP '24]☆25Updated 10 months ago
- Model-less Inference Serving☆92Updated last year
- PArametrized Recommendation and Ai Model benchmark is a repository for development of numerous uBenchmarks as well as end to end nets for…☆151Updated 3 weeks ago
- ML model training for edge devices☆166Updated last year
- A benchmark suite for evaluating FaaS scheduler.☆23Updated 2 years ago
- Serverless for all computation☆42Updated 2 years ago
- WIP. Veloce is a low-code Ray-based parallelization library that makes machine learning computation novel, efficient, and heterogeneous.☆18Updated 3 years ago
- Estimating hardware and cloud costs of LLMs and transformer projects☆18Updated 3 months ago