eis-lab / sage
Experimental deep learning framework written in Rust
☆14Updated 2 years ago
Alternatives and similar repositories for sage:
Users that are interested in sage are comparing it to the libraries listed below
- MobiSys#114☆21Updated last year
- ☆24Updated last year
- Source code for the paper: "A Latency-Predictable Multi-Dimensional Optimization Framework forDNN-driven Autonomous Systems"☆22Updated 4 years ago
- SOTA Learning-augmented Systems☆35Updated 2 years ago
- A Cluster-Wide Model Manager to Accelerate DNN Training via Automated Training Warmup☆34Updated 2 years ago
- [ICML 2024 Oral] Any-Precision LLM: Low-Cost Deployment of Multiple, Different-Sized LLMs☆98Updated 2 months ago
- The open-source project for "Mandheling: Mixed-Precision On-Device DNN Training with DSP Offloading"[MobiCom'2022]☆19Updated 2 years ago
- zTT: Learning-based DVFS with Zero Thermal Throttling for Mobile Devices [MobiSys'21] - Artifact Evaluation☆24Updated 3 years ago
- Fast and Efficient Model Serving Using Multi-GPUs with Direct-Host-Access (ACM EuroSys '23)☆57Updated 11 months ago
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆38Updated 2 years ago
- ☆61Updated 2 months ago
- ☆99Updated last year
- PipeTransformer: Automated Elastic Pipelining for Distributed Training of Large-scale Models. ICML 2021☆56Updated 3 years ago
- A curated list of early exiting (LLM, CV, NLP, etc)☆44Updated 7 months ago
- ☆52Updated 11 months ago
- [HPCA'24] Smart-Infinity: Fast Large Language Model Training using Near-Storage Processing on a Real System☆41Updated last year
- ☆19Updated 2 years ago
- LLMServingSim: A HW/SW Co-Simulation Infrastructure for LLM Inference Serving at Scale☆94Updated 3 weeks ago
- This is a list of awesome edgeAI inference related papers.☆95Updated last year
- Multi-Instance-GPU profiling tool☆57Updated last year
- MAGIS: Memory Optimization via Coordinated Graph Transformation and Scheduling for DNN (ASPLOS'24)☆49Updated 9 months ago
- Ok-Topk is a scheme for distributed training with sparse gradients. Ok-Topk integrates a novel sparse allreduce algorithm (less than 6k c…☆25Updated 2 years ago
- [DATE 2023] Pipe-BD: Pipelined Parallel Blockwise Distillation☆11Updated last year
- Official PyTorch Implementation of HELP: Hardware-adaptive Efficient Latency Prediction for NAS via Meta-Learning (NeurIPS 2021 Spotlight…☆62Updated 7 months ago
- [ICML 2024] Serving LLMs on heterogeneous decentralized clusters.☆22Updated 10 months ago
- ☆42Updated 2 years ago
- A version of XRBench-MAESTRO used for MLSys 2023 publication☆23Updated last year
- PyTorch implementation of paper "Response Length Perception and Sequence Scheduling: An LLM-Empowered LLM Inference Pipeline".☆85Updated last year
- ☆14Updated 7 months ago
- ☆92Updated 2 years ago