eis-lab / sageLinks
Experimental deep learning framework written in Rust
☆15Updated 3 years ago
Alternatives and similar repositories for sage
Users that are interested in sage are comparing it to the libraries listed below
Sorting:
- MobiSys#114☆23Updated 2 years ago
- ☆26Updated 2 years ago
- ☆81Updated 8 months ago
- Source code for the paper: "A Latency-Predictable Multi-Dimensional Optimization Framework forDNN-driven Autonomous Systems"☆22Updated 5 years ago
- This is a list of awesome edgeAI inference related papers.☆98Updated 2 years ago
- ☆21Updated 3 years ago
- [ICML 2024 Oral] Any-Precision LLM: Low-Cost Deployment of Multiple, Different-Sized LLMs☆123Updated 6 months ago
- [ACM EuroSys 2023] Fast and Efficient Model Serving Using Multi-GPUs with Direct-Host-Access☆56Updated 5 months ago
- ☆102Updated 2 years ago
- Study Group of Deep Learning Compiler☆166Updated 3 years ago
- ☆84Updated last year
- This is the implementation for paper: AdaTune: Adaptive Tensor Program CompilationMade Efficient (NeurIPS 2020).☆14Updated 4 years ago
- An external memory allocator example for PyTorch.☆16Updated 5 months ago
- The open-source project for "Mandheling: Mixed-Precision On-Device DNN Training with DSP Offloading"[MobiCom'2022]☆19Updated 3 years ago
- SOTA Learning-augmented Systems☆37Updated 3 years ago
- DietCode Code Release☆65Updated 3 years ago
- Multi-Instance-GPU profiling tool☆58Updated 2 years ago
- SpInfer: Leveraging Low-Level Sparsity for Efficient Large Language Model Inference on GPUs☆62Updated 10 months ago
- ☆164Updated last year
- ☆78Updated 2 years ago
- ☆40Updated last year
- Code for paper "ElasticTrainer: Speeding Up On-Device Training with Runtime Elastic Tensor Selection" (MobiSys'23)☆14Updated 2 years ago
- Chameleon: Adaptive Code Optimization for Expedited Deep Neural Network Compilation☆27Updated 6 years ago
- PET: Optimizing Tensor Programs with Partially Equivalent Transformations and Automated Corrections☆123Updated 3 years ago
- Code for the AAAI 2024 Oral paper "OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Model…☆68Updated last year
- Official PyTorch Implementation of HELP: Hardware-adaptive Efficient Latency Prediction for NAS via Meta-Learning (NeurIPS 2021 Spotlight…☆64Updated last year
- ☆94Updated 3 years ago
- Official Repo for "SplitQuant / LLM-PQ: Resource-Efficient LLM Offline Serving on Heterogeneous GPUs via Phase-Aware Model Partition and …☆36Updated 5 months ago
- LLMServingSim: A HW/SW Co-Simulation Infrastructure for LLM Inference Serving at Scale☆177Updated 6 months ago
- [DATE 2023] Pipe-BD: Pipelined Parallel Blockwise Distillation☆11Updated 2 years ago