pittisl / ElasticTrainerLinks
Code for paper "ElasticTrainer: Speeding Up On-Device Training with Runtime Elastic Tensor Selection" (MobiSys'23)
☆13Updated last year
Alternatives and similar repositories for ElasticTrainer
Users that are interested in ElasticTrainer are comparing it to the libraries listed below
Sorting:
- KVTuner: Sensitivity-Aware Layer-wise Mixed Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference☆15Updated 2 months ago
- ☆32Updated last year
- ☆12Updated last year
- [ICLR 2024] This is the official PyTorch implementation of "QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Mod…☆39Updated last year
- This is a list of awesome edgeAI inference related papers.☆96Updated last year
- ☆61Updated last month
- [Preprint] Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Prunin…☆40Updated 2 years ago
- Experimental deep learning framework written in Rust☆15Updated 2 years ago
- It's All In the Teacher: Zero-Shot Quantization Brought Closer to the Teacher [CVPR 2022 Oral]☆29Updated 2 years ago
- Official implementation for ECCV 2022 paper LIMPQ - "Mixed-Precision Neural Network Quantization via Learned Layer-wise Importance"☆56Updated 2 years ago
- Official implementation of ICML 2024 paper "ExCP: Extreme LLM Checkpoint Compression via Weight-Momentum Joint Shrinking".☆48Updated last year
- [EMNLP 2024] RoLoRA: Fine-tuning Rotated Outlier-free LLMs for Effective Weight-Activation Quantization☆36Updated 9 months ago
- ☆25Updated 3 years ago
- ☆28Updated 11 months ago
- [CVPRW 2021] Dynamic-OFA: Runtime DNN Architecture Switching for Performance Scaling on Heterogeneous Embedded Platforms☆29Updated 2 years ago
- ☆58Updated last year
- ☆21Updated 2 years ago
- [ICML 2024 Oral] Any-Precision LLM: Low-Cost Deployment of Multiple, Different-Sized LLMs☆111Updated last week
- Implementation for the paper: CMoE: Fast Carving of Mixture-of-Experts for Efficient LLM Inference☆22Updated 4 months ago
- To deploy Transformer models in CV to mobile devices.☆18Updated 3 years ago
- Official PyTorch Implementation of HELP: Hardware-adaptive Efficient Latency Prediction for NAS via Meta-Learning (NeurIPS 2021 Spotlight…☆63Updated 11 months ago
- Summary of system papers/frameworks/codes/tools on training or serving large model☆57Updated last year
- ☆22Updated 3 months ago
- [NeurIPS 2024] Search for Efficient LLMs☆14Updated 6 months ago
- ☆11Updated last year
- Official Repo for SparseLLM: Global Pruning of LLMs (NeurIPS 2024)☆64Updated 3 months ago
- [ICLR 2025] Linear Combination of Saved Checkpoints Makes Consistency and Diffusion Models Better☆15Updated 5 months ago
- ☆77Updated 2 months ago
- SQUEEZED ATTENTION: Accelerating Long Prompt LLM Inference☆50Updated 7 months ago
- An external memory allocator example for PyTorch.☆14Updated 3 years ago