LINC-BIT / legodnn
LegoDNN: a block-grained scaling tool for mobile vision systems
☆51Updated last year
Alternatives and similar repositories for legodnn:
Users that are interested in legodnn are comparing it to the libraries listed below
- ☆14Updated last year
- ☆17Updated last year
- Source code and datasets for Ekya, a system for continuous learning on the edge.☆105Updated 3 years ago
- A DNN inference latency prediction toolkit for accurately modeling and predicting the latency on diverse edge devices.☆345Updated 8 months ago
- To deploy Transformer models in CV to mobile devices.☆17Updated 3 years ago
- Deep Compressive Offloading: Speeding Up Neural Network Inference by Trading Edge Computation for Network Latency☆26Updated 4 years ago
- FedNAS: Federated Deep Learning via Neural Architecture Search☆54Updated 3 years ago
- Code for paper "ElasticTrainer: Speeding Up On-Device Training with Runtime Elastic Tensor Selection" (MobiSys'23)☆12Updated last year
- Pytorch-based early exit network inspired by branchynet☆31Updated 2 weeks ago
- ☆199Updated last year
- This project will realize experiments about BranchyNet partitioning using pytorch framework☆27Updated 4 years ago
- About DNN compression and acceleration on Edge Devices.☆56Updated 3 years ago
- InFi is a library for building input filters for resource-efficient inference.☆38Updated last year
- This is the pytorch implementation for the paper: Generalizable Mixed-Precision Quantization via Attribution Rank Preservation, which is…☆25Updated 3 years ago
- This is a list of awesome edgeAI inference related papers.☆95Updated last year
- ☆77Updated last year
- [ICLR 2022] The Unreasonable Effectiveness of Random Pruning: Return of the Most Naive Baseline for Sparse Training by Shiwei Liu, Tianlo…☆73Updated 2 years ago
- Autodidactic Neurosurgeon Collaborative Deep Inference for Mobile Edge Intelligence via Online Learning☆41Updated 3 years ago
- ☆125Updated last year
- ☆25Updated 3 years ago
- [ICLR 2018] Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training☆217Updated 8 months ago
- Adaptive Model Streaming for real-time video inference on edge devices☆41Updated 3 years ago
- Dynamic Channel Pruning: Feature Boosting and Suppression☆17Updated 5 years ago
- Official implementation for paper LIMPQ, "Mixed-Precision Neural Network Quantization via Learned Layer-wise Importance", ECCV 2022☆53Updated 2 years ago
- ☆19Updated 3 years ago
- ☆56Updated 3 years ago
- Code for paper "Real-time Neural Network Inference on Extremely Weak Devices: Agile Offloading with Explainable AI" (MobiCom'22)☆17Updated last year
- [MobiSys 2020] Fast and Scalable In-memory Deep Multitask Learning via Neural Weight Virtualization☆16Updated 4 years ago
- Measuring and predicting on-device metrics (latency, power, etc.) of machine learning models☆66Updated last year
- PipeEdge: Pipeline Parallelism for Large-Scale Model Inference on Heterogeneous Edge Devices☆30Updated last year