LINC-BIT / legodnn
LegoDNN: a block-grained scaling tool for mobile vision systems
☆51Updated last year
Alternatives and similar repositories for legodnn:
Users that are interested in legodnn are comparing it to the libraries listed below
- This is a list of awesome edgeAI inference related papers.☆92Updated last year
- ☆14Updated last year
- Source code and datasets for Ekya, a system for continuous learning on the edge.☆103Updated 2 years ago
- To deploy Transformer models in CV to mobile devices.☆17Updated 3 years ago
- ☆76Updated last year
- InFi is a library for building input filters for resource-efficient inference.☆37Updated last year
- Deep Compressive Offloading: Speeding Up Neural Network Inference by Trading Edge Computation for Network Latency☆26Updated 4 years ago
- ☆195Updated last year
- Autodidactic Neurosurgeon Collaborative Deep Inference for Mobile Edge Intelligence via Online Learning☆40Updated 3 years ago
- A DNN inference latency prediction toolkit for accurately modeling and predicting the latency on diverse edge devices.☆342Updated 6 months ago
- ☆124Updated last year
- About DNN compression and acceleration on Edge Devices.☆55Updated 3 years ago
- ☆56Updated 3 years ago
- Cache design for CNN on mobile☆32Updated 6 years ago
- Measuring and predicting on-device metrics (latency, power, etc.) of machine learning models☆66Updated last year
- Adaptive Model Streaming for real-time video inference on edge devices☆41Updated 3 years ago
- ☆17Updated last year
- This project will realize experiments about BranchyNet partitioning using pytorch framework☆27Updated 4 years ago
- Pytorch-based early exit network inspired by branchynet☆31Updated 2 weeks ago
- [ICLR 2018] Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training☆218Updated 7 months ago
- ☆34Updated 2 years ago
- A curated list of research in System for Edge Intelligence and Computing(Edge MLSys), including Frameworks, Tools, Repository, etc. Paper…☆26Updated 3 years ago
- This is the pytorch implementation for the paper: Generalizable Mixed-Precision Quantization via Attribution Rank Preservation, which is…☆25Updated 3 years ago
- [MobiSys 2020] Fast and Scalable In-memory Deep Multitask Learning via Neural Weight Virtualization☆16Updated 4 years ago
- Server-driven Video Streaming for Deep Learning Inference☆87Updated 2 years ago
- 2-stage pruning to favor distributed inference (local device compute half of the model, upload the feature for further computing on stron…☆23Updated 6 years ago
- Source code for Jellyfish, a soft real-time inference serving system☆12Updated 2 years ago
- Official PyTorch Implementation of HELP: Hardware-adaptive Efficient Latency Prediction for NAS via Meta-Learning (NeurIPS 2021 Spotlight…☆61Updated 6 months ago
- PipeEdge: Pipeline Parallelism for Large-Scale Model Inference on Heterogeneous Edge Devices☆29Updated last year
- MobiSys#114☆21Updated last year