LINC-BIT / legodnnLinks
LegoDNN: a block-grained scaling tool for mobile vision systems
☆50Updated 2 years ago
Alternatives and similar repositories for legodnn
Users that are interested in legodnn are comparing it to the libraries listed below
Sorting:
- A DNN inference latency prediction toolkit for accurately modeling and predicting the latency on diverse edge devices.☆357Updated last year
- ☆205Updated last year
- ☆131Updated last year
- ☆19Updated 3 years ago
- Source code and datasets for Ekya, a system for continuous learning on the edge.☆107Updated 3 years ago
- This is a list of awesome edgeAI inference related papers.☆97Updated last year
- ☆16Updated last year
- Code for paper "ElasticTrainer: Speeding Up On-Device Training with Runtime Elastic Tensor Selection" (MobiSys'23)☆13Updated last year
- ☆13Updated 2 years ago
- To deploy Transformer models in CV to mobile devices.☆18Updated 3 years ago
- [ICLR 2018] Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training☆223Updated last year
- 2021 Summer Research Internship project (UROP) at Imperial College London. Supervised by Prof George Constantinides and Ben Biggs☆17Updated 2 years ago
- InFi is a library for building input filters for resource-efficient inference.☆38Updated last year
- This project will realize experiments about BranchyNet partitioning using pytorch framework☆28Updated 5 years ago
- About DNN compression and acceleration on Edge Devices.☆55Updated 4 years ago
- ☆77Updated 2 years ago
- Pytorch-based early exit network inspired by branchynet☆34Updated 2 months ago
- [MobiSys 2020] Fast and Scalable In-memory Deep Multitask Learning via Neural Weight Virtualization☆15Updated 5 years ago
- ActNN: Reducing Training Memory Footprint via 2-Bit Activation Compressed Training☆200Updated 2 years ago
- [ICLR 2021] HW-NAS-Bench: Hardware-Aware Neural Architecture Search Benchmark☆111Updated 2 years ago
- MobiSys#114☆21Updated last year
- ☆57Updated 3 years ago
- Official implementation for ECCV 2022 paper LIMPQ - "Mixed-Precision Neural Network Quantization via Learned Layer-wise Importance"☆56Updated 2 years ago
- Adaptive Model Streaming for real-time video inference on edge devices☆41Updated 3 years ago
- Deep Compressive Offloading: Speeding Up Neural Network Inference by Trading Edge Computation for Network Latency☆28Updated 4 years ago
- ☆28Updated 2 years ago
- GRACE - GRAdient ComprEssion for distributed deep learning☆139Updated last year
- [NIPS 2016] Learning Structured Sparsity in Deep Neural Networks☆20Updated 3 years ago
- Cache design for CNN on mobile☆34Updated 7 years ago
- [CVPR'20] ZeroQ: A Novel Zero Shot Quantization Framework☆280Updated last year