yuanmu97 / InFiLinks
InFi is a library for building input filters for resource-efficient inference.
☆41Updated 2 years ago
Alternatives and similar repositories for InFi
Users that are interested in InFi are comparing it to the libraries listed below
Sorting:
- ☆213Updated last year
- This is a list of awesome edgeAI inference related papers.☆98Updated 2 years ago
- PipeEdge: Pipeline Parallelism for Large-Scale Model Inference on Heterogeneous Edge Devices☆37Updated last year
- A Portable C Library for Distributed CNN Inference on IoT Edge Clusters☆88Updated 5 years ago
- ☆102Updated last year
- Source code and datasets for Ekya, a system for continuous learning on the edge.☆112Updated 3 years ago
- ☆13Updated 5 years ago
- ☆78Updated 2 years ago
- ☆16Updated 2 years ago
- Autodidactic Neurosurgeon Collaborative Deep Inference for Mobile Edge Intelligence via Online Learning☆42Updated 4 years ago
- ☆57Updated 4 years ago
- A curated list of research in System for Edge Intelligence and Computing(Edge MLSys), including Frameworks, Tools, Repository, etc. Paper…☆32Updated 4 years ago
- ☆117Updated 7 years ago
- ☆47Updated 2 years ago
- a deep learning-driven scheduler for elastic training in deep learning clusters☆31Updated 4 years ago
- Server-driven Video Streaming for Deep Learning Inference☆93Updated 3 years ago
- A DNN inference latency prediction toolkit for accurately modeling and predicting the latency on diverse edge devices.☆360Updated last year
- Adaptive Model Streaming for real-time video inference on edge devices☆41Updated 4 years ago
- PyTorch implementation of the paper: Decomposing Vision Transformers for Collaborative Inference in Edge Devices☆17Updated last year
- FilterForward: Scaling Video Analytics on Constrained Edge Nodes☆28Updated 5 years ago
- A curated list of awesome projects and papers for AI on Mobile/IoT/Edge devices. Everything is continuously updating. Welcome contributio…☆44Updated 2 years ago
- DNN_Partition辅助工具,用于对pytorch模型进行简单的性能分析以及支持模型切分☆14Updated 4 years ago
- INFOCOM 2024: Online Resource Allocation for Edge Intelligence with Colocated Model Retraining and Inference☆34Updated last year
- Deep Compressive Offloading: Speeding Up Neural Network Inference by Trading Edge Computation for Network Latency☆28Updated 4 years ago
- About DNN compression and acceleration on Edge Devices.☆57Updated 4 years ago
- [IEEE Access] "Head Network Distillation: Splitting Distilled Deep Neural Networks for Resource-constrained Edge Computing Systems" and […☆36Updated 2 years ago
- A demo of end-to-end federated learning system.☆69Updated 3 years ago
- LegoDNN: a block-grained scaling tool for mobile vision systems☆51Updated 2 years ago
- A Distributed Camera System for Inference Scheduling and Continuous Learning in Video Analytics☆18Updated 2 years ago
- [ICLR 2018] Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training☆226Updated last year