A DNN inference latency prediction toolkit for accurately modeling and predicting the latency on diverse edge devices.
☆362Jul 30, 2024Updated last year
Alternatives and similar repositories for nn-Meter
Users that are interested in nn-Meter are comparing it to the libraries listed below
Sorting:
- Official PyTorch Implementation of HELP: Hardware-adaptive Efficient Latency Prediction for NAS via Meta-Learning (NeurIPS 2021 Spotlight…☆64Aug 5, 2024Updated last year
- To deploy Transformer models in CV to mobile devices.☆18Jan 20, 2022Updated 4 years ago
- [ICLR 2021] HW-NAS-Bench: Hardware-Aware Neural Architecture Search Benchmark☆116Apr 18, 2023Updated 2 years ago
- [ICLR 2020] Once for All: Train One Network and Specialize it for Efficient Deployment☆1,943Dec 14, 2023Updated 2 years ago
- A flexible and efficient deep neural network (DNN) compiler that generates high-performance executable from a DNN model description.☆1,003Sep 19, 2024Updated last year
- quantize aware training package for NCNN on pytorch☆68Jul 27, 2021Updated 4 years ago
- Autodidactic Neurosurgeon Collaborative Deep Inference for Mobile Edge Intelligence via Online Learning☆42Aug 14, 2021Updated 4 years ago
- Tengine 管子是用来快速生产 demo 的辅助工具☆12Jul 15, 2021Updated 4 years ago
- Model Quantization Benchmark☆861Apr 20, 2025Updated 11 months ago
- [MLSys 2021] IOS: Inter-Operator Scheduler for CNN Acceleration☆199Apr 27, 2022Updated 3 years ago
- ncnn android benchmark app☆86Aug 10, 2021Updated 4 years ago
- [SIGMETRICS 2022] One Proxy Device Is Enough for Hardware-Aware Neural Architecture Search☆13Nov 3, 2021Updated 4 years ago
- prebuild package for cross compiling riscv☆17Dec 28, 2021Updated 4 years ago
- [CVPR 2020] APQ: Joint Search for Network Architecture, Pruning and Quantization Policy☆160Jun 16, 2020Updated 5 years ago
- (HotMobile'24) Salted Inference: Enhancing Privacy while Maintaining Efficiency of Split Inference in Mobile Computing☆17Jan 22, 2024Updated 2 years ago
- This is a list of awesome edgeAI inference related papers.☆99Dec 21, 2023Updated 2 years ago
- ☆266Oct 30, 2019Updated 6 years ago
- [ICLR 2019] ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware☆1,449Aug 30, 2024Updated last year
- Class Project for 18663 - Implementation of FBNet (Hardware-Aware DNAS)☆34Oct 31, 2019Updated 6 years ago
- PyTorch implementation of the paper: Multi-Agent Collaborative Inference via DNN Decoupling: Intermediate Feature Compression and Edge Le…☆46Oct 26, 2023Updated 2 years ago
- code for "AttentiveNAS Improving Neural Architecture Search via Attentive Sampling"☆105Sep 29, 2021Updated 4 years ago
- code for NASViT☆67Apr 25, 2022Updated 3 years ago
- ☆46Nov 25, 2024Updated last year
- Bolt is a deep learning library with high performance and heterogeneous flexibility.☆956Apr 11, 2025Updated 11 months ago
- TVMScript kernel for deformable attention☆25Dec 15, 2021Updated 4 years ago
- NASLib is a Neural Architecture Search (NAS) library for facilitating NAS research for the community by providing interfaces to several …☆586Nov 11, 2024Updated last year
- ☆10May 16, 2021Updated 4 years ago
- AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.☆2,566Mar 14, 2026Updated last week
- benchmark for embededded-ai deep learning inference engines, such as NCNN / TNN / MNN / TensorFlow Lite etc.☆201Feb 18, 2021Updated 5 years ago
- Mobile vision models and code☆918Feb 11, 2026Updated last month
- ☆25Aug 27, 2021Updated 4 years ago
- ☆23Dec 8, 2022Updated 3 years ago
- AutoML tools chain☆851Feb 15, 2023Updated 3 years ago
- Official PyTorch implementation of "Evolving Search Space for Neural Architecture Search"☆12Aug 18, 2021Updated 4 years ago
- (CVPR 2021, Oral) Dynamic Slimmable Network☆231Dec 31, 2021Updated 4 years ago
- ONNX Optimizer☆800Mar 2, 2026Updated 2 weeks ago
- [ACL'20] HAT: Hardware-Aware Transformers for Efficient Natural Language Processing☆336Jul 14, 2024Updated last year
- Single-Path NAS: Designing Hardware-Efficient ConvNets in less than 4 Hours☆395Dec 14, 2020Updated 5 years ago
- NART = NART is not A RunTime, a deep learning inference framework.☆37Mar 2, 2023Updated 3 years ago