lovelyzzkei / QNN-Android-ServerLinks
Let's use Qualcomm NPU in Android
☆15Updated 10 months ago
Alternatives and similar repositories for QNN-Android-Server
Users that are interested in QNN-Android-Server are comparing it to the libraries listed below
Sorting:
- PyTorch Quantization Aware Training(QAT,量化感知训练)☆42Updated 2 years ago
- Run Chinese MobileBert model on SNPE.☆15Updated 2 years ago
- Demonstration of combine YOLO and depth estimation on Android device.☆62Updated last month
- Try to export the ONNX QDQ model that conforms to the AXERA NPU quantization specification. Currently, only w8a8 is supported.☆11Updated last year
- Inference RWKV v5, v6 and v7 with Qualcomm AI Engine Direct SDK☆89Updated 3 weeks ago
- A yolov7-tiny model inference applied on qualcomm snpe for pedestrian detection with embedded system.☆13Updated last year
- Quick and Self-Contained TensorRT Custom Plugin Implementation and Integration☆74Updated 7 months ago
- A simple tutorial of SNPE.☆182Updated 2 years ago
- EasyNN是一个面向教学而开发的神经网络推理框架,旨在让大家0基础也能自主完成推理框架编写!☆34Updated last year
- ☆170Updated 2 weeks ago
- This is a repository to practice multi-thread programming in C++☆26Updated last year
- QAI AppBuilder is designed to help developers easily execute models on WoS and Linux platforms. It encapsulates the Qualcomm® AI Runtime …☆98Updated this week
- Llama3 Streaming Chat Sample☆22Updated last year
- A Toolkit to Help Optimize Large Onnx Model☆162Updated last month
- snpe tutorial☆10Updated 2 years ago
- Speed up image preprocess with cuda when handle image or tensorrt inference☆80Updated last month
- An onnx-based quantitation tool.☆71Updated last year
- the original reference implementation of a specified llama.cpp backend for Qualcomm Hexagon NPU on Android phone, https://github.com/ggml…☆35Updated 5 months ago
- A set of examples around MegEngine☆31Updated 2 years ago
- base quantization methods including: QAT, PTQ, per_channel, per_tensor, dorefa, lsq, adaround, omse, Histogram, bias_correction.etc☆51Updated 3 years ago
- stable diffusion using mnn☆67Updated 2 years ago
- 该代码与B站上的视频 https://www.bilibili.com/video/BV18L41197Uz/?spm_id_from=333.788&vd_source=eefa4b6e337f16d87d87c2c357db8ca7 相关联。☆71Updated 2 years ago
- 本仓库在OpenVINO推理框架下部署Nanodet检测算法,并重写预处理和后处理部分,具有超高性能!让你在Intel CPU平台上的检测速度起飞! 并基于NNCF和PPQ工具将模型量化(PTQ)至int8精度,推理速度更快!☆16Updated 2 years ago
- trt-hackathon-2022 三等奖方案☆10Updated 2 years ago
- Quantize yolov5 using pytorch_quantization.🚀🚀🚀☆14Updated 2 years ago
- ☆26Updated last year
- ☆30Updated last year
- DETR tensor去除推理过程无用辅助头+fp16部署再次加速+解决转tensorrt 输出全为0问题的新方法。☆12Updated last year
- a simple pipline of int8 quantization based on tensorrt.☆69Updated 3 years ago
- 记录量化LLM中的总结。☆49Updated this week