mlcommons / mobile_app_openLinks
Mobile App Open
☆66Updated this week
Alternatives and similar repositories for mobile_app_open
Users that are interested in mobile_app_open are comparing it to the libraries listed below
Sorting:
- Model compression for ONNX☆99Updated last year
- This library empowers users to seamlessly port pretrained models and checkpoints on the HuggingFace (HF) hub (developed using HF transfor…☆85Updated this week
- Count number of parameters / MACs / FLOPS for ONNX models.☆95Updated last year
- A Toolkit to Help Optimize Onnx Model☆288Updated this week
- AI Edge Quantizer: flexible post training quantization for LiteRT models.☆84Updated last week
- C++ implementations for various tokenizers (sentencepiece, tiktoken etc).☆44Updated last week
- A faster implementation of OpenCV-CUDA that uses OpenCV objects, and more!☆54Updated last month
- QAI AppBuilder is designed to help developers easily execute models on WoS and Linux platforms. It encapsulates the Qualcomm® AI Runtime …☆98Updated last week
- An easy way to run, test, benchmark and tune OpenCL kernel files☆24Updated 2 years ago
- Export utility for unconstrained channel pruned models☆72Updated 2 years ago
- A Toolkit to Help Optimize Large Onnx Model☆162Updated 2 months ago
- [EMNLP Findings 2024] MobileQuant: Mobile-friendly Quantization for On-device Language Models☆68Updated last year
- Snapdragon Neural Processing Engine (SNPE) SDKThe Snapdragon Neural Processing Engine (SNPE) is a Qualcomm Snapdragon software accelerate…☆36Updated 3 years ago
- MegEngine到其他框架的转换器☆69Updated 2 years ago
- ☆172Updated last week
- torch::deploy (multipy for non-torch uses) is a system that lets you get around the GIL problem by running multiple Python interpreters i…☆182Updated 2 weeks ago
- Common utilities for ONNX converters☆289Updated 2 weeks ago
- Inference RWKV v5, v6 and v7 with Qualcomm AI Engine Direct SDK☆89Updated 3 weeks ago
- Inference Vision Transformer (ViT) in plain C/C++ with ggml☆304Updated last year
- ☆340Updated 2 years ago
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆168Updated this week
- Large Language Model Onnx Inference Framework☆36Updated last month
- Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. Th…☆428Updated last week
- ☆43Updated 3 years ago
- AMD's graph optimization engine.☆268Updated last week
- Notes and artifacts from the ONNX steering committee☆27Updated last week
- Parse TFLite models (*.tflite) EASILY with Python. Check the API at https://zhenhuaw.me/tflite/docs/☆103Updated 11 months ago
- Customized matrix multiplication kernels☆57Updated 3 years ago
- Utility to test the performance of CoreML models.☆70Updated 5 years ago
- Inference of quantization aware trained networks using TensorRT☆83Updated 2 years ago