mlcommons / mobile_app_open
Mobile App Open
☆54Updated last week
Alternatives and similar repositories for mobile_app_open
Users that are interested in mobile_app_open are comparing it to the libraries listed below
Sorting:
- MLPerf Mobile benchmarks☆10Updated last week
- Count number of parameters / MACs / FLOPS for ONNX models.☆92Updated 6 months ago
- Model compression for ONNX☆92Updated 6 months ago
- This repository is a read-only mirror of https://gitlab.arm.com/kleidi/kleidiai☆37Updated this week
- This library empowers users to seamlessly port pretrained models and checkpoints on the HuggingFace (HF) hub (developed using HF transfor…☆65Updated this week
- ☆69Updated 2 years ago
- ☆40Updated 2 years ago
- An easy way to run, test, benchmark and tune OpenCL kernel files☆23Updated last year
- Simple tool for partial optimization of ONNX. Further optimize some models that cannot be optimized with onnx-optimizer and onnxsim by se…☆19Updated last year
- This repository contains the results and code for the MLPerf™ Inference v2.1 benchmark.☆18Updated 2 years ago
- C++ implementations for various tokenizers (sentencepiece, tiktoken etc).☆22Updated this week
- torch::deploy (multipy for non-torch uses) is a system that lets you get around the GIL problem by running multiple Python interpreters i…☆177Updated 5 months ago
- MegEngine到其他框架的转换器☆69Updated 2 years ago
- A tool convert TensorRT engine/plan to a fake onnx☆39Updated 2 years ago
- Snapdragon Neural Processing Engine (SNPE) SDKThe Snapdragon Neural Processing Engine (SNPE) is a Qualcomm Snapdragon software accelerate…☆34Updated 3 years ago
- ☆146Updated 2 years ago
- ☆37Updated this week
- Common utilities for ONNX converters☆269Updated 5 months ago
- A Toolkit to Help Optimize Large Onnx Model☆156Updated last year
- How to export PyTorch models with unsupported layers to ONNX and then to Intel OpenVINO☆27Updated 3 months ago
- AI Edge Quantizer: flexible post training quantization for LiteRT models.☆32Updated this week
- ☆106Updated last month
- edge/mobile transformer based Vision DNN inference benchmark☆16Updated 4 months ago
- IntLLaMA: A fast and light quantization solution for LLaMA☆18Updated last year
- New operators for the ReferenceEvaluator, new kernels for onnxruntime, CPU, CUDA☆32Updated 2 months ago
- A very simple tool for situations where optimization with onnx-simplifier would exceed the Protocol Buffers upper file size limit of 2GB,…☆16Updated last year
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆148Updated last week
- Customized matrix multiplication kernels☆54Updated 3 years ago
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆34Updated 3 years ago
- Convert tflite to JSON and make it editable in the IDE. It also converts the edited JSON back to tflite binary.☆27Updated 2 years ago