quic / ai-hub-models
The Qualcomm® AI Hub Models are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
☆672Updated 2 weeks ago
Alternatives and similar repositories for ai-hub-models:
Users that are interested in ai-hub-models are comparing it to the libraries listed below
- The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) a…☆180Updated 2 weeks ago
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆543Updated last week
- ☆132Updated last month
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆260Updated this week
- Generative AI extensions for onnxruntime☆693Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆460Updated this week
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆784Updated last week
- LiteRT is the new name for TensorFlow Lite (TFLite). While the name is new, it's still the same trusted, high-performance runtime for on-…☆365Updated this week
- nvidia-modelopt is a unified library of state-of-the-art model optimization techniques like quantization, pruning, distillation, speculat…☆870Updated last week
- AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.☆2,282Updated this week
- ☆321Updated last year
- Fast Multimodal LLM on Mobile Devices☆824Updated last month
- Demonstration of running a native LLM on Android device.☆129Updated last week
- Low-bit LLM inference on CPU with lookup table☆735Updated 3 months ago
- On-device AI across mobile, embedded and edge for PyTorch☆2,747Updated this week
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,002Updated this week
- Tools for easier OpenVINO development/debugging☆9Updated last month
- A tool to modify ONNX models in a visualization fashion, based on Netron and Flask.☆1,473Updated 2 months ago
- ☆34Updated 2 weeks ago
- A parser, editor and profiler tool for ONNX models.☆425Updated 3 months ago
- Efficient Inference of Transformer models☆432Updated 8 months ago
- LLaMa/RWKV onnx models, quantization and testcase☆361Updated last year
- A Toolkit to Help Optimize Onnx Model☆140Updated this week
- Repository for OpenVINO's extra modules☆119Updated 2 weeks ago
- Conversion of PyTorch Models into TFLite☆375Updated 2 years ago
- A simple tutorial of SNPE.☆168Updated 2 years ago
- PyTorch Neural Network eXchange☆574Updated last week
- Examples for using ONNX Runtime for machine learning inferencing.☆1,354Updated last week
- A set of simple tools for splitting, merging, OP deletion, size compression, rewriting attributes and constants, OP generation, change op…☆290Updated last year
- Qualcomm Cloud AI SDK (Platform and Apps) enable high performance deep learning inference on Qualcomm Cloud AI platforms delivering high …☆59Updated 6 months ago