quic / ai-hub-apps
The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
☆113Updated last week
Alternatives and similar repositories for ai-hub-apps:
Users that are interested in ai-hub-apps are comparing it to the libraries listed below
- The Qualcomm® AI Hub Models are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.)…☆581Updated this week
- ☆120Updated last month
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆429Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆201Updated this week
- LiteRT is the new name for TensorFlow Lite (TFLite). While the name is new, it's still the same trusted, high-performance runtime for on-…☆253Updated this week
- Demonstration of running a native LLM on Android device.☆106Updated this week
- Generative AI extensions for onnxruntime☆594Updated this week
- onnxruntime-extensions: A specialized pre- and post- processing library for ONNX Runtime☆352Updated this week
- On-device Speech Recognition for Android☆47Updated this week
- A simple tutorial of SNPE.☆153Updated last year
- 本项目是一个通过文字生成图片的项目,基于开源模型Stable Diffusion V1.5生成可以在手机的CPU和NPU上运行的模型,包括其配套的模型运行框架。☆135Updated 10 months ago
- a Android demo of depth_anything_v1 and depth_anything_v2☆54Updated 7 months ago
- workbench for learing&practising AI tech in real scenario on Android device, powered by GGML(Georgi Gerganov Machine Learning) and NCNN(T…☆136Updated this week
- Demonstration of combine YOLO and depth estimation on Android device.☆34Updated this week
- High-performance, optimized pre-trained template AI application pipelines for systems using Hailo devices☆109Updated 3 weeks ago
- ☆312Updated last year
- My develoopment fork of llama.cpp. For now working on RK3588 NPU and Tenstorrent backend☆80Updated last week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆436Updated this week
- Inference RWKV v5, v6 and (WIP) v7 with Qualcomm AI Engine Direct SDK☆49Updated last week
- LLM inference in C/C++☆21Updated this week
- ☆26Updated last week
- A Toolkit to Help Optimize Onnx Model☆106Updated this week
- Optimized local inference for LLMs with HuggingFace-like APIs for quantization, vision/language models, multimodal agents, speech, vector…☆231Updated 3 months ago
- A mobile Implementation of llama.cpp☆299Updated 11 months ago
- A parser, editor and profiler tool for ONNX models.☆414Updated 2 weeks ago
- TensorRT Model Optimizer is a unified library of state-of-the-art model optimization techniques such as quantization, pruning, distillati…☆683Updated this week
- ☆23Updated 2 months ago
- Common utilities for ONNX converters☆257Updated last month
- The no-code AI toolchain☆82Updated this week
- ☆59Updated 3 weeks ago