quic / ai-hub-appsLinks
The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
☆306Updated this week
Alternatives and similar repositories for ai-hub-apps
Users that are interested in ai-hub-apps are comparing it to the libraries listed below
Sorting:
- The Qualcomm® AI Hub Models are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.)…☆798Updated last week
- ☆157Updated 3 months ago
- LiteRT continues the legacy of TensorFlow Lite as the trusted, high-performance runtime for on-device AI. Now with LiteRT Next, we're exp…☆845Updated this week
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆793Updated last week
- LLM inference in C/C++☆45Updated this week
- QAI AppBuilder is designed to help developers easily execute models on WoS and Linux platforms. It encapsulates the Qualcomm® AI Runtime …☆73Updated last week
- Fast Multimodal LLM on Mobile Devices☆1,087Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆345Updated this week
- 使用Android手机的CPU推理stable diffusion☆159Updated last year
- Inference RWKV v5, v6 and v7 with Qualcomm AI Engine Direct SDK☆83Updated last week
- ☆38Updated 6 months ago
- workbench for learning and practicing on-device AI technology in real scenario with online-TV on Android phone, powered by ggml(llama.cpp…☆177Updated 3 months ago
- llama.cpp tutorial on Android phone☆132Updated 5 months ago
- Generative AI extensions for onnxruntime☆843Updated this week
- This repository is a read-only mirror of https://gitlab.arm.com/kleidi/kleidiai☆86Updated this week
- ☆388Updated this week
- Low-bit LLM inference on CPU/NPU with lookup table☆862Updated 4 months ago
- A Toolkit to Help Optimize Onnx Model☆220Updated last week
- 本项目是一个通过文字生成图片的项目,基于开源模型Stable Diffusion V1.5生成可以在手机的CPU和NPU上运行的模型,包括其配套的模型运行框架。☆219Updated last year
- the original reference implementation of a specified llama.cpp backend for Qualcomm Hexagon NPU on Android phone, https://github.com/ggml…☆31Updated 2 months ago
- Awesome Mobile LLMs☆246Updated last week
- Demonstration of combine YOLO and depth estimation on Android device.☆57Updated last month
- [EMNLP Findings 2024] MobileQuant: Mobile-friendly Quantization for On-device Language Models☆68Updated last year
- High-speed and easy-use LLM serving framework for local deployment☆122Updated 2 months ago
- ☆152Updated 3 weeks ago
- AI Edge Quantizer: flexible post training quantization for LiteRT models.☆68Updated this week
- IRIS is an android app for interfacing with GGUF / llama.cpp models locally.☆239Updated 8 months ago
- On-device Speech Recognition for Android☆136Updated last week
- Let's use Qualcomm NPU in Android☆13Updated 7 months ago
- A unified library of state-of-the-art model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. …☆1,422Updated this week