quic / ai-hub-appsLinks
The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
☆232Updated 2 weeks ago
Alternatives and similar repositories for ai-hub-apps
Users that are interested in ai-hub-apps are comparing it to the libraries listed below
Sorting:
- The Qualcomm® AI Hub Models are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.)…☆733Updated 2 weeks ago
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆706Updated this week
- LLM inference in C/C++☆42Updated this week
- LiteRT continues the legacy of TensorFlow Lite as the trusted, high-performance runtime for on-device AI. Now with LiteRT Next, we're exp…☆638Updated last week
- ☆144Updated 3 weeks ago
- Demonstration of running a native LLM on Android device.☆151Updated this week
- A Toolkit to Help Optimize Onnx Model☆174Updated last week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆303Updated this week
- Generative AI extensions for onnxruntime☆753Updated this week
- Fast Multimodal LLM on Mobile Devices☆948Updated last month
- QAI AppBuilder is designed to help developers easily execute models on WoS and Linux platforms. It encapsulates the Qualcomm® AI Runtime …☆55Updated this week
- 本项目是一个通过文字生成图片的项目,基于开源模型Stable Diffusion V1.5生成可以在手机的CPU和NPU上运行的模型,包括其配套的模型运行框架。☆204Updated last year
- Inference RWKV v5, v6 and v7 with Qualcomm AI Engine Direct SDK☆75Updated last week
- llama.cpp tutorial on Android phone☆112Updated 2 months ago
- ☆25Updated this week
- [EMNLP Findings 2024] MobileQuant: Mobile-friendly Quantization for On-device Language Models☆65Updated 9 months ago
- Demonstration of combine YOLO and depth estimation on Android device.☆52Updated 2 months ago
- Run Chinese MobileBert model on SNPE.☆15Updated 2 years ago
- 基于MNN-llm的安卓手机部署大语言模型:Qwen1.5-0.5B-Chat☆80Updated last year
- Port of Facebook's LLaMA model in C/C++☆97Updated last week
- ☆36Updated 3 months ago
- ☆248Updated last week
- IRIS is an android app for interfacing with GGUF / llama.cpp models locally.☆223Updated 5 months ago
- This repository is a read-only mirror of https://gitlab.arm.com/kleidi/kleidiai☆56Updated last week
- llm-export can export llm model to onnx.☆299Updated 6 months ago
- Low-bit LLM inference on CPU/NPU with lookup table☆823Updated last month
- Run generative AI models in sophgo BM1684X/BM1688☆224Updated last week
- Optimized local inference for LLMs with HuggingFace-like APIs for quantization, vision/language models, multimodal agents, speech, vector…☆292Updated 8 months ago
- onnxruntime-extensions: A specialized pre- and post- processing library for ONNX Runtime☆399Updated this week
- PyTorch Neural Network eXchange☆602Updated last week