quic / ai-hub-models
The Qualcomm® AI Hub Models are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
☆651Updated this week
Alternatives and similar repositories for ai-hub-models:
Users that are interested in ai-hub-models are comparing it to the libraries listed below
- The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) a…☆167Updated 2 weeks ago
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆507Updated this week
- ☆129Updated 3 weeks ago
- LiteRT is the new name for TensorFlow Lite (TFLite). While the name is new, it's still the same trusted, high-performance runtime for on-…☆330Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆247Updated this week
- Fast Multimodal LLM on Mobile Devices☆781Updated last week
- Generative AI extensions for onnxruntime☆665Updated this week
- On-device AI across mobile, embedded and edge for PyTorch☆2,646Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆454Updated this week
- ☆316Updated last year
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆772Updated this week
- A unified library of state-of-the-art model optimization techniques such as quantization, pruning, distillation, speculative decoding, et…☆821Updated last week
- A simple tutorial of SNPE.☆167Updated 2 years ago
- Strong and Open Vision Language Assistant for Mobile Devices☆1,177Updated 11 months ago
- Low-bit LLM inference on CPU with lookup table☆703Updated 2 months ago
- MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.☆1,269Updated last month
- TinyChatEngine: On-Device LLM Inference Library☆826Updated 8 months ago
- SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX R…☆2,365Updated this week
- The Hailo Model Zoo includes pre-trained models and a full building and evaluation environment☆405Updated last week
- Awesome Mobile LLMs☆156Updated last week
- A pytorch quantization backend for optimum☆907Updated 3 weeks ago
- Examples for using ONNX Runtime for machine learning inferencing.☆1,341Updated 2 months ago
- High-performance, optimized pre-trained template AI application pipelines for systems using Hailo devices☆123Updated this week
- Run generative AI models in sophgo BM1684X☆190Updated this week
- Inference Vision Transformer (ViT) in plain C/C++ with ggml☆264Updated 11 months ago
- ☆694Updated last month
- PyTorch Neural Network eXchange☆564Updated this week
- Conversion of PyTorch Models into TFLite☆370Updated 2 years ago
- Qualcomm Cloud AI SDK (Platform and Apps) enable high performance deep learning inference on Qualcomm Cloud AI platforms delivering high …☆56Updated 5 months ago
- Efficient Inference of Transformer models☆427Updated 7 months ago