google-ai-edge / LiteRTLinks
LiteRT, successor to TensorFlow Lite. is Google's On-device framework for high-performance ML & GenAI deployment on edge platforms, via efficient conversion, runtime, and optimization
☆1,267Updated last week
Alternatives and similar repositories for LiteRT
Users that are interested in LiteRT are comparing it to the libraries listed below
Sorting:
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆903Updated this week
- ☆728Updated this week
- The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) a…☆355Updated last month
- Qualcomm® AI Hub Models is our collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) an…☆890Updated this week
- On-device AI across mobile, embedded and edge for PyTorch☆4,126Updated this week
- Generative AI extensions for onnxruntime☆930Updated this week
- TFLite Support is a toolkit that helps users to develop ML and deploy TFLite models onto mobile / ioT devices.☆432Updated last week
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆905Updated last week
- onnxruntime-extensions: A specialized pre- and post- processing library for ONNX Runtime☆434Updated 3 weeks ago
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆414Updated this week
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆2,223Updated this week
- Examples for using ONNX Runtime for machine learning inferencing.☆1,584Updated last week
- This repository is a read-only mirror of https://gitlab.arm.com/kleidi/kleidiai☆111Updated 3 weeks ago
- Awesome Mobile LLMs☆290Updated last month
- 🤗 Optimum ExecuTorch☆101Updated last week
- ☆170Updated last week
- Example apps and demos using PyTorch's ExecuTorch framework☆53Updated last week
- AI Edge Quantizer: flexible post training quantization for LiteRT models.☆88Updated this week
- A Toolkit to Help Optimize Onnx Model☆308Updated this week
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆418Updated this week
- ☆177Updated 3 weeks ago
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆528Updated this week
- Intel® NPU Acceleration Library☆701Updated 8 months ago
- No-code CLI designed for accelerating ONNX workflows☆224Updated 7 months ago
- A modern model graph visualizer and debugger☆1,365Updated last week
- On-device Speech Recognition for Android☆196Updated 3 months ago
- TinyChatEngine: On-Device LLM Inference Library☆939Updated last year
- MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.☆1,405Updated 8 months ago
- Inference Vision Transformer (ViT) in plain C/C++ with ggml☆305Updated last year
- Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. Th…☆430Updated last week