LiteRT, successor to TensorFlow Lite. is Google's On-device framework for high-performance ML & GenAI deployment on edge platforms, via efficient conversion, runtime, and optimization
☆1,527Mar 6, 2026Updated this week
Alternatives and similar repositories for LiteRT
Users that are interested in LiteRT are comparing it to the libraries listed below
Sorting:
- Support PyTorch model conversion with LiteRT.☆944Feb 28, 2026Updated last week
- ☆249Feb 26, 2026Updated last week
- ☆877Updated this week
- AI Edge Quantizer: flexible post training quantization for LiteRT models.☆102Updated this week
- On-device AI across mobile, embedded and edge for PyTorch☆4,336Updated this week
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆2,267Updated this week
- Infrastructure to enable deployment of ML models to low-power resource-constrained embedded targets (including microcontrollers and digit…☆2,780Feb 27, 2026Updated last week
- A modern model graph visualizer and debugger☆1,394Feb 27, 2026Updated last week
- The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) a…☆382Feb 26, 2026Updated last week
- TFLite Support is a toolkit that helps users to develop ML and deploy TFLite models onto mobile / ioT devices.☆433Feb 23, 2026Updated last week
- ☆18Jul 22, 2025Updated 7 months ago
- Qualcomm® AI Hub Models is our collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) an…☆936Feb 25, 2026Updated last week
- This repository is a read-only mirror of https://gitlab.arm.com/kleidi/kleidiai☆116Feb 27, 2026Updated last week
- ☆2,563Feb 4, 2026Updated last month
- ☆15Dec 4, 2024Updated last year
- ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator☆19,435Updated this week
- A gallery that showcases on-device ML/GenAI use cases and allows people to try and use models locally.☆15,213Feb 26, 2026Updated last week
- Cross-platform, customizable ML solutions for live and streaming media.☆33,988Updated this week
- A retargetable MLIR-based machine learning compiler and runtime toolkit.☆3,621Updated this week
- A machine learning compiler for GPUs, CPUs, and ML accelerators☆4,023Updated this week
- [EMNLP Findings 2024] MobileQuant: Mobile-friendly Quantization for On-device Language Models☆67Sep 22, 2024Updated last year
- Tensor library for machine learning☆14,152Feb 27, 2026Updated last week
- Examples for using ONNX Runtime for machine learning inferencing.☆1,616Feb 24, 2026Updated last week
- This repository is an implementation of converting the YOLOv10 object detection model to LiteRT (.tflite) format and deploy it on Android…☆36Sep 25, 2024Updated last year
- ☆182Feb 16, 2026Updated 2 weeks ago
- Low-latency AI engine for mobile devices & wearables☆4,403Updated this week
- lightweight, standalone C++ inference engine for Google's Gemma models.☆6,742Updated this week
- Generative AI extensions for onnxruntime☆965Updated this week
- On-device Neural Engine☆557Updated this week
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆926Updated this week
- Universal LLM Deployment Engine with ML Compilation☆22,082Updated this week
- Development repository for the Triton language and compiler☆18,501Updated this week
- Fast Multimodal LLM on Mobile Devices☆1,412Feb 27, 2026Updated last week
- TT-NN operator library, and TT-Metalium low level kernel programming model.☆1,367Feb 28, 2026Updated last week
- LLM inference in C/C++☆96,322Updated this week
- ☆46Updated this week
- A VS Code Workspace for developing Zephyr Projects☆11Jun 7, 2023Updated 2 years ago
- Official inference framework for 1-bit LLMs☆28,697Feb 3, 2026Updated last month
- MNN is a blazing fast, lightweight deep learning framework, battle-tested by business-critical use cases in Alibaba. Full multimodal LLM …☆14,276Feb 28, 2026Updated last week