LiteRT, successor to TensorFlow Lite. is Google's On-device framework for high-performance ML & GenAI deployment on edge platforms, via efficient conversion, runtime, and optimization
☆2,143Apr 9, 2026Updated last week
Alternatives and similar repositories for LiteRT
Users that are interested in LiteRT are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Support PyTorch model conversion with LiteRT.☆988Apr 3, 2026Updated 2 weeks ago
- ☆270Mar 31, 2026Updated 2 weeks ago
- On-device AI across mobile, embedded and edge for PyTorch☆4,502Updated this week
- ☆3,144Apr 9, 2026Updated last week
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆2,309Updated this week
- GPU virtual machines on DigitalOcean Gradient AI • AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- A modern model graph visualizer and debugger☆1,436Updated this week
- Infrastructure to enable deployment of ML models to low-power resource-constrained embedded targets (including microcontrollers and digit…☆2,853Updated this week
- The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) a…☆396Updated this week
- This repository is a read-only mirror of https://gitlab.arm.com/kleidi/kleidiai☆133Apr 8, 2026Updated last week
- TFLite Support is a toolkit that helps users to develop ML and deploy TFLite models onto mobile / ioT devices.☆437Mar 19, 2026Updated 3 weeks ago
- [EMNLP Findings 2024] MobileQuant: Mobile-friendly Quantization for On-device Language Models☆67Sep 22, 2024Updated last year
- ☆15Dec 4, 2024Updated last year
- Qualcomm® AI Hub Models is our collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) an…☆988Updated this week
- A gallery that showcases on-device ML/GenAI use cases and allows people to try and use models locally.☆20,307Apr 8, 2026Updated last week
- Managed Kubernetes at scale on DigitalOcean • AdDigitalOcean Kubernetes includes the control plane, bandwidth allowance, container registry, automatic updates, and more for free.
- ☆2,642Apr 9, 2026Updated last week
- ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator☆19,864Updated this week
- Cross-platform, customizable ML solutions for live and streaming media.☆34,738Updated this week
- A machine learning compiler for GPUs, CPUs, and ML accelerators☆4,151Updated this week
- A retargetable MLIR-based machine learning compiler and runtime toolkit.☆3,714Updated this week
- ☆18Jul 22, 2025Updated 8 months ago
- On-device Neural Engine☆558Updated this week
- Low-latency AI engine for mobile devices & wearables☆4,640Updated this week
- Tensor library for machine learning☆14,394Apr 9, 2026Updated last week
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- lightweight, standalone C++ inference engine for Google's Gemma models.☆6,846Apr 8, 2026Updated last week
- Fast Multimodal LLM on Mobile Devices☆1,463Mar 29, 2026Updated 2 weeks ago
- Universal LLM Deployment Engine with ML Compilation☆22,414Apr 6, 2026Updated last week
- MNN: A blazing-fast, lightweight inference engine battle-tested by Alibaba, powering high-performance on-device LLMs and Edge AI.☆14,845Updated this week
- Official inference framework for 1-bit LLMs☆38,049Mar 10, 2026Updated last month
- Let's use Qualcomm NPU in Android☆18Feb 18, 2025Updated last year
- LLM inference in C/C++☆103,237Updated this week
- ☆552Apr 9, 2026Updated last week
- A tool for converting ONNX files to LiteRT/TFLite/TensorFlow, PyTorch native code (nn.Module), TorchScript (.pt), state_dict (.pt), Expor…☆947Apr 1, 2026Updated 2 weeks ago
- Managed Database hosting by DigitalOcean • AdPostgreSQL, MySQL, MongoDB, Kafka, Valkey, and OpenSearch available. Automatically scale up storage and focus on building your apps.
- Development repository for the Triton language and compiler☆18,902Updated this week
- A python library for converting Pytorch modules into a circle model that is a lightweight and efficient representation in ONE designed fo…☆16Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆76,536Updated this week
- Examples for using ONNX Runtime for machine learning inferencing.☆1,634Feb 24, 2026Updated last month
- TT-NN operator library, and TT-Metalium low level kernel programming model.☆1,402Updated this week
- Generative AI extensions for onnxruntime☆1,004Updated this week
- Simple tool for partial optimization of ONNX. Further optimize some models that cannot be optimized with onnx-optimizer and onnxsim by se…☆19May 7, 2024Updated last year