google-ai-edge / LiteRTLinks
LiteRT, successor to TensorFlow Lite. is Google's On-device framework for high-performance ML & GenAI deployment on edge platforms, via efficient conversion, runtime, and optimization
☆1,444Updated this week
Alternatives and similar repositories for LiteRT
Users that are interested in LiteRT are comparing it to the libraries listed below
Sorting:
- Support PyTorch model conversion with LiteRT.☆935Updated this week
- ☆797Updated this week
- The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) a…☆372Updated 2 weeks ago
- Qualcomm® AI Hub Models is our collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) an…☆915Updated 2 weeks ago
- Generative AI extensions for onnxruntime☆957Updated this week
- On-device AI across mobile, embedded and edge for PyTorch☆4,258Updated this week
- TFLite Support is a toolkit that helps users to develop ML and deploy TFLite models onto mobile / ioT devices.☆433Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆433Updated this week
- AI Edge Quantizer: flexible post training quantization for LiteRT models.☆99Updated this week
- ☆178Updated last week
- 🤗 Optimum ExecuTorch☆108Updated last week
- onnxruntime-extensions: A specialized pre- and post- processing library for ONNX Runtime☆441Updated this week
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆2,245Updated last week
- Kernels & AI inference engine for mobile devices.☆4,238Updated last week
- Awesome Mobile LLMs☆301Updated 2 months ago
- This repository is a read-only mirror of https://gitlab.arm.com/kleidi/kleidiai☆113Updated last week
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆422Updated this week
- On-device Speech Recognition for Android☆201Updated 2 weeks ago
- Example apps and demos using PyTorch's ExecuTorch framework☆60Updated this week
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆923Updated this week
- Intel® NPU Acceleration Library☆703Updated 9 months ago
- A modern model graph visualizer and debugger☆1,384Updated this week
- No-code CLI designed for accelerating ONNX workflows☆227Updated 8 months ago
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆532Updated this week
- Olive: Simplify ML Model Finetuning, Conversion, Quantization, and Optimization for CPUs, GPUs and NPUs.☆2,249Updated this week
- ☆181Updated 3 weeks ago
- 🎯An accuracy-first, highly efficient quantization toolkit for LLMs, designed to minimize quality degradation across Weight-Only Quantiza…☆845Updated last week
- Examples for using ONNX Runtime for machine learning inferencing.☆1,605Updated this week
- A Toolkit to Help Optimize Onnx Model☆409Updated last week
- On-device LLM Inference Powered by X-Bit Quantization☆278Updated 2 weeks ago