meta-pytorch / executorch-examplesLinks
Example apps and demos using PyTorch's ExecuTorch framework
☆40Updated 3 weeks ago
Alternatives and similar repositories for executorch-examples
Users that are interested in executorch-examples are comparing it to the libraries listed below
Sorting:
- The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) a…☆345Updated this week
- LiteRT, successor to TensorFlow Lite. is Google's On-device framework for high-performance ML & GenAI deployment on edge platforms, via e…☆1,035Updated this week
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆855Updated this week
- 🤗 Optimum ExecuTorch☆87Updated last week
- On-device Speech Recognition for Android☆160Updated 2 months ago
- Android app for running transformers locally using LLama.cpp & Whisper.cpp☆27Updated last year
- The Qualcomm® AI Hub Models are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.)…☆849Updated 2 weeks ago
- Awesome Mobile LLMs☆276Updated last week
- ☆509Updated this week
- AI Edge Quantizer: flexible post training quantization for LiteRT models.☆81Updated 2 weeks ago
- ☆189Updated 2 weeks ago
- Demonstration of running a native LLM on Android device.☆202Updated this week
- A Toolkit to Help Optimize Onnx Model☆267Updated this week
- Running any GGUF SLMs/LLMs locally, on-device in Android☆588Updated 3 weeks ago
- This repository is a read-only mirror of https://gitlab.arm.com/kleidi/kleidiai☆101Updated last week
- Visualize ONNX models with model-explorer☆64Updated last month
- workbench for learning and practicing on-device AI technology in real scenario with online-TV on Android phone, powered by ggml(llama.cpp…☆178Updated 5 months ago
- Inference RWKV v5, v6 and v7 with Qualcomm AI Engine Direct SDK☆88Updated this week
- Embeddings from sentence-transformers in Android! Supports all-MiniLM-L6-V2, bge-small-en, snowflake-arctic, model2vec models and more☆61Updated 2 months ago
- Optimized OpenAI's Whisper TFLite Port for Efficient Offline Inference on Edge Devices☆266Updated last year
- Advanced quantization toolkit for LLMs and VLMs. Native support for WOQ, MXFP4, NVFP4, GGUF, Adaptive Bits and seamless integration with …☆735Updated this week
- IRIS is an android app for interfacing with GGUF / llama.cpp models locally.☆252Updated 10 months ago
- DFloat11: Lossless LLM Compression for Efficient GPU Inference☆569Updated 2 weeks ago
- Low-bit LLM inference on CPU/NPU with lookup table☆898Updated 6 months ago
- No-code CLI designed for accelerating ONNX workflows☆219Updated 5 months ago
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆381Updated this week
- ☆709Updated last week