google-ai-edge / LiteRT-LMLinks
☆509Updated this week
Alternatives and similar repositories for LiteRT-LM
Users that are interested in LiteRT-LM are comparing it to the libraries listed below
Sorting:
- ☆161Updated 2 weeks ago
- LiteRT, successor to TensorFlow Lite. is Google's On-device framework for high-performance ML & GenAI deployment on edge platforms, via e…☆1,035Updated this week
- ☆712Updated last week
- Gemma 2 optimized for your local machine.☆378Updated last year
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆946Updated last week
- Welcome to the official repository of SINQ! A novel, fast and high-quality quantization method designed to make any Large Language Model …☆579Updated 2 weeks ago
- ☆414Updated 3 weeks ago
- Inference, Fine Tuning and many more recipes with Gemma family of models☆275Updated 4 months ago
- Sparse Inferencing for transformer based LLMs☆215Updated 3 months ago
- FastMLX is a high performance production ready API to host MLX models.☆335Updated 8 months ago
- Train Large Language Models on MLX.☆223Updated last week
- ☆301Updated 4 months ago
- Tiny Model, Big Logic: Diversity-Driven Optimization Elicits Large-Model Reasoning Ability in VibeThinker-1.5B☆527Updated 2 weeks ago
- No-code CLI designed for accelerating ONNX workflows☆219Updated 5 months ago
- Fast Streaming TTS with Orpheus + WebRTC (with FastRTC)☆345Updated 7 months ago
- A command-line interface tool for serving LLM using vLLM.☆454Updated this week
- Big & Small LLMs working together☆1,211Updated last week
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆148Updated 5 months ago
- Fast parallel LLM inference for MLX☆234Updated last year
- An implementation of the CSM(Conversation Speech Model) for Apple Silicon using MLX.☆385Updated 3 months ago
- On-device LLM Inference Powered by X-Bit Quantization☆272Updated 2 weeks ago
- API Server for Transformer Lab☆80Updated 2 weeks ago
- Liquid Audio - Speech-to-Speech audio models by Liquid AI☆285Updated 2 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆226Updated last year
- ☆144Updated this week
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆275Updated 3 weeks ago
- A pure MLX-based training pipeline for fine-tuning LLMs using GRPO on Apple Silicon.☆210Updated last month
- Docs for GGUF quantization (unofficial)☆330Updated 4 months ago
- Code to accompany the Universal Deep Research paper (https://arxiv.org/abs/2509.00244)☆450Updated 3 months ago
- Real Time Speech Transcription with FastRTC ⚡️and Local Whisper 🤗☆692Updated 4 months ago