google-ai-edge / LiteRT-LMLinks
☆322Updated last week
Alternatives and similar repositories for LiteRT-LM
Users that are interested in LiteRT-LM are comparing it to the libraries listed below
Sorting:
- 1.58 Bit LLM on Apple Silicon using MLX☆223Updated last year
- Train Large Language Models on MLX.☆159Updated last month
- FastMLX is a high performance production ready API to host MLX models.☆331Updated 5 months ago
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆272Updated last year
- ☆619Updated 3 weeks ago
- ☆295Updated last month
- Fast Streaming TTS with Orpheus + WebRTC (with FastRTC)☆308Updated 5 months ago
- Distributed Inference for mlx LLm☆95Updated last year
- Inference, Fine Tuning and many more recipes with Gemma family of models☆266Updated last month
- Fast parallel LLM inference for MLX☆216Updated last year
- An implementation of the CSM(Conversation Speech Model) for Apple Silicon using MLX.☆376Updated last month
- API Server for Transformer Lab☆76Updated last week
- Sparse Inferencing for transformer based LLMs☆197Updated last month
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. I…☆547Updated 2 weeks ago
- A command-line interface tool for serving LLM using vLLM.☆403Updated 3 weeks ago
- Qwen Image models through MPS☆188Updated last week
- ☆140Updated last week
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆886Updated 2 weeks ago
- A flexible, adaptive classification system for dynamic text classification☆442Updated 3 weeks ago
- Gemma 2 optimized for your local machine.☆375Updated last year
- ☆209Updated 7 months ago
- Montelimar - Extract text from anywhere☆83Updated 3 months ago
- Docs for GGUF quantization (unofficial)☆258Updated last month
- MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.☆204Updated last week
- Blazing fast whisper turbo for ASR (speech-to-text) tasks☆215Updated 10 months ago
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆110Updated 2 months ago
- 📋 NotebookMLX - An Open Source version of NotebookLM (Ported NotebookLlama)☆316Updated 6 months ago
- Enhancing LLMs with LoRA☆135Updated last week
- Real Time Speech Transcription with FastRTC ⚡️and Local Whisper 🤗☆680Updated 2 months ago
- Code to accompany the Universal Deep Research paper (https://arxiv.org/abs/2509.00244)☆370Updated 3 weeks ago