google-ai-edge / LiteRT-LMLinks
☆290Updated this week
Alternatives and similar repositories for LiteRT-LM
Users that are interested in LiteRT-LM are comparing it to the libraries listed below
Sorting:
- TPI-LLM: Serving 70b-scale LLMs Efficiently on Low-resource Edge Devices☆186Updated 2 months ago
- Train Large Language Models on MLX.☆138Updated this week
- Distributed Inference for mlx LLm☆94Updated last year
- ☆207Updated this week
- Inference, Fine Tuning and many more recipes with Gemma family of models☆262Updated 2 weeks ago
- 1.58 Bit LLM on Apple Silicon using MLX☆217Updated last year
- Fast Streaming TTS with Orpheus + WebRTC (with FastRTC)☆302Updated 3 months ago
- FastMLX is a high performance production ready API to host MLX models.☆319Updated 4 months ago
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆270Updated 10 months ago
- Sparse Inferencing for transformer based LLMs☆196Updated this week
- Fast parallel LLM inference for MLX☆204Updated last year
- Lemonade helps users run local LLMs with the highest performance by configuring state-of-the-art inference engines for their NPUs and GPU…☆350Updated this week
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆97Updated last month
- ☆155Updated 3 months ago
- ☆102Updated 11 months ago
- Official python implementation of the UTCP☆364Updated this week
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆90Updated last month
- SmolVLM2 Demo☆170Updated 4 months ago
- A Tree Search Library with Flexible API for LLM Inference-Time Scaling☆423Updated this week
- ☆102Updated last month
- Gemma 2 optimized for your local machine.☆376Updated 11 months ago
- Verifiers for LLM Reinforcement Learning☆69Updated this week
- Kyutai with an "eye"☆212Updated 4 months ago
- Blazing fast whisper turbo for ASR (speech-to-text) tasks☆213Updated 9 months ago
- Examples on how to use various LLM providers with a Wine Classification problem☆97Updated last month
- Minimal Linux OS with a Model Context Protocol (MCP) gateway to expose local capabilities to LLMs.☆259Updated last month
- A flexible, adaptive classification system for dynamic text classification☆351Updated last week
- MLX-GUI MLX Inference Server☆107Updated last week
- Pipecat voice AI agents running locally on macOS☆88Updated last week
- A simple MLX implementation for pretraining LLMs on Apple Silicon.☆83Updated 3 months ago