A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.
☆102Jun 29, 2025Updated 8 months ago
Alternatives and similar repositories for MLX-Textgen
Users that are interested in MLX-Textgen are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- For inferring and serving local LLMs using the MLX framework☆112Mar 24, 2024Updated 2 years ago
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. I…☆682Mar 10, 2026Updated 2 weeks ago
- A little file for doing LLM-assisted prompt expansion and image generation using Flux.schnell - complete with prompt history, prompt queu…☆26Aug 16, 2024Updated last year
- FastMLX is a high performance production ready API to host MLX models.☆347Mar 18, 2025Updated last year
- A simple Jupyter Notebook for learning MLX text-completion fine-tuning!☆124Nov 10, 2024Updated last year
- Fast parallel LLM inference for MLX☆249Jul 7, 2024Updated last year
- An example implementation of RLHF (or, more accurately, RLAIF) built on MLX and HuggingFace.☆38Jun 21, 2024Updated last year
- ☆38Mar 12, 2024Updated 2 years ago
- ☆21Oct 9, 2024Updated last year
- Roberta Question Answering using MLX.☆24Feb 22, 2026Updated last month
- This repo maintains a 'cheat sheet' for LLMs that are undertrained on mlx☆27Mar 12, 2026Updated last week
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆286Jun 16, 2025Updated 9 months ago
- MLX Transformers is a library that provides model implementation in MLX. It uses a similar model interface as HuggingFace Transformers an…☆75Nov 19, 2024Updated last year
- Gradio chat interface for FastMLX☆12Sep 22, 2024Updated last year
- 🧠 Retrieval Augmented Generation (RAG) example☆19Feb 19, 2026Updated last month
- Generate train.jsonl and valid.jsonl files to use for fine-tuning Mistral and other LLMs.☆97Feb 5, 2024Updated 2 years ago
- Minimal Claude Code alternative powered by MLX☆46Jan 11, 2026Updated 2 months ago
- A CLI in Rust to generate synthetic data for MLX friendly training☆25Jan 13, 2024Updated 2 years ago
- MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.☆317Mar 14, 2026Updated last week
- A tiny server to run local inference on MLX model in the style of OpenAI☆13Jan 31, 2024Updated 2 years ago
- An OpenAI API compatible LLM inference server based on ExLlamaV2.☆25Feb 9, 2024Updated 2 years ago
- o1lama: Use Ollama with Llama 3.2 3B and other models locally to create reasoning chains that are similar in appearance to OpenAI's o1.☆22Jun 1, 2025Updated 9 months ago
- GenAI & agent toolkit for Apple Silicon Mac, implementing JSON schema-steered structured output (3SO) and tool-calling in Python. For mor…☆132Feb 27, 2026Updated 3 weeks ago
- ☆92Jan 24, 2025Updated last year
- MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.☆2,338Updated this week
- Introduction to MLX for Swift developers☆46Jun 23, 2025Updated 9 months ago
- Scripts to create your own moe models using mlx☆89Feb 26, 2024Updated 2 years ago
- Ultra-minimal autoregressive diffusion model for image generation☆21Dec 26, 2025Updated 2 months ago
- mlx image models for Apple Silicon machines☆93Mar 17, 2026Updated last week
- ☆49Mar 17, 2026Updated last week
- Discord Docsbot, Built on bgent☆11Jun 17, 2024Updated last year
- Examples for using the SiLLM framework for training and running Large Language Models (LLMs) on Apple Silicon☆16May 8, 2025Updated 10 months ago
- CLI tool for text to image generation using the FLUX.1 model.☆67Jun 28, 2025Updated 8 months ago
- On-device Image Generation for Apple Silicon☆701Apr 11, 2025Updated 11 months ago
- MLX native implementations of state-of-the-art generative image models☆1,910Updated this week
- This is a FastAPI based LLM server. Load multiple LLM models (MLX or llama.cpp) simultaneously using multiprocessing.☆17Mar 8, 2026Updated 2 weeks ago
- Transcribe and summarize videos using whisper and llms on apple mlx framework☆79Jan 28, 2024Updated 2 years ago
- Start a server from the MLX library.☆199Jul 26, 2024Updated last year
- The easiest way to run the fastest MLX-based LLMs locally☆319Oct 30, 2024Updated last year