cubist38 / mlx-openai-serverLinks
A high-performance API server that provides OpenAI-compatible endpoints for MLX models. Developed using Python and powered by the FastAPI framework, it provides an efficient, scalable, and user-friendly solution for running MLX-based vision and language models locally with an OpenAI-compatible interface.
☆104Updated this week
Alternatives and similar repositories for mlx-openai-server
Users that are interested in mlx-openai-server are comparing it to the libraries listed below
Sorting:
- Train Large Language Models on MLX.☆154Updated last month
- MLX-GUI MLX Inference Server for Apple Silicone☆120Updated 3 weeks ago
- FastMLX is a high performance production ready API to host MLX models.☆325Updated 5 months ago
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. I…☆542Updated this week
- ollama like cli tool for MLX models on huggingface (pull, rm, list, show, serve etc.)☆93Updated this week
- MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.☆193Updated 2 weeks ago
- A command-line utility to manage MLX models between your Hugging Face cache and LM Studio.☆62Updated 6 months ago
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆278Updated 2 months ago
- Find the hidden meaning of LLMs☆25Updated last month
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆273Updated 11 months ago
- Guaranteed Structured Output from any Language Model via Hierarchical State Machines☆145Updated 3 months ago
- GenAI & agent toolkit for Apple Silicon Mac, implementing JSON schema-steered structured output (3SO) and tool-calling in Python. For mor…☆129Updated this week
- Distributed Inference for mlx LLm☆93Updated last year
- A simple Jupyter Notebook for learning MLX text-completion fine-tuning!☆120Updated 9 months ago
- chrome & firefox extension to chat with webpages: local llms☆125Updated 8 months ago
- Qwen Image models through MPS☆171Updated this week
- Blazing fast whisper turbo for ASR (speech-to-text) tasks☆214Updated 10 months ago
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆93Updated 2 months ago
- ☆315Updated this week
- Your gateway to both Ollama & Apple MlX models☆143Updated 6 months ago
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆102Updated last month
- ☆96Updated 2 weeks ago
- API Server for Transformer Lab☆72Updated this week
- A flexible, adaptive classification system for dynamic text classification☆436Updated last week
- Enhancing LLMs with LoRA☆128Updated 3 weeks ago
- Explore a simple example of utilizing MLX for RAG application running locally on your Apple Silicon device.☆174Updated last year
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆259Updated 5 months ago
- Fast parallel LLM inference for MLX☆209Updated last year
- An implementation of the CSM(Conversation Speech Model) for Apple Silicon using MLX.☆372Updated 2 weeks ago
- For inferring and serving local LLMs using the MLX framework☆109Updated last year