Start a server from the MLX library.
☆200Jul 26, 2024Updated last year
Alternatives and similar repositories for mlxserver
Users that are interested in mlxserver are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- The easiest way to run the fastest MLX-based LLMs locally☆321Oct 30, 2024Updated last year
- mlx image models for Apple Silicon machines☆95Apr 8, 2026Updated last week
- FastMLX is a high performance production ready API to host MLX models.☆352Mar 18, 2025Updated last year
- For inferring and serving local LLMs using the MLX framework☆114Mar 24, 2024Updated 2 years ago
- A simple script to enhance text editing across your Mac, leveraging the power of MLX. Designed for seamless integration, it offers real-t…☆109Mar 4, 2024Updated 2 years ago
- Wordpress hosting with auto-scaling - Free Trial • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- Run large models from the terminal using Apple MLX.☆31Mar 18, 2024Updated 2 years ago
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. I…☆703Mar 10, 2026Updated last month
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆284Jun 16, 2025Updated 10 months ago
- A simple UI / Web / Frontend for MLX mlx-lm using Streamlit.☆262Oct 25, 2025Updated 5 months ago
- Gradio chat interface for FastMLX☆12Sep 22, 2024Updated last year
- A multi-platform SwiftUI frontend for running local LLMs with Apple's MLX framework.☆432Oct 27, 2024Updated last year
- huggingface chat-ui integration with mlx-lm server☆62Feb 13, 2024Updated 2 years ago
- An example implementation of RLHF (or, more accurately, RLAIF) built on MLX and HuggingFace.☆38Jun 21, 2024Updated last year
- Minimal, clean code implementation of RAG with mlx using gguf model weights☆53Apr 27, 2024Updated last year
- Wordpress hosting with auto-scaling - Free Trial • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- ☆44Jun 27, 2025Updated 9 months ago
- Explore a simple example of utilizing MLX for RAG application running locally on your Apple Silicon device.☆180Jan 31, 2024Updated 2 years ago
- An all-in-one LLMs Chat UI for Apple Silicon Mac using MLX Framework.☆1,594Sep 6, 2024Updated last year
- Transcribe and summarize videos using whisper and llms on apple mlx framework☆79Jan 28, 2024Updated 2 years ago
- Chat with MLX is a high-performance macOS application that connects your local documents to a personalized large language model (LLM).☆178Mar 8, 2024Updated 2 years ago
- Examples in the MLX framework☆8,498Apr 6, 2026Updated last week
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.☆461Jan 29, 2025Updated last year
- On-device Image Generation for Apple Silicon☆701Apr 11, 2025Updated last year
- A Next.js chatbot app demonstrating seamless integration with window.ai.☆15Jun 25, 2023Updated 2 years ago
- Wordpress hosting with auto-scaling - Free Trial • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆119May 22, 2024Updated last year
- Code to train Sentence BERT Japanese model for Hugging Face Model Hub☆11Aug 8, 2021Updated 4 years ago
- Implementation of nougat that focuses on processing pdf locally.☆85Jan 15, 2025Updated last year
- MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.☆4,333Apr 9, 2026Updated last week
- Distributed Inference for mlx LLm☆100Aug 1, 2024Updated last year
- Generate train.jsonl and valid.jsonl files to use for fine-tuning Mistral and other LLMs.☆97Feb 5, 2024Updated 2 years ago
- MLX native implementations of state-of-the-art generative image models☆1,981Updated this week
- ☆14Oct 18, 2023Updated 2 years ago
- tiny_fnc_engine is a minimal python library that provides a flexible engine for calling functions extracted from a LLM.☆38Sep 11, 2024Updated last year
- Serverless GPU API endpoints on Runpod - Bonus Credits • AdSkip the infrastructure headaches. Auto-scaling, pay-as-you-go, no-ops approach lets you focus on innovating your application.
- A reinforcement learning framework based on MLX.☆254Dec 1, 2025Updated 4 months ago
- ☆20Oct 25, 2025Updated 5 months ago
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆102Jun 29, 2025Updated 9 months ago
- ☆283Jun 4, 2024Updated last year
- ☆21Mar 10, 2024Updated 2 years ago
- A simple Python sandbox for helpful LLM data agents☆309Updated this week
- Examples using MLX Swift☆2,496Updated this week