mustafaaljadery / mlxserver
Start a server from the MLX library.
☆169Updated 6 months ago
Alternatives and similar repositories for mlxserver:
Users that are interested in mlxserver are comparing it to the libraries listed below
- FastMLX is a high performance production ready API to host MLX models.☆256Updated 2 months ago
- MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.☆93Updated 3 months ago
- Fast parallel LLM inference for MLX☆153Updated 6 months ago
- A simple UI / Web / Frontend for MLX mlx-lm using Streamlit.☆237Updated last month
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆222Updated 9 months ago
- An mlx project to train a base model on your whatsapp chats using (Q)Lora finetuning☆161Updated last year
- The easiest way to run the fastest MLX-based LLMs locally☆240Updated 2 months ago
- ☆207Updated 6 months ago
- run embeddings in MLX☆81Updated 4 months ago
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆246Updated last week
- For inferring and serving local LLMs using the MLX framework☆91Updated 10 months ago
- ☆147Updated last month
- Benchmark of Apple MLX operations on all Apple Silicon chips (GPU, CPU) + MPS and CUDA.☆149Updated 2 months ago
- The Batched API provides a flexible and efficient way to process multiple requests in a batch, with a primary focus on dynamic batching o…☆119Updated last month
- Explore a simple example of utilizing MLX for RAG application running locally on your Apple Silicon device.☆154Updated 11 months ago
- ☆315Updated 3 months ago
- Routing on Random Forest (RoRF)☆98Updated 4 months ago
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. I…☆241Updated 2 weeks ago
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.☆396Updated 2 weeks ago
- Scripts to create your own moe models using mlx☆86Updated 11 months ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆225Updated 2 months ago
- ☆125Updated 3 weeks ago
- A simple Python sandbox for helpful LLM data agents☆215Updated 7 months ago
- ☆196Updated 8 months ago
- ☆38Updated 10 months ago
- Create keyboard shortcuts for an LLM using OpenAI GPT, Ollama, HuggingFace with Automator on macOS.☆143Updated 10 months ago
- Generate train.jsonl and valid.jsonl files to use for fine-tuning Mistral and other LLMs.☆84Updated 11 months ago
- ☆109Updated last month
- ☆136Updated last year
- A simple Jupyter Notebook for learning MLX text-completion fine-tuning!☆105Updated 2 months ago