1.58 Bit LLM on Apple Silicon using MLX
☆265May 10, 2024Updated last year
Alternatives and similar repositories for mlx-bitnet
Users that are interested in mlx-bitnet are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Supporting code for "LLMs for your iPhone: Whole-Tensor 4 Bit Quantization"☆11Mar 31, 2024Updated 2 years ago
- Very basic framework for composable parameterized large language model (Q)LoRA / (Q)Dora fine-tuning using mlx, mlx_lm, and OgbujiPT.☆43Jun 20, 2025Updated 9 months ago
- A tiny server to run local inference on MLX model in the style of OpenAI☆13Jan 31, 2024Updated 2 years ago
- Distributed Inference for mlx LLm☆100Aug 1, 2024Updated last year
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆284Jun 16, 2025Updated 9 months ago
- Wordpress hosting with auto-scaling - Free Trial • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- REAP expert pruning for MoE LLMs on Apple Silicon via MLX☆53Mar 16, 2026Updated 3 weeks ago
- A simple UI / Web / Frontend for MLX mlx-lm using Streamlit.☆262Oct 25, 2025Updated 5 months ago
- Fast parallel LLM inference for MLX☆249Jul 7, 2024Updated last year
- ☆10Nov 16, 2024Updated last year
- MLX Transformers is a library that provides model implementation in MLX. It uses a similar model interface as HuggingFace Transformers an…☆76Mar 23, 2026Updated 3 weeks ago
- import documents for LLMs☆48Mar 30, 2026Updated 2 weeks ago
- Code for the examples presented in the talk "Training a Llama in your backyard: fine-tuning very large models on consumer hardware" given…☆15Oct 16, 2023Updated 2 years ago
- 📋 NotebookMLX - An Open Source version of NotebookLM (Ported NotebookLlama)☆340Mar 3, 2025Updated last year
- Run frontier AI locally.☆43,503Updated this week
- Deploy open-source AI quickly and easily - Bonus Offer • AdRunpod Hub is built for open source. One-click deployment and autoscaling endpoints without provisioning your own infrastructure.
- Implementation of F5-TTS in Swift using MLX☆90Dec 11, 2024Updated last year
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.☆462Jan 29, 2025Updated last year
- Minimal Claude Code alternative powered by MLX☆46Jan 11, 2026Updated 3 months ago
- Implementing the BitNet model in Rust☆48Apr 18, 2024Updated last year
- Implementation of the Mamba SSM with hf_integration.☆55Aug 31, 2024Updated last year
- Clawdbot skills for agentic coding workflows - ACFS stack, cloud CLIs, and dev tools☆62Mar 27, 2026Updated 2 weeks ago
- Implementation of nougat that focuses on processing pdf locally.☆85Jan 15, 2025Updated last year
- LLM training in simple, raw C/Metal Shading Language☆63Apr 24, 2024Updated last year
- A simple LLaMA implementation using MLX.☆15Apr 22, 2024Updated last year
- Wordpress hosting with auto-scaling - Free Trial • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- Examples in the MLX framework☆8,459Apr 6, 2026Updated last week
- It's a baby compiler. (Lean btw.)☆16May 19, 2025Updated 10 months ago
- The easiest way to run the fastest MLX-based LLMs locally☆321Oct 30, 2024Updated last year
- MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.☆349Mar 24, 2026Updated 3 weeks ago
- MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.☆4,138Apr 6, 2026Updated last week
- Efficient framework-agnostic data loading☆465Oct 1, 2025Updated 6 months ago
- Tiny ASIC implementation for "The Era of 1-bit LLMs All Large Language Models are in 1.58 Bits" matrix multiplication unit☆186Apr 19, 2024Updated last year
- Your gateway to both Ollama & Apple MlX models☆153Mar 2, 2025Updated last year
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. I…☆694Mar 10, 2026Updated last month
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- ☆11Jul 17, 2023Updated 2 years ago
- Run GreenBitAI's Quantized LLMs on Apple Devices with MLX☆31Aug 27, 2025Updated 7 months ago
- TerDiT: Ternary Diffusion Models with Transformers☆74Jun 17, 2024Updated last year
- A reinforcement learning framework based on MLX.☆254Dec 1, 2025Updated 4 months ago
- run embeddings in MLX☆98Sep 27, 2024Updated last year
- Open-source implementation of Google's TurboQuant (ICLR 2026) — KV cache compression to 2.5–4 bits with near-zero quality loss. 3.8–5.7x …☆46Mar 29, 2026Updated 2 weeks ago
- TOPLOC: is a novel method for verifiable inference that enables users to verify that LLM providers are using the correct model configurat…☆54Apr 14, 2025Updated last year