ml-explore / mlxLinks
MLX: An array framework for Apple silicon
☆22,522Updated this week
Alternatives and similar repositories for mlx
Users that are interested in mlx are comparing it to the libraries listed below
Sorting:
- Examples in the MLX framework☆7,926Updated 2 weeks ago
- ☆8,655Updated last year
- Official inference library for Mistral models☆10,506Updated 7 months ago
- 20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.☆12,840Updated last week
- Tensor library for machine learning☆13,302Updated last week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆60,385Updated this week
- An Extensible Deep Learning Library☆2,271Updated this week
- CoreNet: A library for training deep neural networks☆7,024Updated 2 weeks ago
- lightweight, standalone C++ inference engine for Google's Gemma models.☆6,591Updated last week
- Universal LLM Deployment Engine with ML Compilation☆21,497Updated this week
- Open-source search and retrieval database for AI applications.☆23,924Updated this week
- SGLang is a fast serving framework for large language models and vision language models.☆19,094Updated this week
- Reference implementation of the Transformer architecture optimized for Apple Neural Engine (ANE)☆2,661Updated 2 years ago
- Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.☆47,063Updated this week
- High-speed Large Language Model Serving for Local Deployment☆8,367Updated 2 months ago
- Stable Diffusion with Core ML on Apple Silicon☆17,650Updated 3 months ago
- Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚☆31,859Updated last week
- Go ahead and axolotl questions☆10,634Updated this week
- The official PyTorch implementation of Google's Gemma models☆5,560Updated 4 months ago
- The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.☆8,775Updated last year
- Structured Outputs☆12,712Updated last week
- Inference Llama 2 in one file of pure C☆18,848Updated last year
- Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We als…☆17,954Updated this week
- LLM inference in C/C++☆87,889Updated last week
- LLM training in simple, raw C/CUDA☆27,923Updated 3 months ago
- ☆4,100Updated last year
- [ICLR 2024] Efficient Streaming Language Models with Attention Sinks☆7,070Updated last year
- Gemma open-weight LLM library, from Google DeepMind☆3,767Updated this week
- Large Language Model Text Generation Inference☆10,580Updated last month
- Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.☆10,021Updated last year