neobundy / Deep-Dive-Into-AI-With-MLX-PyTorch
"Deep Dive into AI with MLX and PyTorch" is an educational initiative designed to help anyone interested in AI, specifically in machine learning and deep learning, using Apple's MLX and Meta's PyTorch frameworks.
☆380Updated 7 months ago
Related projects ⓘ
Alternatives and complementary repositories for Deep-Dive-Into-AI-With-MLX-PyTorch
- MLX-VLM is a package for running Vision LLMs locally on your Mac using MLX.☆496Updated this week
- On-device Inference of Diffusion Models for Apple Silicon☆509Updated 3 weeks ago
- Fast parallel LLM inference for MLX☆149Updated 4 months ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆221Updated 3 weeks ago
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.☆348Updated 2 months ago
- Official implementation of Half-Quadratic Quantization (HQQ)☆701Updated last week
- A simple UI / Web / Frontend for MLX mlx-lm using Streamlit.☆227Updated last month
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024☆229Updated 3 weeks ago
- A reinforcement learning framework based on MLX.☆220Updated 9 months ago
- Explore a simple example of utilizing MLX for RAG application running locally on your Apple Silicon device.☆145Updated 9 months ago
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆237Updated 2 months ago
- System 2 Reasoning Link Collection☆693Updated 3 weeks ago
- A comprehensive deep dive into the world of tokens☆214Updated 4 months ago
- ☆641Updated this week
- An Open Source Toolkit For LLM Distillation☆356Updated 2 months ago
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆226Updated this week
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines☆195Updated 6 months ago
- LoRA and DoRA from Scratch Implementations☆188Updated 8 months ago
- FastMLX is a high performance production ready API to host MLX models.☆218Updated 3 weeks ago
- Fine-tune mistral-7B on 3090s, a100s, h100s☆702Updated last year
- Fast bare-bones BPE for modern tokenizer training☆142Updated last month
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆173Updated 4 months ago
- Start a server from the MLX library.☆161Updated 3 months ago
- Automatically evaluate your LLMs in Google Colab☆559Updated 6 months ago
- ☆191Updated 5 months ago
- The Tensor (or Array)☆411Updated 3 months ago
- ☆532Updated 3 weeks ago
- ☆411Updated last year
- ☆448Updated 7 months ago
- Official implementation of "Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling"☆803Updated 3 months ago