maxbbraun / llama4micro
A "large" language model running on a microcontroller
☆520Updated last year
Alternatives and similar repositories for llama4micro:
Users that are interested in llama4micro are comparing it to the libraries listed below
- Mistral7B playing DOOM☆130Updated 8 months ago
- Let's make sand talk☆589Updated last year
- run paligemma in real time☆131Updated 10 months ago
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆566Updated last year
- Fine-tune mistral-7B on 3090s, a100s, h100s☆709Updated last year
- CLIP inference in plain C/C++ with no extra dependencies☆487Updated 7 months ago
- Alex Krizhevsky's original code from Google Code☆191Updated 9 years ago
- An mlx project to train a base model on your whatsapp chats using (Q)Lora finetuning☆165Updated last year
- gpt-2 from scratch in mlx☆378Updated 9 months ago
- Finetune llama2-70b and codellama on MacBook Air without quantization☆448Updated last year
- Run GGML models with Kubernetes.☆174Updated last year
- LLaVA server (llama.cpp).☆179Updated last year
- WebGPU LLM inference tuned by hand☆149Updated last year
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.☆432Updated 2 months ago
- A reinforcement learning framework based on MLX.☆232Updated last month
- A really tiny autograd engine☆90Updated 11 months ago
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines☆198Updated 10 months ago
- ☆412Updated last year
- SoTA Transformers with C-backend for fast inference on your CPU.☆311Updated last year
- 1.58 Bit LLM on Apple Silicon using MLX☆194Updated 10 months ago
- Llama 2 Everywhere (L2E)☆1,516Updated 2 months ago
- Understanding large language models☆118Updated 2 years ago
- A personal wearable AI that runs locally☆570Updated last year
- ☆1,274Updated last year
- llama.cpp with BakLLaVA model describes what does it see☆384Updated last year
- C++ implementation for BLOOM☆809Updated last year
- Efficient Inference of Transformer models☆427Updated 7 months ago
- LLM-powered lossless compression tool☆274Updated 7 months ago
- C++ implementation for 💫StarCoder☆453Updated last year
- Gradient descent is cool and all, but what if we could delete it?☆103Updated 3 weeks ago