maxbbraun / llama4microLinks
A "large" language model running on a microcontroller
☆531Updated last year
Alternatives and similar repositories for llama4micro
Users that are interested in llama4micro are comparing it to the libraries listed below
Sorting:
- run paligemma in real time☆131Updated last year
- Running a LLM on the ESP32☆331Updated 10 months ago
- Instructions on how to run LLMs on Raspberry PI☆207Updated 11 months ago
- Efficient Inference of Transformer models☆439Updated 11 months ago
- An mlx project to train a base model on your whatsapp chats using (Q)Lora finetuning☆168Updated last year
- llama.cpp with BakLLaVA model describes what does it see☆382Updated last year
- Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wi…☆348Updated 11 months ago
- a small code base for training large models☆305Updated 2 months ago
- A really tiny autograd engine☆94Updated last month
- Let's make sand talk☆591Updated last year
- Mistral7B playing DOOM☆132Updated last year
- ☆111Updated last year
- gpt-2 from scratch in mlx☆391Updated last year
- ☆96Updated last year
- CLIP inference in plain C/C++ with no extra dependencies☆508Updated last month
- Inference Llama 2 in one file of pure Python☆418Updated 9 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆214Updated last year
- Following master Karpathy with GPT-2 implementation and training, writing lots of comments cause I have memory of a goldfish☆173Updated 11 months ago
- Alex Krizhevsky's original code from Google Code☆194Updated 9 years ago
- Llama 2 Everywhere (L2E)☆1,519Updated 6 months ago
- programmable e-paper tag with RFID☆320Updated 8 months ago
- Fine-tune mistral-7B on 3090s, a100s, h100s☆715Updated last year
- Gradient descent is cool and all, but what if we could delete it?☆104Updated this week
- The Tensor (or Array)☆437Updated 11 months ago
- LLM-powered lossless compression tool☆283Updated 11 months ago
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆567Updated last year
- ☆864Updated last year
- A reinforcement learning framework based on MLX.☆234Updated 4 months ago
- An implementation of bucketMul LLM inference☆220Updated last year
- SoTA Transformers with C-backend for fast inference on your CPU.☆309Updated last year