maxbbraun / llama4microLinks
A "large" language model running on a microcontroller
☆547Updated 2 years ago
Alternatives and similar repositories for llama4micro
Users that are interested in llama4micro are comparing it to the libraries listed below
Sorting:
- run paligemma in real time☆133Updated last year
- Instructions on how to run LLMs on Raspberry PI☆209Updated last year
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆569Updated 2 years ago
- Efficient Inference of Transformer models☆478Updated last year
- gpt-2 from scratch in mlx☆414Updated last year
- LLaVA server (llama.cpp).☆183Updated 2 years ago
- Inference Llama 2 in one file of pure Python☆425Updated 2 months ago
- ☆1,028Updated 2 years ago
- llama.cpp with BakLLaVA model describes what does it see☆380Updated 2 years ago
- Inference Vision Transformer (ViT) in plain C/C++ with ggml☆31Updated 2 years ago
- Alex Krizhevsky's original code from Google Code☆199Updated 9 years ago
- A really tiny autograd engine☆99Updated 8 months ago
- Running a LLM on the ESP32☆474Updated last year
- A reinforcement learning framework based on MLX.☆248Updated 2 months ago
- Fine-tune mistral-7B on 3090s, a100s, h100s☆725Updated 2 years ago
- a small code base for training large models☆322Updated 9 months ago
- llama3.np is a pure NumPy implementation for Llama 3 model.☆991Updated 9 months ago
- An mlx project to train a base model on your whatsapp chats using (Q)Lora finetuning☆173Updated 2 years ago
- Llama 2 Everywhere (L2E)☆1,526Updated 5 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆243Updated last year
- ☆867Updated 2 years ago
- The repository for the code of the UltraFastBERT paper