maxbbraun / llama4microLinks
A "large" language model running on a microcontroller
☆545Updated 2 years ago
Alternatives and similar repositories for llama4micro
Users that are interested in llama4micro are comparing it to the libraries listed below
Sorting:
- Instructions on how to run LLMs on Raspberry PI☆210Updated last year
- run paligemma in real time☆133Updated last year
- LLaVA server (llama.cpp).☆183Updated 2 years ago
- a small code base for training large models☆318Updated 8 months ago
- Llama 2 Everywhere (L2E)☆1,526Updated 4 months ago
- llama.cpp with BakLLaVA model describes what does it see☆380Updated 2 years ago
- Inference Llama 2 in one file of pure Python☆425Updated last month
- An mlx project to train a base model on your whatsapp chats using (Q)Lora finetuning☆172Updated 2 years ago
- Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wi…☆355Updated last year
- Efficient Inference of Transformer models☆478Updated last year
- llama3.np is a pure NumPy implementation for Llama 3 model.☆993Updated 8 months ago
- A modern model graph visualizer and debugger☆1,365Updated last week
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆568Updated 2 years ago
- Mistral7B playing DOOM☆138Updated last year
- Fine-tune mistral-7B on 3090s, a100s, h100s☆723Updated 2 years ago
- gpt-2 from scratch in mlx☆414Updated last year
- 1.58 Bit LLM on Apple Silicon using MLX☆238Updated last year
- Let's make sand talk☆591Updated 2 years ago
- Gradient descent is cool and all, but what if we could delete it?☆105Updated 4 months ago
- The repository for the code of the UltraFastBERT paper☆519Updated last year
- SoTA Transformers with C-backend for fast inference on your CPU.☆311Updated 2 years ago
- A simplistic linear and multiprocessed approach to sentiment analysis using Gzip Normalized Compression Distances with k nearest neighbor…☆144Updated 2 years ago
- ☆96Updated last year
- C++ implementation for BLOOM☆808Updated 2 years ago
- CLIP inference in plain C/C++ with no extra dependencies☆547Updated 6 months ago
- ☆866Updated 2 years ago
- Alex Krizhevsky's original code from Google Code☆198Updated 9 years ago
- LLM-based code completion engine☆190Updated 11 months ago
- LLM-powered lossless compression tool☆298Updated 2 weeks ago
- A really tiny autograd engine☆99Updated 7 months ago