maxbbraun / llama4microLinks
A "large" language model running on a microcontroller
☆539Updated last year
Alternatives and similar repositories for llama4micro
Users that are interested in llama4micro are comparing it to the libraries listed below
Sorting:
- a small code base for training large models☆310Updated 6 months ago
- run paligemma in real time☆133Updated last year
- Instructions on how to run LLMs on Raspberry PI☆208Updated last year
- llama3.np is a pure NumPy implementation for Llama 3 model.☆989Updated 6 months ago
- An mlx project to train a base model on your whatsapp chats using (Q)Lora finetuning☆169Updated last year
- LLaVA server (llama.cpp).☆183Updated 2 years ago
- 1.58 Bit LLM on Apple Silicon using MLX☆225Updated last year
- gpt-2 from scratch in mlx☆401Updated last year
- Llama 2 Everywhere (L2E)☆1,522Updated 2 months ago
- Alex Krizhevsky's original code from Google Code☆199Updated 9 years ago
- llama.cpp with BakLLaVA model describes what does it see☆382Updated last year
- Mistral7B playing DOOM☆138Updated last year
- ☆96Updated last year
- LLM-powered lossless compression tool☆288Updated last year
- An implementation of bucketMul LLM inference☆223Updated last year
- Open weights language model from Google DeepMind, based on Griffin.☆653Updated 4 months ago
- Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wi…☆350Updated last year
- TinyChatEngine: On-Device LLM Inference Library☆906Updated last year
- Computer Vision and Machine Learning Jupyter Notebooks for Educational Purposes☆79Updated 10 months ago
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆568Updated 2 years ago
- GGUF implementation in C as a library and a tools CLI program☆292Updated 2 months ago
- SoTA Transformers with C-backend for fast inference on your CPU.☆308Updated last year
- Efficient Inference of Transformer models☆461Updated last year
- Inference Llama 2 in one file of pure Python☆422Updated last year
- The Tensor (or Array)☆451Updated last year
- Gradient descent is cool and all, but what if we could delete it?☆104Updated 2 months ago
- A really tiny autograd engine☆96Updated 5 months ago
- ☆865Updated last year
- Visualize the intermediate output of Mistral 7B☆375Updated 9 months ago
- throwaway GPT inference☆140Updated last year