maxbbraun / llama4microLinks
A "large" language model running on a microcontroller
☆533Updated last year
Alternatives and similar repositories for llama4micro
Users that are interested in llama4micro are comparing it to the libraries listed below
Sorting:
- Llama 2 Everywhere (L2E)☆1,523Updated this week
- a small code base for training large models☆309Updated 4 months ago
- Inference Llama 2 in one file of pure Python☆422Updated 10 months ago
- run paligemma in real time☆131Updated last year
- Instructions on how to run LLMs on Raspberry PI☆208Updated last year
- gpt-2 from scratch in mlx☆396Updated last year
- Fine-tune mistral-7B on 3090s, a100s, h100s☆718Updated last year
- The Tensor (or Array)☆441Updated last year
- A really tiny autograd engine☆95Updated 3 months ago
- llama.cpp with BakLLaVA model describes what does it see☆382Updated last year
- Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wi…☆349Updated last year
- Alex Krizhevsky's original code from Google Code☆196Updated 9 years ago
- The repository for the code of the UltraFastBERT paper☆517Updated last year
- llama3.np is a pure NumPy implementation for Llama 3 model.☆988Updated 4 months ago
- An mlx project to train a base model on your whatsapp chats using (Q)Lora finetuning☆168Updated last year
- Efficient Inference of Transformer models☆449Updated last year
- Accelerate your Hugging Face Transformers 7.6-9x. Native to Hugging Face and PyTorch.☆686Updated last year
- Following master Karpathy with GPT-2 implementation and training, writing lots of comments cause I have memory of a goldfish☆172Updated last year
- ☆864Updated last year
- ☆96Updated last year
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆568Updated 2 years ago
- Run GGML models with Kubernetes.☆174Updated last year
- C++ implementation for BLOOM☆809Updated 2 years ago
- A reinforcement learning framework based on MLX.☆237Updated 6 months ago
- Mistral7B playing DOOM☆135Updated last year
- TinyChatEngine: On-Device LLM Inference Library☆887Updated last year
- An implementation of bucketMul LLM inference☆223Updated last year
- 1.58 Bit LLM on Apple Silicon using MLX☆221Updated last year
- LLaVA server (llama.cpp).☆181Updated last year
- ☆249Updated last year