maxbbraun / llama4micro
A "large" language model running on a microcontroller
☆494Updated 11 months ago
Related projects ⓘ
Alternatives and complementary repositories for llama4micro
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆557Updated last year
- An mlx project to train a base model on your whatsapp chats using (Q)Lora finetuning☆159Updated 9 months ago
- A minimal Tensor Processing Unit (TPU) inspired by Google's TPUv1.☆116Updated 3 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆136Updated 6 months ago
- A really tiny autograd engine☆87Updated 7 months ago
- Start a server from the MLX library.☆159Updated 3 months ago
- prime (previously called ZeroBand) is a framework for efficient, globally distributed training of AI models over the internet.☆207Updated this week
- LLaVA server (llama.cpp).☆177Updated last year
- Mistral7B playing DOOM☆122Updated 3 months ago
- The Tensor (or Array)☆408Updated 3 months ago
- Inference Vision Transformer (ViT) in plain C/C++ with ggml☆229Updated 7 months ago
- a small code base for training large models☆264Updated last week
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.