Distribute and run LLMs with a single file.
☆24,349May 1, 2026Updated this week
Alternatives and similar repositories for llamafile
Users that are interested in llamafile are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- LLM inference in C/C++☆107,892Updated this week
- Get up and running with Kimi-K2.5, GLM-5, MiniMax, DeepSeek, gpt-oss, Qwen, Gemma and other models.☆170,289Updated this week
- build-once run-anywhere c library☆20,780Mar 6, 2026Updated last month
- Port of OpenAI's Whisper model in C/C++☆49,148Apr 20, 2026Updated 2 weeks ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆78,385Updated this week
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- Jan is an open source alternative to ChatGPT that runs 100% offline on your computer.☆42,247Updated this week
- Python SDK, Proxy Server (AI Gateway) to call 100+ LLM APIs in OpenAI (or native) format, with cost tracking, guardrails, loadbalancing a…☆45,153Updated this week
- Web UI for training and running open models like Gemma 4, Qwen3.6, DeepSeek, gpt-oss locally.☆63,070Apr 27, 2026Updated last week
- Tensor library for machine learning☆14,560Updated this week
- LocalAI is the open-source AI engine. Run any model - LLMs, vision, voice, image, video - on any hardware. No GPU required.☆45,883Updated this week
- aider is AI pair programming in your terminal☆44,187Apr 25, 2026Updated last week
- A natural language interface for computers☆63,344Apr 27, 2026Updated last week
- LlamaIndex is the leading document agent and OCR platform☆48,997Updated this week
- Inference Llama 2 in one file of pure C☆19,460Aug 6, 2024Updated last year
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- Universal LLM Deployment Engine with ML Compilation☆22,557Apr 22, 2026Updated last week
- GPT4All: Run Local LLMs on Any Device. Open-source and available for commercial use.☆77,367May 27, 2025Updated 11 months ago
- High-performance In-browser LLM Inference Engine☆17,858Apr 24, 2026Updated last week
- User-friendly AI Interface (Supports Ollama, OpenAI API, ...)☆135,272Updated this week
- DSPy: The framework for programming—not prompting—language models☆34,016Apr 24, 2026Updated last week
- The original local LLM interface. Text, vision, tool-calling, training. UI + API, 100% offline and private.☆46,874Apr 27, 2026Updated last week
- Interact with your documents using the power of GPT, 100% privately, no data leaks☆57,207Feb 26, 2026Updated 2 months ago
- The all-in-one AI productivity accelerator. On device and privacy first with no annoying setup or configuration.☆59,383Updated this week
- 🙌 OpenHands: AI-Driven Development☆72,542Updated this week
- Serverless GPU API endpoints on Runpod - Get Bonus Credits • AdSkip the infrastructure headaches. Auto-scaling, pay-as-you-go, no-ops approach lets you focus on innovating your application.
- Self-hosted AI coding assistant☆33,473Mar 2, 2026Updated 2 months ago
- ⏩ Source-controlled AI checks, enforceable in CI. Powered by the open-source Continue CLI☆32,849Updated this week
- LLM training in simple, raw C/CUDA☆29,780Jun 26, 2025Updated 10 months ago
- 20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.☆13,337Updated this week
- Fast, flexible LLM inference☆7,074Apr 15, 2026Updated 2 weeks ago
- Letta is the platform for building stateful agents: AI with advanced memory that can learn and self-improve over time.☆22,391Apr 12, 2026Updated 3 weeks ago
- Structured Outputs☆13,776Apr 16, 2026Updated 2 weeks ago
- A vector search SQLite extension that runs anywhere!☆7,520Apr 8, 2026Updated 3 weeks ago
- Run frontier AI locally.☆44,293Updated this week
- Serverless GPU API endpoints on Runpod - Get Bonus Credits • AdSkip the infrastructure headaches. Auto-scaling, pay-as-you-go, no-ops approach lets you focus on innovating your application.
- MLX: An array framework for Apple silicon☆25,814Updated this week
- Vane is an AI-powered answering engine.☆34,009Apr 11, 2026Updated 3 weeks ago
- Universal memory layer for AI Agents☆54,199Apr 25, 2026Updated last week
- A programming framework for agentic AI☆57,588Apr 15, 2026Updated 2 weeks ago
- Go ahead and axolotl questions☆11,779Updated this week
- Access large language models from the command-line☆11,762Updated this week
- Convert PDF to markdown + JSON quickly with high accuracy☆34,606Apr 24, 2026Updated last week