mozilla-ai / llamafileLinks
Distribute and run LLMs with a single file.
☆23,553Updated last week
Alternatives and similar repositories for llamafile
Users that are interested in llamafile are comparing it to the libraries listed below
Sorting:
- Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.☆158,406Updated last week
- ⏩ Ship faster with Continuous AI. Open-source CLI that can be used in TUI mode as a coding agent or Headless mode to run background agent…☆30,543Updated this week
- The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on…☆40,785Updated this week
- LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a ch…☆5,964Updated 2 weeks ago
- Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.☆20,400Updated last week
- Tensor library for machine learning☆13,764Updated 2 weeks ago
- A vector search SQLite extension that runs anywhere!☆6,584Updated 11 months ago
- Inference Llama 2 in one file of pure C☆19,063Updated last year
- User-friendly AI Interface (Supports Ollama, OpenAI API, ...)☆118,987Updated this week
- 🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading☆9,864Updated last year
- MLX: An array framework for Apple silicon☆23,275Updated last week
- High-speed Large Language Model Serving for Local Deployment☆8,503Updated 4 months ago
- Self-hosted AI coding assistant☆32,635Updated this week
- LLM inference in C/C++☆92,005Updated this week
- 20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.☆13,052Updated last week
- Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚☆38,251Updated this week
- Blazingly fast LLM inference.☆6,310Updated last week
- Private & local AI personal knowledge management app for high entropy people.☆8,440Updated 7 months ago
- Fully private LLM chatbot that runs entirely with a browser with no server needed. Supports Mistral and LLama 3.☆2,674Updated last year
- aider is AI pair programming in your terminal☆39,329Updated last week
- Examples in the MLX framework☆8,085Updated 2 weeks ago
- SWE-agent takes a GitHub issue and tries to automatically fix it, using your LM of choice. It can also be employed for offensive cybersec…☆18,109Updated last week
- The definitive Web UI for local AI, with powerful features and easy setup.☆45,706Updated last week
- Ollama Python library☆9,069Updated 3 weeks ago
- The all-in-one Desktop & Docker AI application with built-in RAG, AI agents, No-code agent builder, MCP compatibility, and more.☆52,729Updated last week
- Open source codebase powering the HuggingChat app☆10,397Updated last week
- The most no-nonsense, locally or API-hosted AI code completion plugin for Visual Studio Code - like GitHub Copilot but 100% free.☆3,621Updated 4 months ago
- Inference code for CodeLlama models☆16,372Updated last year
- High-performance In-browser LLM Inference Engine☆17,027Updated last month
- A self-hosted, offline, ChatGPT-like chatbot. Powered by Llama 2. 100% private, with no data leaving your device. New: Code Llama support…☆10,990Updated last year