The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM models, execute structured function calls and get structured output. Works also with models not fine-tuned to JSON output and function calls.
☆630Mar 9, 2026Updated last month
Alternatives and similar repositories for llama-cpp-agent
Users that are interested in llama-cpp-agent are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- ToolAgents is a lightweight and flexible framework for creating function-calling agents with various language models and APIs.☆35Apr 5, 2026Updated 3 weeks ago
- ☆32Dec 29, 2023Updated 2 years ago
- Locally running LLM with internet access☆96Jun 30, 2025Updated 10 months ago
- function calling-based LLM agents☆290Sep 16, 2024Updated last year
- Python bindings for llama.cpp☆10,240Updated this week
- Deploy on Railway without the complexity - Free Credits Offer • AdConnect your repo and Railway handles the rest with instant previews. Quickly provision container image services, databases, and storage volumes.
- A Comprehensive survey on business use cases of AI that help them thrive in the digital economy☆13Oct 7, 2020Updated 5 years ago
- Chat language model that can use tools and interpret the results☆1,594Dec 3, 2025Updated 4 months ago
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces☆142Jul 9, 2024Updated last year
- Modified Beam Search with periodical restart☆12Sep 12, 2024Updated last year
- Inference of Mamba, Mamba2 and Mamba3 models in pure C☆200Mar 18, 2026Updated last month
- An experimental desktop client for using Claude Desktop's MCP with Novelcrafter codices.☆11Dec 3, 2024Updated last year
- This GUI aims to simplify the process of converting GGUF files to llamafile format by providing an intuitive and convenient way for users…☆14Jan 2, 2026Updated 3 months ago
- Harness LLMs with Multi-Agent Programming☆3,989Apr 7, 2026Updated 3 weeks ago
- Your Trusty Memory-enabled AI Companion - Simple RAG chatbot optimized for local LLMs | 12 Languages Supported | OpenAI API Compatible☆352Feb 28, 2025Updated last year
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- A guidance compatibility layer for llama-cpp-python☆36Sep 11, 2023Updated 2 years ago
- Create Custom LLMs☆1,831Apr 24, 2026Updated last week
- Python bindings for the Transformer models implemented in C/C++ using GGML library.☆1,886Jan 28, 2024Updated 2 years ago
- A multimodal, function calling powered LLM webui.☆213Sep 23, 2024Updated last year
- Enforce the output format (JSON Schema, Regex etc) of a language model☆2,011Apr 4, 2026Updated 3 weeks ago
- Simple agent framework using Ollama tool calling☆10Aug 27, 2024Updated last year
- ☆1,307Dec 22, 2025Updated 4 months ago
- ☆135Apr 8, 2026Updated 3 weeks ago
- cli tool to quantize gguf, gptq, awq, hqq and exl2 models☆79Dec 17, 2024Updated last year
- Managed Database hosting by DigitalOcean • AdPostgreSQL, MySQL, MongoDB, Kafka, Valkey, and OpenSearch available. Automatically scale up storage and focus on building your apps.
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆1,197Updated this week
- An AI assistant beyond the chat box.☆329Mar 11, 2024Updated 2 years ago
- ☆345Mar 5, 2026Updated last month
- ☆19Jun 5, 2023Updated 2 years ago
- Open-source LLM/VLM load balancer and serving platform for self-hosting LLMs (and VLMs) at scale 🏓🦙 Alternative to projects like llm-d,…☆1,540Updated this week
- A tool for generating function arguments and choosing what function to call with local LLMs☆438Mar 12, 2024Updated 2 years ago
- WilmerAI is one of the oldest LLM semantic routers. It uses multi-layer prompt routing and complex workflows to allow you to not only cre…☆811Apr 18, 2026Updated last week
- ☆38Mar 12, 2024Updated 2 years ago
- A library for working with GBNF files☆29Nov 2, 2025Updated 5 months ago
- Deploy on Railway without the complexity - Free Credits Offer • AdConnect your repo and Railway handles the rest with instant previews. Quickly provision container image services, databases, and storage volumes.
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆150Jan 7, 2026Updated 3 months ago
- Python package wrapping llama.cpp for on-device LLM inference☆105Apr 2, 2026Updated 3 weeks ago
- Tools for merging pretrained large language models.☆7,023Mar 15, 2026Updated last month
- A simple experiment on letting two local LLM have a conversation about anything!☆112Jul 3, 2024Updated last year
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆4,511Mar 4, 2026Updated last month
- Efficient visual programming for AI language models☆361May 13, 2025Updated 11 months ago
- Serving LLMs in the HF-Transformers format via a PyFlask API☆72Sep 10, 2024Updated last year