Maximilian-Winter / llama-cpp-agent
The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM models, execute structured function calls and get structured output. Works also with models not fine-tuned to JSON output and function calls.
☆545Updated last month
Alternatives and similar repositories for llama-cpp-agent:
Users that are interested in llama-cpp-agent are comparing it to the libraries listed below
- ☆832Updated 6 months ago
- function calling-based LLM agents☆283Updated 6 months ago
- Web UI for ExLlamaV2☆486Updated last month
- A fast batching API to serve LLM models☆182Updated 10 months ago
- A multimodal, function calling powered LLM webui.☆215Updated 5 months ago
- An application for running LLMs locally on your device, with your documents, facilitating detailed citations in generated responses.☆566Updated 4 months ago
- An AI assistant beyond the chat box.☆322Updated last year
- Large-scale LLM inference engine☆1,342Updated this week
- Dataset Crafting w/ RAG/Wikipedia ground truth and Efficient Fine-Tuning Using MLX and Unsloth. Includes configurable dataset annotation …☆176Updated 8 months ago
- This is our own implementation of 'Layer Selective Rank Reduction'☆233Updated 9 months ago
- Efficient visual programming for AI language models☆349Updated 6 months ago
- ☆273Updated last month
- An OAI compatible exllamav2 API that's both lightweight and fast☆858Updated this week
- ☆196Updated this week
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆232Updated 2 weeks ago
- FastMLX is a high performance production ready API to host MLX models.☆272Updated last week
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆224Updated 10 months ago
- Automatically evaluate your LLMs in Google Colab☆604Updated 10 months ago
- 🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.☆343Updated 3 months ago
- Your Trusty Memory-enabled AI Companion - Simple RAG chatbot optimized for local LLMs | 12 Languages Supported | OpenAI API Compatible☆305Updated 2 weeks ago
- Customizable implementation of the self-instruct paper.☆1,039Updated last year
- A python package for developing AI applications with local LLMs.☆145Updated 2 months ago
- Convenience scripts to finetune (chat-)LLaMa3 and other models for any language☆301Updated 9 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆147Updated 10 months ago
- Building open version of OpenAI o1 via reasoning traces (Groq, ollama, Anthropic, Gemini, OpenAI, Azure supported) Demo: https://hugging…☆175Updated 5 months ago
- A library for easily merging multiple LLM experts, and efficiently train the merged LLM.☆450Updated 6 months ago
- Official implementation of Half-Quadratic Quantization (HQQ)☆765Updated this week
- A comprehensive repository of reasoning tasks for LLMs (and beyond)☆423Updated 5 months ago
- This project demonstrates a basic chain-of-thought interaction with any LLM (Large Language Model)☆319Updated 5 months ago
- Low-Rank adapter extraction for fine-tuned transformers models☆171Updated 10 months ago