Maximilian-Winter / llama-cpp-agentLinks
The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM models, execute structured function calls and get structured output. Works also with models not fine-tuned to JSON output and function calls.
☆573Updated 4 months ago
Alternatives and similar repositories for llama-cpp-agent
Users that are interested in llama-cpp-agent are comparing it to the libraries listed below
Sorting:
- ☆900Updated 9 months ago
- A multimodal, function calling powered LLM webui.☆214Updated 8 months ago
- An application for running LLMs locally on your device, with your documents, facilitating detailed citations in generated responses.☆597Updated 7 months ago
- Web UI for ExLlamaV2☆501Updated 4 months ago
- A fast batching API to serve LLM models☆183Updated last year
- Large-scale LLM inference engine☆1,453Updated this week
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆987Updated this week
- Dataset Crafting w/ RAG/Wikipedia ground truth and Efficient Fine-Tuning Using MLX and Unsloth. Includes configurable dataset annotation …☆185Updated 11 months ago
- function calling-based LLM agents☆288Updated 9 months ago
- The RunPod worker template for serving our large language model endpoints. Powered by vLLM.☆322Updated last week
- An AI assistant beyond the chat box.☆328Updated last year
- TheBloke's Dockerfiles☆304Updated last year
- Efficient visual programming for AI language models☆363Updated last month
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆154Updated last year
- This is our own implementation of 'Layer Selective Rank Reduction'☆239Updated last year
- Your Trusty Memory-enabled AI Companion - Simple RAG chatbot optimized for local LLMs | 12 Languages Supported | OpenAI API Compatible☆316Updated 3 months ago
- A python package for developing AI applications with local LLMs.☆150Updated 5 months ago
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆222Updated last year
- 🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.☆381Updated last month
- FastMLX is a high performance production ready API to host MLX models.☆308Updated 3 months ago
- Pure C++ implementation of several models for real-time chatting on your computer (CPU & GPU)☆627Updated this week
- Simple Python library/structure to ablate features in LLMs which are supported by TransformerLens☆478Updated last year
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆255Updated 3 months ago
- Stateful load balancer custom-tailored for llama.cpp 🏓🦙☆779Updated this week
- Convenience scripts to finetune (chat-)LLaMa3 and other models for any language☆309Updated last year
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆162Updated last year
- AlwaysReddy is a LLM voice assistant that is always just a hotkey away.☆744Updated 3 months ago
- ☆203Updated last month
- Low-Rank adapter extraction for fine-tuned transformers models☆173Updated last year
- A tool for generating function arguments and choosing what function to call with local LLMs☆427Updated last year