mostlygeek / llama-swapLinks
Model swapping for llama.cpp (or any local OpenAPI compatible server)
☆1,010Updated last week
Alternatives and similar repositories for llama-swap
Users that are interested in llama-swap are comparing it to the libraries listed below
Sorting:
- What If Language Models Expertly Routed All Inference? WilmerAI allows prompts to be routed to specialized workflows based on the domain …☆721Updated last week
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆997Updated last week
- Manifold is a platform for enabling workflow automation using AI assistants.☆448Updated this week
- llama.cpp fork with additional SOTA quants and improved performance☆652Updated this week
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆430Updated this week
- LLM Frontend in a single html file☆508Updated 5 months ago
- ☆205Updated 2 months ago
- Effortlessly run LLM backends, APIs, frontends, and services with one command.☆1,906Updated this week
- ☆600Updated this week
- Large-scale LLM inference engine☆1,471Updated this week
- The AI toolkit for the AI developer☆777Updated this week
- A multi-platform desktop application to evaluate and compare LLM models, written in Rust and React.☆776Updated 2 months ago
- VS Code extension for LLM-assisted code/text completion☆835Updated last week
- The Fastest Way to Fine-Tune LLMs Locally☆311Updated 3 months ago
- Web UI for ExLlamaV2☆503Updated 5 months ago
- Your Trusty Memory-enabled AI Companion - Simple RAG chatbot optimized for local LLMs | 12 Languages Supported | OpenAI API Compatible☆319Updated 4 months ago
- Local LLM Powered Recursive Search & Smart Knowledge Explorer☆244Updated 5 months ago
- Code execution utilities for Open WebUI & Ollama☆290Updated 8 months ago
- An application for running LLMs locally on your device, with your documents, facilitating detailed citations in generated responses.☆600Updated 8 months ago
- An OpenAI API compatible text to speech server using Coqui AI's xtts_v2 and/or piper tts as the backend.☆791Updated 5 months ago
- Big & Small LLMs working together☆1,058Updated this week
- An AI memory layer with short- and long-term storage, semantic clustering, and optional memory decay for context-aware applications.☆639Updated 5 months ago
- Easy to use interface for the Whisper model optimized for all GPUs!☆234Updated 2 weeks ago
- Collection of LLM system prompts.☆122Updated last month
- A proxy server for multiple ollama instances with Key security☆461Updated this week
- Chat with your current directory's files using a local or API LLM.☆382Updated 2 weeks ago
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆575Updated 4 months ago
- 🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.☆385Updated 2 months ago
- Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,209Updated this week
- Stateful load balancer custom-tailored for llama.cpp 🏓🦙☆790Updated this week