mostlygeek / llama-swap
HTTP proxy for on-demand model loading with llama.cpp (or other OpenAI compatible backends)
☆41Updated this week
Related projects ⓘ
Alternatives and complementary repositories for llama-swap
- Easily view and modify JSON datasets for large language models☆62Updated last month
- Large Model Proxy is designed to make it easy to run multiple resource-heavy Large Models (LM) on the same machine with limited amount of…☆47Updated last month
- Something similar to Apple Intelligence?☆57Updated 4 months ago
- klmbr - a prompt pre-processing technique to break through the barrier of entropy while generating text with LLMs☆58Updated last month
- ☆128Updated this week
- 4 million public stable diffusion prompts -- interactive neural search and llama chat☆19Updated 2 months ago
- Serving LLMs in the HF-Transformers format via a PyFlask API☆68Updated 2 months ago
- A bot that checks your grammar and phrasing using LLM of choice☆27Updated 5 months ago
- idea: https://github.com/nyxkrage/ebook-groupchat/☆82Updated 3 months ago
- Realtime tts reading of large textfiles by your favourite voice. +Translation via LLM (Python script)☆47Updated last month
- Cohere Toolkit is a collection of prebuilt components enabling users to quickly build and deploy RAG applications.☆25Updated this week
- A frontend for creative writing with LLMs☆108Updated 4 months ago
- An extension that lets the AI take the wheel, allowing it to use the mouse and keyboard, recognize UI elements, and prompt itself :3...no…☆96Updated 3 weeks ago
- After my server ui improvements were successfully merged, consider this repo a playground for experimenting, tinkering and hacking around…☆56Updated 3 months ago
- AnyModal is a Flexible Multimodal Language Model Framework☆40Updated this week
- ☆25Updated last month
- ☆30Updated 6 months ago
- Gradio based tool to run opensource LLM models directly from Huggingface☆87Updated 4 months ago
- This small API downloads and exposes access to NeuML's txtai-wikipedia and full wikipedia datasets, taking in a query and returning full …☆51Updated this week
- A python application that routes incoming prompts to an LLM by category, and can support a single incoming connection from a front end to…☆167Updated this week
- A fast batching API to serve LLM models☆172Updated 6 months ago
- Local LLM inference & management server with built-in OpenAI API☆31Updated 7 months ago
- The hearth of The Pulsar App, fast, secure and shared inference with modern UI☆35Updated 3 weeks ago
- ☆18Updated 3 weeks ago
- Experimental LLM Inference UX to aid in creative writing☆106Updated 4 months ago
- ☆112Updated this week
- Creates an index of images, queries a local LLM and adds tags to the image metadata☆71Updated last month
- 5X faster 60% less memory QLoRA finetuning☆21Updated 5 months ago
- Open source LLM UI, compatible with all local LLM providers.☆167Updated 2 months ago
- A local AI companion that uses a collection of free, open source AI models in order to create two virtual companions that will follow you…☆112Updated this week