iohub / collama
VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.
☆176Updated this week
Alternatives and similar repositories for collama:
Users that are interested in collama are comparing it to the libraries listed below
- transparent proxy server for llama.cpp's server to provide automatic model swapping☆135Updated this week
- AI Studio is an independent app for utilizing LLM.☆190Updated this week
- Dolphin System Messages☆226Updated 3 months ago
- On-device LLM Inference Powered by X-Bit Quantization☆203Updated this week
- Lightweight, standalone, multi-platform, and privacy focused local LLM chat interface with optional encryption☆78Updated 2 months ago
- automatically quant GGUF models☆150Updated this week
- MinimalChat is a lightweight, open-source chat application that allows you to interact with various large language models.☆150Updated 5 months ago
- Web UI for ExLlamaV2☆460Updated 3 weeks ago
- LM Studio JSON configuration file format and a collection of example config files.☆189Updated 5 months ago
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆217Updated last month
- Open source alternative to Perplexity AI with ability to run locally☆175Updated 3 months ago
- Parse files (e.g. code repos) and websites to clipboard or a file for ingestions by AI / LLMs☆112Updated last month
- Review/Check GGUF files and estimate the memory usage and maximum tokens per second.☆69Updated this week
- A multimodal, function calling powered LLM webui.☆213Updated 3 months ago
- ☆73Updated 3 weeks ago
- A fast batching API to serve LLM models☆177Updated 8 months ago
- 100% Local AGI with LocalAI☆431Updated 6 months ago
- ☆88Updated this week
- Serving LLMs in the HF-Transformers format via a PyFlask API☆68Updated 4 months ago
- An open-source VSCode extension, the AI coding assistant, integrates with Ollama, HuggingFace, OpenAI, and Anthropic.☆170Updated this week
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆113Updated 7 months ago
- Link you Ollama models to LM-Studio☆118Updated 6 months ago
- ☆181Updated last month
- Local ML voice chat using high-end models.☆155Updated last week
- An AI assistant beyond the chat box.☆317Updated 10 months ago
- An extension that lets the AI take the wheel, allowing it to use the mouse and keyboard, recognize UI elements, and prompt itself :3...no…☆103Updated 2 months ago
- Open source LLM UI, compatible with all local LLM providers.☆169Updated 3 months ago
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated 3 months ago
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆485Updated this week
- ☆121Updated this week