janhq / cortex.llamacppLinks
cortex.llamacpp is a high-efficiency C++ inference engine for edge computing. It is a dynamic library that can be loaded by any server at runtime.
☆42Updated 7 months ago
Alternatives and similar repositories for cortex.llamacpp
Users that are interested in cortex.llamacpp are comparing it to the libraries listed below
Sorting:
- Lightweight C inference for Qwen3 GGUF. Multiturn prefix caching & batch processing.☆21Updated 5 months ago
- AirLLM 70B inference with single 4GB GPU☆17Updated 7 months ago
- TTS support with GGML☆218Updated 4 months ago
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated last year
- Yet Another (LLM) Web UI, made with Gemini☆12Updated last year
- ☆29Updated last year
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆88Updated last week
- Thin wrapper around GGML to make life easier☆42Updated 3 months ago
- A chat UI for Llama.cpp☆15Updated 2 months ago
- ☆109Updated 5 months ago
- Something similar to Apple Intelligence?☆60Updated last year
- Running Microsoft's BitNet via Electron, React & Astro☆52Updated 4 months ago
- Locally running LLM with internet access☆97Updated 7 months ago
- instinct.cpp provides ready to use alternatives to OpenAI Assistant API and built-in utilities for developing AI Agent applications (RAG,…☆57Updated last year
- Generate a llama-quantize command to copy the quantization parameters of any GGUF☆30Updated 2 weeks ago
- Experiments with BitNet inference on CPU☆55Updated last year
- Tcurtsni: Reverse Instruction Chat, ever wonder what your LLM wants to ask you?☆23Updated last year
- SPLAA is an AI assistant framework that utilizes voice recognition, text-to-speech, and tool-calling capabilities to provide a conversati…☆28Updated 9 months ago
- GGML implementation of BERT model with Python bindings and quantization.☆58Updated last year
- 33B Chinese LLM, DPO QLORA, 100K context, AirLLM 70B inference with single 4GB GPU☆13Updated last year
- ☆32Updated 2 years ago
- llama.cpp fork used by GPT4All☆55Updated 11 months ago
- ☆24Updated last year
- Port of Suno AI's Bark in C/C++ for fast inference☆54Updated last year
- A sleek, customizable interface for managing LLMs with responsive design and easy agent personalization.☆17Updated last year
- Spotlight-like client for Ollama on Windows.☆28Updated last year
- Light WebUI for lm.rs☆24Updated last year
- The hearth of The Pulsar App, fast, secure and shared inference with modern UI☆59Updated last year
- PowerShell automation to rebuild llama.cpp for a Windows environment.☆35Updated 3 weeks ago
- A ggml (C++) re-implementation of tortoise-tts☆193Updated last year