RobinQu / instinct.cppLinks
instinct.cpp provides ready to use alternatives to OpenAI Assistant API and built-in utilities for developing AI Agent applications (RAG, Chatbot, Code interpreter) powered by language models. Call it langchain.cpp if you like.
☆57Updated last year
Alternatives and similar repositories for instinct.cpp
Users that are interested in instinct.cpp are comparing it to the libraries listed below
Sorting:
- GGML implementation of BERT model with Python bindings and quantization.☆58Updated last year
- ggml implementation of embedding models including SentenceTransformer and BGE☆63Updated 2 years ago
- General purpose GPU compute framework built on Vulkan to support 1000s of cross vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). …☆51Updated 11 months ago
- LLM based agents with proactive interactions, long-term memory, external tool integration, and local deployment capabilities.☆107Updated 6 months ago
- cortex.llamacpp is a high-efficiency C++ inference engine for edge computing. It is a dynamic library that can be loaded by any server a…☆42Updated 7 months ago
- Inference of Mamba and Mamba2 models in pure C☆196Updated 2 weeks ago
- ggml implementation of BERT Embedding☆26Updated 2 years ago
- Inference of Large Multimodal Models in C/C++. LLaVA and others☆48Updated 2 years ago
- Lightweight C inference for Qwen3 GGUF. Multiturn prefix caching & batch processing.☆21Updated 5 months ago
- ☆70Updated 2 years ago
- LLM inference in C/C++☆104Updated last week
- Python bindings for ggml☆147Updated last year
- ☆51Updated last year
- 🚀 Scale your RAG pipeline using Ragswift: A scalable centralized embeddings management platform☆38Updated 2 years ago
- EdgeInfer enables efficient edge intelligence by running small AI models, including embeddings and OnnxModels, on resource-constrained de…☆50Updated last year
- Self-host LLMs with vLLM and BentoML☆168Updated 2 weeks ago
- AirLLM 70B inference with single 4GB GPU☆17Updated 7 months ago
- GPT-4 Level Conversational QA Trained In a Few Hours☆65Updated last year
- Thin wrapper around GGML to make life easier☆42Updated 3 months ago
- High-Performance Text Deduplication Toolkit☆61Updated 5 months ago
- A minimalistic C++ Jinja templating engine for LLM chat templates☆202Updated 4 months ago
- Deployment a light and full OpenAI API for production with vLLM to support /v1/embeddings with all embeddings models.☆44Updated last year
- Open Source Text Embedding Models with OpenAI Compatible API☆167Updated last year
- LLaVA server (llama.cpp).☆183Updated 2 years ago
- A chat UI for Llama.cpp☆15Updated 2 months ago
- A memory framework for Large Language Models and Agents.☆181Updated last year
- A collection of all available inference solutions for the LLMs☆94Updated 11 months ago
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated last year
- Local ML voice chat using high-end models.☆182Updated last month
- Port of Microsoft's BioGPT in C/C++ using ggml☆85Updated last year