LLaVA server (llama.cpp).
☆183Oct 20, 2023Updated 2 years ago
Alternatives and similar repositories for llava-cpp-server
Users that are interested in llava-cpp-server are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- 生成训练文本检测数据集☆12Jul 1, 2020Updated 5 years ago
- iterate quickly with llama.cpp hot reloading. use the llama.cpp bindings with bun.sh☆50Oct 30, 2023Updated 2 years ago
- A simple "Be My Eyes" web app with a llama.cpp/llava backend☆494Nov 28, 2023Updated 2 years ago
- Port of Suno AI's Bark in C/C++ for fast inference☆55Apr 15, 2024Updated last year
- ☆1,274Oct 24, 2023Updated 2 years ago
- Simple, predictable pricing with DigitalOcean hosting • AdAlways know what you'll pay with monthly caps and flat pricing. Enterprise-grade infrastructure trusted by 600k+ customers.
- CLIP inference in plain C/C++ with no extra dependencies☆557Jun 19, 2025Updated 9 months ago
- Suno AI's Bark model in C/C++ for fast text-to-speech generation☆854Nov 16, 2024Updated last year
- Port of Microsoft's BioGPT in C/C++ using ggml☆86Feb 21, 2024Updated 2 years ago
- Semantic emoji finder. Python/dash UI. Uses sentence transformer embeddings and duckdb☆19Sep 15, 2025Updated 6 months ago
- Inference of Large Multimodal Models in C/C++. LLaVA and others☆48Oct 1, 2023Updated 2 years ago
- Inference Vision Transformer (ViT) in plain C/C++ with ggml☆313Apr 11, 2024Updated 2 years ago
- The Codec 2 speech codec, compiled to WASM using Emscripten.☆13Apr 27, 2023Updated 2 years ago
- Diffusion model(SD,Flux,Wan,Qwen Image,Z-Image,...) inference in pure C/C++☆5,726Updated this week
- Fine-tuning, DPO, RLHF, RLAIF on LLMs - Qwen3, Zephyr 7B GPTQ with 4-Bit Quantization, Mistral-7B-GPTQ☆15Jul 5, 2025Updated 9 months ago
- Simple, predictable pricing with DigitalOcean hosting • AdAlways know what you'll pay with monthly caps and flat pricing. Enterprise-grade infrastructure trusted by 600k+ customers.
- Implementation of "LM-Infinite: Simple On-the-Fly Length Generalization for Large Language Models"☆40Nov 11, 2024Updated last year
- GPT-2 small trained on phi-like data☆68Feb 18, 2024Updated 2 years ago
- Demo python script app to interact with llama.cpp server using whisper API, microphone and webcam devices.☆46Nov 6, 2023Updated 2 years ago
- Friendly Terminal Assistant for Developers☆17Mar 23, 2024Updated 2 years ago
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆569Aug 8, 2023Updated 2 years ago
- Fine-tune mistral-7B on 3090s, a100s, h100s☆724Oct 11, 2023Updated 2 years ago
- A Javascript library (with Typescript types) to parse metadata of GGML based GGUF files.☆52Jul 30, 2024Updated last year
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆626Mar 9, 2026Updated last month
- Python bindings for llama.cpp☆10,147Apr 5, 2026Updated last week
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- Tensor library for machine learning☆274Apr 23, 2023Updated 2 years ago
- ☆12Jan 25, 2023Updated 3 years ago
- transformer tokenizers (e.g. BERT tokenizer) in C++ (WIP)☆18Apr 7, 2022Updated 4 years ago
- LLM-based code completion engine☆191Jan 23, 2025Updated last year
- ☆135Nov 24, 2023Updated 2 years ago
- Extracts structured data from unstructured input. Programming language agnostic. Uses llama.cpp☆45May 16, 2024Updated last year
- High accuracy code-switching whisper / qwen3 transcription☆25Updated this week
- Distributed LLM inference. Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,892Feb 10, 2026Updated 2 months ago
- GGML implementation of BERT model with Python bindings and quantization.☆57Feb 19, 2024Updated 2 years ago
- GPU virtual machines on DigitalOcean Gradient AI • AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- This repo is for handling Question Answering, especially for Multi-hop Question Answering☆69Dec 20, 2023Updated 2 years ago
- Python bindings for the Transformer models implemented in C/C++ using GGML library.☆1,883Jan 28, 2024Updated 2 years ago
- ☆15Sep 8, 2023Updated 2 years ago
- Visual Studio Code extension for WizardCoder☆148Aug 1, 2023Updated 2 years ago
- [ICML 2025] EffiCoder: Enhancing Code Generation in Large Language Models through Efficiency-Aware Fine-tuning☆16May 24, 2025Updated 10 months ago
- Guess the Hacker News titles☆12Mar 24, 2022Updated 4 years ago
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆4,493Mar 4, 2026Updated last month