High-level, optionally asynchronous Rust bindings to llama.cpp
☆243Jun 5, 2024Updated last year
Alternatives and similar repositories for llama_cpp-rs
Users that are interested in llama_cpp-rs are comparing it to the libraries listed below
Sorting:
- LLama.cpp rust bindings☆414Jun 27, 2024Updated last year
- ☆468Updated this week
- Yet another `llama.cpp` Rust wrapper☆12Jun 19, 2024Updated last year
- ☆13Nov 4, 2023Updated 2 years ago
- Rust bindings to https://github.com/leejet/stable-diffusion.cpp☆37Updated this week
- Rust library for vector embeddings and reranking.☆780Feb 23, 2026Updated last week
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibes☆244Aug 6, 2025Updated 6 months ago
- Rust bindings to https://github.com/ggerganov/whisper.cpp☆933Jul 30, 2025Updated 7 months ago
- A simple and easy-to-use library for interacting with the Ollama API.☆989Feb 23, 2026Updated last week
- llm_utils: Basic LLM tools, best practices, and minimal abstraction.☆48Feb 18, 2025Updated last year
- asynchronous/distributed speculative evaluation for llama3☆40Aug 8, 2024Updated last year
- Rust client for the huggingface hub aiming for minimal subset of features over `huggingface-hub` python package☆265Feb 19, 2026Updated last week
- Fast ML inference & training for ONNX models in Rust☆2,021Updated this week
- ⚡ Edgen: Local, private GenAI server alternative to OpenAI. No GPU required. Run AI models locally: LLMs (Llama2, Mistral, Mixtral...), …☆369May 23, 2024Updated last year
- Rust bindings to https://github.com/k2-fsa/sherpa-onnx☆297Nov 1, 2025Updated 4 months ago
- Fast, flexible LLM inference☆6,623Updated this week
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆1,003Dec 17, 2025Updated 2 months ago
- A cross-platform browser ML framework.☆747Nov 23, 2024Updated last year
- pyannote audio diarization in rust☆103Sep 7, 2025Updated 5 months ago
- Speexdsp bindings and pure-rust implementation☆25Feb 2, 2026Updated last month
- Minimalist ML framework for Rust☆19,509Updated this week
- Spacedrive native dependencies☆13Apr 8, 2025Updated 10 months ago
- ☆24Jan 22, 2025Updated last year
- LLM inference in C/C++☆23Oct 4, 2024Updated last year
- a GGUF file parser☆16Updated this week
- YOLOv5 Inference with ONNX & OpenCV in Rust☆14Dec 26, 2023Updated 2 years ago
- ☆11Jun 22, 2023Updated 2 years ago
- Rust+OpenCL+AVX2 implementation of LLaMA inference code☆554Feb 12, 2024Updated 2 years ago
- A fast llama2 decoder in pure Rust.☆1,060Nov 30, 2023Updated 2 years ago
- Rust bindings to Apple media frameworks☆36Feb 2, 2026Updated last month
- A simple and clear way of hosting llama.cpp as a private HTTP API using Rust☆27Jun 22, 2024Updated last year
- Inference of Mamba and Mamba2 models in pure C☆197Jan 22, 2026Updated last month
- An implementation of the diffusers api in Rust☆586Apr 4, 2024Updated last year
- llama.cpp bindings and utilities for zig☆68Apr 14, 2025Updated 10 months ago
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Face☆47May 3, 2024Updated last year
- Rust bindings for OpenNMT/CTranslate2☆49Feb 7, 2026Updated 3 weeks ago
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆603Updated this week
- WebAssembly (Wasm) Build and Bindings for llama.cpp☆285Jul 23, 2024Updated last year
- [Unmaintained, see README] An ecosystem of Rust libraries for working with large language models☆6,150Jun 24, 2024Updated last year