Fast, flexible LLM inference
☆6,713Mar 15, 2026Updated last week
Alternatives and similar repositories for mistral.rs
Users that are interested in mistral.rs are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Minimalist ML framework for Rust☆19,735Updated this week
- Burn is a next generation tensor library and Deep Learning Framework that doesn't compromise on flexibility, efficiency and portability.☆14,679Updated this week
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆614Updated this week
- [Unmaintained, see README] An ecosystem of Rust libraries for working with large language models☆6,152Jun 24, 2024Updated last year
- Qdrant - High-performance, massive-scale Vector Database and Vector Search Engine for the next generation of AI. Also available in the cl…☆29,790Updated this week
- ⚙️🦀 Build modular and scalable LLM Applications in Rust☆6,620Updated this week
- Deep learning at the speed of light.☆2,783Updated this week
- Instant, controllable, local pre-trained AI models in Rust☆2,158Updated this week
- A blazing fast inference solution for text embeddings models☆4,600Mar 13, 2026Updated last week
- Fullstack app framework for web, desktop, and mobile.☆35,326Mar 18, 2026Updated last week
- Distribute and run LLMs with a single file.☆23,859Updated this week
- Open Lakehouse Format for Multimodal AI. Convert from Parquet in 2 lines of code for 100x faster random access, vector index, and data ve…☆6,187Updated this week
- Build, run, manage agentic software at scale.☆38,835Updated this week
- Developer-friendly OSS embedded retrieval library for multimodal AI. Search More; Manage Less.☆9,536Mar 17, 2026Updated last week
- Tantivy is a full-text search engine library inspired by Apache Lucene and written in Rust☆14,778Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆74,135Updated this week
- Self-hosted AI coding assistant☆33,022Mar 2, 2026Updated 3 weeks ago
- screenpipe turns your computer into a personal AI that knows everything you've done. record. search. automate. all local, all private, al…☆17,389Updated this week
- Extremely fast Query Engine for DataFrames, written in Rust☆37,810Updated this week
- Unsloth Studio is a web UI for training and running open models like Qwen, DeepSeek, gpt-oss and Gemma locally.☆57,673Updated this week
- SGLang is a high-performance serving framework for large language models and multimodal models.☆24,829Updated this week
- Turso is an in-process SQL database, compatible with SQLite.☆17,933Updated this week
- Fast ML inference & training for ONNX models in Rust☆2,093Mar 16, 2026Updated last week
- LLM inference in C/C++☆98,911Updated this week
- Large Language Model Text Generation Inference☆10,812Jan 8, 2026Updated 2 months ago
- Cloud-native search engine for observability. An open-source alternative to Datadog, Elasticsearch, Loki, and Tempo.☆11,002Updated this week
- Python SDK, Proxy Server (AI Gateway) to call 100+ LLM APIs in OpenAI (or native) format, with cost tracking, guardrails, loadbalancing a…☆39,597Updated this week
- A cross-platform browser ML framework.☆750Nov 23, 2024Updated last year
- 🚂 🦀 The one-person framework for Rust for side-projects and startups☆8,774Mar 4, 2026Updated 2 weeks ago
- `llm-chain` is a powerful rust crate for building chains in large language models allowing you to summarise text and complete complex tas…☆1,597Oct 31, 2024Updated last year
- 20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.☆13,256Updated this week
- An open source SDK for logging, storing, querying, and visualizing multimodal and multi-rate data☆10,403Updated this week
- An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.☆28,028Sep 30, 2025Updated 5 months ago
- Build fast web applications with Rust.☆20,458Updated this week
- Vane is an AI-powered answering engine.☆33,329Mar 10, 2026Updated 2 weeks ago
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆4,468Mar 4, 2026Updated 3 weeks ago
- Minimal LLM inference in Rust☆1,035Oct 24, 2024Updated last year
- A library for building fast, reliable and evolvable network services.☆26,286Updated this week
- Rust library for vector embeddings and reranking.☆805Mar 17, 2026Updated last week