EricLBuehler / mistral.rsLinks
Blazingly fast LLM inference.
☆6,149Updated this week
Alternatives and similar repositories for mistral.rs
Users that are interested in mistral.rs are comparing it to the libraries listed below
Sorting:
- Deep learning at the speed of light.☆2,564Updated last week
- Local AI API Platform☆2,758Updated 3 months ago
- [Unmaintained, see README] An ecosystem of Rust libraries for working with large language models☆6,134Updated last year
- Instant, controllable, local pre-trained AI models in Rust☆2,041Updated this week
- Bionic is an on-premise replacement for ChatGPT, offering the advantages of Generative AI while maintaining strict data confidentiality☆2,265Updated this week
- Minimalist ML framework for Rust☆18,341Updated this week
- A vector search SQLite extension that runs anywhere!☆6,300Updated 9 months ago
- Distributed LLM and StableDiffusion inference for mobile, desktop and server.☆2,886Updated last year
- Developer-friendly, embedded retrieval engine for multimodal AI. Search More; Manage Less.☆7,762Updated last week
- Burn is a next generation Deep Learning Framework that doesn't compromise on flexibility, efficiency and portability.☆13,140Updated this week
- Open-source LLM load balancer and serving platform for self-hosting LLMs at scale 🏓🦙☆1,337Updated 2 weeks ago
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆1,516Updated last week
- Run PyTorch LLMs locally on servers, desktop and mobile☆3,615Updated last month
- A blazing fast inference solution for text embeddings models☆4,103Updated 2 weeks ago
- Cohere Toolkit is a collection of prebuilt components enabling users to quickly build and deploy RAG applications.☆3,133Updated 2 weeks ago
- AICI: Prompts as (Wasm) Programs☆2,050Updated 9 months ago
- Fast ML inference & training for ONNX models in Rust☆1,642Updated last week
- A fast llama2 decoder in pure Rust.☆1,052Updated last year
- 🦜️🔗LangChain for Rust, the easiest way to write LLM-based programs in Rust☆1,116Updated 3 weeks ago
- Distributed LLM inference. Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,705Updated 2 weeks ago
- ⚙️🦀 Build modular and scalable LLM Applications in Rust☆4,717Updated this week
- Examples in the MLX framework☆7,926Updated 2 weeks ago
- Inference Llama 2 in one file of pure 🔥☆2,118Updated last week
- peer-2-peer that just works☆7,045Updated last week
- Rust bindings for the C++ api of PyTorch.☆5,073Updated last week
- the terminal client for Ollama☆2,219Updated last week
- Minimal LLM inference in Rust☆1,013Updated last year
- Large Action Model framework to develop AI Web Agents☆6,184Updated 9 months ago
- Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs☆3,518Updated 5 months ago
- A simple and easy-to-use library for interacting with the Ollama API.☆924Updated last week