oxideai / mlx-rsLinks
Unofficial Rust bindings to Apple's mlx framework
☆157Updated this week
Alternatives and similar repositories for mlx-rs
Users that are interested in mlx-rs are comparing it to the libraries listed below
Sorting:
- Low rank adaptation (LoRA) for Candle.☆147Updated last month
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibes☆201Updated 3 months ago
- Andrej Karpathy's Let's build GPT: from scratch video & notebook implemented in Rust + candle☆71Updated last year
- Use multiple LLM backends in a single crate, simple builder-based configuration, and built-in prompt chaining & templating.☆130Updated last week
- Llama2 LLM ported to Rust burn☆277Updated last year
- ☆29Updated 6 months ago
- Structured outputs for LLMs☆46Updated 10 months ago
- Experimental compiler for deep learning models☆67Updated last week
- ☆129Updated last year
- Rust client for the huggingface hub aiming for minimal subset of features over `huggingface-hub` python package☆207Updated 3 months ago
- High-level, optionally asynchronous Rust bindings to llama.cpp☆222Updated 11 months ago
- Example of tch-rs on M1☆53Updated last year
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆80Updated last year
- A lightweight, high-performance text embedding model implemented in Rust.☆66Updated 3 weeks ago
- Library for doing RAG☆72Updated 2 weeks ago
- Inference Llama 2 in one file of pure Rust 🦀☆232Updated last year
- Tutorial for Porting PyTorch Transformer Models to Candle (Rust)☆296Updated 10 months ago
- LLM Orchestrator built in Rust☆276Updated last year
- A powerful Rust library and CLI tool to unify and orchestrate multiple LLM and voice backends (OpenAI, Claude, Gemini, Ollama, ElevenLabs…☆112Updated this week
- Extract core logic from qdrant and make it available as a library.☆58Updated last year
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆372Updated this week
- ONNX neural network inference engine☆210Updated this week
- pgvector support for Rust☆158Updated last week
- A set of Rust macros for working with OpenAI function/tool calls.☆49Updated last year
- Run Generative AI models directly on your hardware☆36Updated 9 months ago
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedback☆92Updated 2 months ago
- LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed!☆104Updated last year
- allms: One Rust Library to rule them aLLMs☆81Updated this week
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Face☆37Updated last year
- Rust SDK for the Model Context Protocol (MCP)☆128Updated 6 months ago