reinterpretcat / qwen3-rsLinks
An educational Rust project for exporting and running inference on Qwen3 LLM family
☆38Updated 5 months ago
Alternatives and similar repositories for qwen3-rs
Users that are interested in qwen3-rs are comparing it to the libraries listed below
Sorting:
- *NIX SHELL with Local AI/LLM integration☆24Updated 10 months ago
- AI Assistant☆20Updated 8 months ago
- Light WebUI for lm.rs☆24Updated last year
- Super-simple, fully Rust powered "memory" (doc store + semantic search) for LLM projects, semantic search, etc.☆64Updated 2 years ago
- A Pure Rust based LLM (Any LLM based MLLM such as Spark-TTS) Inference Engine, powering by Candle framework.☆222Updated last week
- Yet another `llama.cpp` Rust wrapper☆12Updated last year
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Face☆46Updated last year
- Rust implementation of VibeVoice text-to-speech with voice cloning and multi-speaker synthesis.☆49Updated 2 weeks ago
- Official Rust Implementation of Model2Vec☆145Updated 3 months ago
- AirLLM 70B inference with single 4GB GPU☆14Updated 6 months ago
- Rust implementation of Surya☆64Updated 10 months ago
- git-like rag pipeline☆251Updated 2 weeks ago
- A Fish Speech implementation in Rust, with Candle.rs☆106Updated 7 months ago
- A high-performance BPE tokenizer built with Rust with Python bindings, focused on speed, safety, and resource optimization.☆55Updated 2 weeks ago
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆99Updated 6 months ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆86Updated this week
- Kroko ASR - Speech-to-text☆126Updated 3 months ago
- fast state-of-the-art speech models and a runtime that runs anywhere 💥☆57Updated 6 months ago
- ☆24Updated 11 months ago
- Fast serverless LLM inference, in Rust.☆108Updated 2 months ago
- ⚡️Lightning fast in-memory VectorDB written in rust🦀☆28Updated 10 months ago
- powerful and fast tool calling agents☆79Updated 9 months ago
- A simple no-install web UI for Ollama and OAI-Compatible APIs!☆31Updated 11 months ago
- Orchestrator Kit for Agentic Reasoning - OrKa is a modular AI orchestration system that transforms Large Language Models (LLMs) into comp…☆83Updated this week
- Rust standalone inference of Namo-500M series models. Extremly tiny, runing VLM on CPU.☆24Updated 9 months ago
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆151Updated 6 months ago
- TensorRT-LLM server with Structured Outputs (JSON) built with Rust☆65Updated 8 months ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- 33B Chinese LLM, DPO QLORA, 100K context, AirLLM 70B inference with single 4GB GPU☆13Updated last year
- Library for doing RAG☆80Updated last week