reinterpretcat / qwen3-rsLinks
An educational Rust project for exporting and running inference on Qwen3 LLM family
☆39Updated 5 months ago
Alternatives and similar repositories for qwen3-rs
Users that are interested in qwen3-rs are comparing it to the libraries listed below
Sorting:
- Light WebUI for lm.rs☆24Updated last year
- A Pure Rust based LLM (Any LLM based MLLM such as Spark-TTS) Inference Engine, powering by Candle framework.☆234Updated last week
- A Fish Speech implementation in Rust, with Candle.rs☆106Updated 7 months ago
- AI Assistant☆20Updated 9 months ago
- Yet another `llama.cpp` Rust wrapper☆12Updated last year
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated 2 years ago
- Super-simple, fully Rust powered "memory" (doc store + semantic search) for LLM projects, semantic search, etc.☆65Updated 2 years ago
- Built for demanding AI workflows, this gateway offers low-latency, provider-agnostic access, ensuring your AI applications run smoothly a…☆88Updated 8 months ago
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Face☆47Updated last year
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedback☆114Updated 10 months ago
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆156Updated 6 months ago
- Rust implementation of Surya☆64Updated 11 months ago
- A high-performance BPE tokenizer built with Rust with Python bindings, focused on speed, safety, and resource optimization.☆54Updated last month
- ☆108Updated 2 months ago
- Implementing the BitNet model in Rust☆44Updated last year
- Rust bindings for OpenNMT/CTranslate2☆49Updated 3 weeks ago
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆100Updated 7 months ago
- Yet Another (LLM) Web UI, made with Gemini☆12Updated last year
- fast state-of-the-art speech models and a runtime that runs anywhere 💥☆57Updated 7 months ago
- A simple no-install web UI for Ollama and OAI-Compatible APIs!☆31Updated last year
- 33B Chinese LLM, DPO QLORA, 100K context, AirLLM 70B inference with single 4GB GPU☆13Updated last year
- Chat WebUI is an easy-to-use user interface for interacting with AI, and it comes with multiple useful built-in tools such as web search …☆48Updated 5 months ago
- Rust standalone inference of Namo-500M series models. Extremly tiny, runing VLM on CPU.☆24Updated 10 months ago
- A relatively basic implementation of RWKV in Rust written by someone with very little math and ML knowledge. Supports 32, 8 and 4 bit eva…☆94Updated 2 years ago
- ☆24Updated last year
- ☆27Updated last year
- *NIX SHELL with Local AI/LLM integration☆24Updated 11 months ago
- Orchestrator Kit for Agentic Reasoning - OrKa is a modular AI orchestration system that transforms Large Language Models (LLMs) into comp…☆86Updated 3 weeks ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆88Updated this week
- Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllm…☆146Updated 2 months ago