reinterpretcat / qwen3-rsLinks
An educational Rust project for exporting and running inference on Qwen3 LLM family
☆32Updated 2 months ago
Alternatives and similar repositories for qwen3-rs
Users that are interested in qwen3-rs are comparing it to the libraries listed below
Sorting:
- ☆66Updated last week
- AI Assistant☆20Updated 6 months ago
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Face☆46Updated last year
- Light WebUI for lm.rs☆24Updated last year
- A Pure Rust based LLM (Any LLM based MLLM such as Spark-TTS) Inference Engine, powering by Candle framework.☆171Updated last week
- Lightweight C inference for Qwen3 GGUF. Multiturn prefix caching & batch processing.☆17Updated 2 months ago
- git-like rag pipeline☆246Updated 3 weeks ago
- Orchestrator Kit for Agentic Reasoning - OrKa is a modular AI orchestration system that transforms Large Language Models (LLMs) into comp…☆51Updated last week
- Super-simple, fully Rust powered "memory" (doc store + semantic search) for LLM projects, semantic search, etc.☆62Updated 2 years ago
- *NIX SHELL with Local AI/LLM integration☆24Updated 8 months ago
- 33B Chinese LLM, DPO QLORA, 100K context, AirLLM 70B inference with single 4GB GPU☆13Updated last year
- Yet another `llama.cpp` Rust wrapper☆12Updated last year
- A Fish Speech implementation in Rust, with Candle.rs☆98Updated 4 months ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆81Updated this week
- ☆14Updated 2 months ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆140Updated 3 months ago
- Rust implementation of Surya☆61Updated 8 months ago
- A simple no-install web UI for Ollama and OAI-Compatible APIs!☆31Updated 9 months ago
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆98Updated 4 months ago
- Built for demanding AI workflows, this gateway offers low-latency, provider-agnostic access, ensuring your AI applications run smoothly a…☆79Updated 5 months ago
- AirLLM 70B inference with single 4GB GPU☆14Updated 4 months ago
- Implementing the BitNet model in Rust☆40Updated last year
- The MCP enterprise actors-based server or mcp-ectors for short☆31Updated 5 months ago
- Official Rust Implementation of Model2Vec☆139Updated last month
- fast state-of-the-art speech models and a runtime that runs anywhere 💥☆57Updated 4 months ago
- A relatively basic implementation of RWKV in Rust written by someone with very little math and ML knowledge. Supports 32, 8 and 4 bit eva…☆94Updated 2 years ago
- Inference Llama 2 in one file of zero-dependency, zero-unsafe Rust☆39Updated 2 years ago
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedback☆108Updated 7 months ago
- VT Code - CLI Semantic Coding Agent☆154Updated last week