reinterpretcat / qwen3-rsLinks
An educational Rust project for exporting and running inference on Qwen3 LLM family
☆25Updated this week
Alternatives and similar repositories for qwen3-rs
Users that are interested in qwen3-rs are comparing it to the libraries listed below
Sorting:
- AI Assistant☆20Updated 3 months ago
- Lightweight C inference for Qwen3 GGUF with the smallest (0.6B) at the fullest (FP32)☆15Updated last week
- Light WebUI for lm.rs☆24Updated 9 months ago
- A Pure Rust based LLM (Any LLM based MLLM such as Spark-TTS) Inference Engine, powering by Candle framework.☆144Updated 2 weeks ago
- Super-simple, fully Rust powered "memory" (doc store + semantic search) for LLM projects, semantic search, etc.☆62Updated last year
- *NIX SHELL with Local AI/LLM integration☆23Updated 5 months ago
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Face☆37Updated last year
- Rust implementation of Surya☆58Updated 5 months ago
- Implementing the BitNet model in Rust☆38Updated last year
- Built for demanding AI workflows, this gateway offers low-latency, provider-agnostic access, ensuring your AI applications run smoothly a…☆74Updated 2 months ago
- A simple no-install web UI for Ollama and OAI-Compatible APIs!☆31Updated 6 months ago
- ☆24Updated 6 months ago
- Official Rust Implementation of Model2Vec☆123Updated last month
- git-like rag pipeline☆237Updated last week
- Chat WebUI is an easy-to-use user interface for interacting with AI, and it comes with multiple useful built-in tools.☆32Updated last month
- 33B Chinese LLM, DPO QLORA, 100K context, AirLLM 70B inference with single 4GB GPU☆13Updated last year
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆80Updated last year
- Inference Llama 2 in one file of zero-dependency, zero-unsafe Rust☆38Updated 2 years ago
- AirLLM 70B inference with single 4GB GPU☆14Updated last month
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆90Updated last month
- Fast serverless LLM inference, in Rust.☆88Updated 5 months ago
- A Fish Speech implementation in Rust, with Candle.rs☆94Updated 2 months ago
- A relatively basic implementation of RWKV in Rust written by someone with very little math and ML knowledge. Supports 32, 8 and 4 bit eva…☆94Updated last year
- A simple and clear way of hosting llama.cpp as a private HTTP API using Rust☆27Updated last year
- fast state-of-the-art speech models and a runtime that runs anywhere 💥☆55Updated last month
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆98Updated last month
- a lightweight, open-source blueprint for building powerful and scalable LLM chat applications☆28Updated last year
- JavaScript bindings for the ggml-js library☆43Updated 4 months ago
- Rust bindings for OpenNMT/CTranslate2☆36Updated 2 weeks ago
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibes☆222Updated this week