reinterpretcat / qwen3-rsLinks
An educational Rust project for exporting and running inference on Qwen3 LLM family
☆27Updated 3 weeks ago
Alternatives and similar repositories for qwen3-rs
Users that are interested in qwen3-rs are comparing it to the libraries listed below
Sorting:
- Super-simple, fully Rust powered "memory" (doc store + semantic search) for LLM projects, semantic search, etc.☆62Updated last year
- AI Assistant☆20Updated 4 months ago
- Light WebUI for lm.rs☆24Updated 10 months ago
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆102Updated last month
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Face☆37Updated last year
- AirLLM 70B inference with single 4GB GPU☆14Updated 2 months ago
- Lightweight C inference for Qwen3 GGUF with the smallest (0.6B) at the fullest (FP32)☆16Updated 2 weeks ago
- A Pure Rust based LLM (Any LLM based MLLM such as Spark-TTS) Inference Engine, powering by Candle framework.☆155Updated last month
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆80Updated last year
- Rust implementation of Surya☆60Updated 6 months ago
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedback☆103Updated 5 months ago
- *NIX SHELL with Local AI/LLM integration☆23Updated 6 months ago
- A Fish Speech implementation in Rust, with Candle.rs☆94Updated 2 months ago
- git-like rag pipeline☆243Updated this week
- A relatively basic implementation of RWKV in Rust written by someone with very little math and ML knowledge. Supports 32, 8 and 4 bit eva…☆94Updated last year
- Built for demanding AI workflows, this gateway offers low-latency, provider-agnostic access, ensuring your AI applications run smoothly a…☆73Updated 3 months ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆73Updated last week
- Official Rust Implementation of Model2Vec☆127Updated last week
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆93Updated 2 months ago
- fast state-of-the-art speech models and a runtime that runs anywhere 💥☆55Updated 2 months ago
- Thin wrapper around GGML to make life easier☆40Updated 2 months ago
- High-level, optionally asynchronous Rust bindings to llama.cpp☆228Updated last year
- Inference Llama 2 in one file of zero-dependency, zero-unsafe Rust☆38Updated 2 years ago
- Implementing the BitNet model in Rust☆39Updated last year
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibes☆230Updated 3 weeks ago
- 33B Chinese LLM, DPO QLORA, 100K context, AirLLM 70B inference with single 4GB GPU☆13Updated last year
- ⚡️Lightning fast in-memory VectorDB written in rust🦀☆24Updated 5 months ago
- OpenAI compatible API for serving LLAMA-2 model☆218Updated last year
- Proof of concept for a generative AI application framework powered by WebAssembly and Extism☆14Updated 2 years ago
- Rust bindings to https://github.com/k2-fsa/sherpa-onnx☆207Updated 3 months ago