RWKV-APP / RWKV_APPLinks
A fast, lightweight, and extensible RWKV chat UI powered by Flutter. Offline-ready, multi-backend support, ideal for local RWKV inference.
☆63Updated this week
Alternatives and similar repositories for RWKV_APP
Users that are interested in RWKV_APP are comparing it to the libraries listed below
Sorting:
- An educational Rust project for exporting and running inference on Qwen3 LLM family☆40Updated 6 months ago
- Inference RWKV v7 in pure C.☆44Updated 3 months ago
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆156Updated 7 months ago
- RWKV-LM-V7(https://github.com/BlinkDL/RWKV-LM) Under Lightning Framework☆55Updated last month
- Transplants vocabulary between language models, enabling the creation of draft models for speculative decoding WITHOUT retraining.☆49Updated 3 months ago
- A Pure Rust based LLM, VLM, VLA, TTS, OCR Inference Engine, powering by Candle & Rust. Alternate to your llama.cpp but much more simpler …☆241Updated last week
- REAP: Router-weighted Expert Activation Pruning for SMoE compression☆222Updated last month
- Sparse Inferencing for transformer based LLMs☆218Updated 5 months ago
- Implementation of the RWKV language model in pure WebGPU/Rust.☆338Updated 3 weeks ago
- Thin wrapper around GGML to make life easier☆42Updated 3 months ago
- Inference RWKV with multiple supported backends.☆77Updated this week
- ☆169Updated 3 weeks ago
- Mini Model Daemon☆12Updated last year
- TTS support with GGML☆218Updated 4 months ago
- Course Project for COMP4471 on RWKV☆17Updated last year
- Create 3D files in the CLI with Small Language Model☆43Updated 3 months ago
- Efficient RWKV inference engine. RWKV7 7.2B fp16 decoding 10250 tps @ single 5090.☆77Updated this week
- Review/Check GGUF files and estimate the memory usage and maximum tokens per second.☆238Updated last month
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆88Updated this week
- ☆90Updated last month
- ☆135Updated 9 months ago
- Tiny Model, Big Logic: Diversity-Driven Optimization Elicits Large-Model Reasoning Ability in VibeThinker-1.5B☆569Updated 2 months ago
- Running Microsoft's BitNet via Electron, React & Astro☆52Updated 4 months ago
- ☆109Updated 5 months ago
- 1.58-bit LLaMa model☆82Updated last year
- RWKV models and examples powered by candle.☆24Updated 2 weeks ago
- Prepare for DeekSeek R1 inference: Benchmark CPU, DRAM, SSD, iGPU, GPU, ... with efficient code.☆74Updated last year
- LLM inference in C/C++☆104Updated last week
- 1.58 Bit LLM on Apple Silicon using MLX☆242Updated last year
- automatically quant GGUF models☆219Updated last month