RWKV-APP / RWKV_APPLinks
A fast, lightweight, and extensible RWKV chat UI powered by Flutter. Offline-ready, multi-backend support, ideal for local RWKV inference.
☆63Updated this week
Alternatives and similar repositories for RWKV_APP
Users that are interested in RWKV_APP are comparing it to the libraries listed below
Sorting:
- An educational Rust project for exporting and running inference on Qwen3 LLM family☆40Updated 6 months ago
- A Pure Rust based LLM, VLM, VLA, TTS, OCR Inference Engine, powering by Candle & Rust. Alternate to your llama.cpp but much more simpler …☆244Updated last week
- RWKV-LM-V7(https://github.com/BlinkDL/RWKV-LM) Under Lightning Framework☆56Updated last month
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆157Updated 7 months ago
- Implementation of the RWKV language model in pure WebGPU/Rust.☆340Updated last month
- Sparse Inferencing for transformer based LLMs☆217Updated 6 months ago
- Inference RWKV v7 in pure C.☆44Updated 4 months ago
- TTS support with GGML☆218Updated 4 months ago
- REAP: Router-weighted Expert Activation Pruning for SMoE compression☆232Updated 2 months ago
- Transplants vocabulary between language models, enabling the creation of draft models for speculative decoding WITHOUT retraining.☆49Updated 3 months ago
- Efficient RWKV inference engine. RWKV7 7.2B fp16 decoding 10250 tps @ single 5090.☆78Updated last week
- Thin wrapper around GGML to make life easier☆42Updated 3 months ago
- ☆171Updated 3 weeks ago
- Tiny Model, Big Logic: Diversity-Driven Optimization Elicits Large-Model Reasoning Ability in VibeThinker-1.5B☆573Updated 2 months ago
- Modified Mamba code to run on CPU☆30Updated 2 years ago
- SVGBench: A challenging LLM benchmark that tests knowledge, coding, physical reasoning capabilities of LLMs.☆62Updated this week
- Course Project for COMP4471 on RWKV☆17Updated 2 years ago
- ☆113Updated 3 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆243Updated last year
- ☆29Updated last year
- Prepare for DeekSeek R1 inference: Benchmark CPU, DRAM, SSD, iGPU, GPU, ... with efficient code.☆74Updated last year
- 1.58-bit LLaMa model☆82Updated last year
- The DPAB-α Benchmark☆32Updated last year
- Welcome to the official repository of SINQ! A novel, fast and high-quality quantization method designed to make any Large Language Model …☆595Updated this week
- ☆135Updated 9 months ago
- Chat WebUI is an easy-to-use user interface for interacting with AI, and it comes with multiple useful built-in tools such as web search …☆49Updated 5 months ago
- Mini Model Daemon☆12Updated last year
- Lightweight toolkit package to train and fine-tune 1.58bit Language models☆112Updated 8 months ago
- Review/Check GGUF files and estimate the memory usage and maximum tokens per second.☆240Updated last month
- Create 3D files in the CLI with Small Language Model☆43Updated 3 months ago