cryscan / web-rwkv
Implementation of the RWKV language model in pure WebGPU/Rust.
☆297Updated this week
Alternatives and similar repositories for web-rwkv:
Users that are interested in web-rwkv are comparing it to the libraries listed below
- The all-in-one RWKV runtime box with embed, RAG, AI agents, and more.☆548Updated 3 weeks ago
- A highly customizable, full scale web backend for web-rwkv, built on axum with websocket protocol.☆26Updated 11 months ago
- A torchless, c++ rwkv implementation using 8bit quantization, written in cuda/hip/vulkan for maximum compatibility and minimum dependenci…☆310Updated last year
- ☆116Updated 3 weeks ago
- RWKV infctx trainer, for training arbitary context sizes, to 10k and beyond!☆148Updated 7 months ago
- RWKV in nanoGPT style☆189Updated 9 months ago
- RWKV models and examples powered by candle.☆18Updated last month
- This project is established for real-time training of the RWKV model.☆49Updated 10 months ago
- A relatively basic implementation of RWKV in Rust written by someone with very little math and ML knowledge. Supports 32, 8 and 4 bit eva…☆93Updated last year
- High-level, optionally asynchronous Rust bindings to llama.cpp☆215Updated 9 months ago
- ☆82Updated 10 months ago
- RWKV is a RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best …☆414Updated last year
- Inference of Mamba models in pure C☆187Updated last year
- RWKV centralised docs for the community☆22Updated last week
- RAG SYSTEM FOR RWKV☆45Updated 4 months ago
- ☆13Updated 3 months ago
- Fine-tuning RWKV-World model☆25Updated last year
- A converter and basic tester for rwkv onnx☆42Updated last year
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆340Updated 2 weeks ago
- ☆32Updated 2 years ago
- This is an inference framework for the RWKV large language model implemented purely in native PyTorch. The official native implementation…☆127Updated 8 months ago
- A Fish Speech implementation in Rust, with Candle.rs☆75Updated last month
- llama.cpp fork with additional SOTA quants and improved performance☆231Updated this week
- Rust+OpenCL+AVX2 implementation of LLaMA inference code☆545Updated last year
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model☆1,492Updated last week
- ☆132Updated 4 months ago
- ☆10Updated last year
- WebGPU LLM inference tuned by hand☆149Updated last year
- ChatGPT-like Web UI for RWKVstic☆100Updated last year