Joluck / mod-rwkvLinks
The WorldRWKV project aims to implement training and inference across various modalities using the RWKV7 architecture. By leveraging different encoders, the project allows for flexible modality switching and aspires to achieve end-to-end cross-modal inference.
☆62Updated 3 weeks ago
Alternatives and similar repositories for mod-rwkv
Users that are interested in mod-rwkv are comparing it to the libraries listed below
Sorting:
- ☆164Updated last week
- RWKV-LM-V7(https://github.com/BlinkDL/RWKV-LM) Under Lightning Framework☆54Updated 2 weeks ago
- Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton☆46Updated 4 months ago
- ☆17Updated last year
- ☆39Updated 8 months ago
- VisualRWKV is the visual-enhanced version of the RWKV language model, enabling RWKV to handle various visual tasks.☆238Updated 7 months ago
- Reinforcement Learning Toolkit for RWKV.(v6,v7,ARWKV) Distillation,SFT,RLHF(DPO,ORPO), infinite context training, Aligning. Exploring the…☆59Updated 3 months ago
- [EMNLP 2024] RWKV-CLIP: A Robust Vision-Language Representation Learner☆148Updated 3 weeks ago
- State tuning tunes the state☆35Updated 10 months ago
- RWKV-X is a Linear Complexity Hybrid Language Model based on the RWKV architecture, integrating Sparse Attention to improve the model's l…☆53Updated 5 months ago
- A large-scale RWKV v7(World, PRWKV, Hybrid-RWKV) inference. Capable of inference by combining multiple states(Pseudo MoE). Easy to deploy…☆46Updated 2 months ago
- A specialized RWKV-7 model for Othello(a.k.a. Reversi) that predicts legal moves, evaluates positions, and performs in-context search. It…☆43Updated 11 months ago
- Efficient RWKV inference engine. RWKV7 7.2B fp16 decoding 10250 tps @ single 5090.☆74Updated 3 weeks ago
- ☆34Updated last year
- https://x.com/BlinkDL_AI/status/1884768989743882276☆28Updated 8 months ago
- This is an inference framework for the RWKV large language model implemented purely in native PyTorch. The official native implementation…☆132Updated last year
- RWKV infctx trainer, for training arbitary context sizes, to 10k and beyond!☆148Updated last year
- MiSS is a novel PEFT method that features a low-rank structure but introduces a new update mechanism distinct from LoRA, achieving an exc…☆25Updated 2 months ago
- ☆23Updated last year
- RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best…☆59Updated 9 months ago
- RWKV-7: Surpassing GPT☆103Updated last year
- 😊 TPTT: Transforming Pretrained Transformers into Titans☆49Updated last month
- Evaluating LLMs with Dynamic Data☆105Updated this week
- RAG SYSTEM FOR RWKV☆52Updated last year
- Pytorch implementation of https://arxiv.org/html/2404.07143v1☆21Updated last year
- RWKV in nanoGPT style☆197Updated last year
- [ICML 2025] Fourier Position Embedding: Enhancing Attention’s Periodic Extension for Length Generalization☆105Updated 7 months ago
- rwkv finetuning☆37Updated last year
- [ICML'24 Oral] The official code of "DiJiang: Efficient Large Language Models through Compact Kernelization", a novel DCT-based linear at…☆104Updated last year
- [ICML 2025] |TokenSwift: Lossless Acceleration of Ultra Long Sequence Generation☆118Updated 7 months ago