JL-er / WorldRWKVLinks
The WorldRWKV project aims to implement training and inference across various modalities using the RWKV7 architecture. By leveraging different encoders, the project allows for flexible modality switching and aspires to achieve end-to-end cross-modal inference.
☆54Updated 3 weeks ago
Alternatives and similar repositories for WorldRWKV
Users that are interested in WorldRWKV are comparing it to the libraries listed below
Sorting:
- Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton☆44Updated last week
- RWKV-LM-V7(https://github.com/BlinkDL/RWKV-LM) Under Lightning Framework☆45Updated last month
- ☆140Updated last week
- VisualRWKV is the visual-enhanced version of the RWKV language model, enabling RWKV to handle various visual tasks.☆233Updated 3 months ago
- ☆38Updated 4 months ago
- RWKV-X is a Linear Complexity Hybrid Language Model based on the RWKV architecture, integrating Sparse Attention to improve the model's l…☆46Updated last month
- ☆17Updated 7 months ago
- State tuning tunes the state☆35Updated 6 months ago
- [EMNLP 2024] RWKV-CLIP: A Robust Vision-Language Representation Learner☆141Updated 3 months ago
- Reinforcement Learning Toolkit for RWKV.(v6,v7,ARWKV) Distillation,SFT,RLHF(DPO,ORPO), infinite context training, Aligning. Exploring the…☆51Updated last week
- This project is to extend RWKV LM's capabilities including sequence classification/embedding/peft/cross encoder/bi encoder/multi modaliti…☆10Updated last year
- imagetokenizer is a python package, helps you encoder visuals and generate visuals token ids from codebook, supports both image and video…☆35Updated last year
- A large-scale RWKV v7(World, PRWKV, Hybrid-RWKV) inference. Capable of inference by combining multiple states(Pseudo MoE). Easy to deploy…☆42Updated this week
- This is an inference framework for the RWKV large language model implemented purely in native PyTorch. The official native implementation…☆133Updated last year
- [ICML'24 Oral] The official code of "DiJiang: Efficient Large Language Models through Compact Kernelization", a novel DCT-based linear at…☆102Updated last year
- ☆34Updated last year
- [ICML 2025] Fourier Position Embedding: Enhancing Attention’s Periodic Extension for Length Generalization☆90Updated 2 months ago
- https://x.com/BlinkDL_AI/status/1884768989743882276☆28Updated 3 months ago
- RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best…☆51Updated 5 months ago
- RWKV infctx trainer, for training arbitary context sizes, to 10k and beyond!☆147Updated last year
- A specialized RWKV-7 model for Othello(a.k.a. Reversi) that predicts legal moves, evaluates positions, and performs in-context search. It…☆41Updated 7 months ago
- ☆22Updated 8 months ago
- Pytorch implementation of https://arxiv.org/html/2404.07143v1☆21Updated last year
- rwkv finetuning☆37Updated last year
- RWKV in nanoGPT style☆192Updated last year
- A collection of tricks and tools to speed up transformer models☆170Updated 2 months ago
- 😊 TPTT: Transforming Pretrained Transformers into Titans☆25Updated this week
- RWKV-7: Surpassing GPT☆94Updated 9 months ago
- Official repository for ICML 2024 paper "MoRe Fine-Tuning with 10x Fewer Parameters"☆20Updated 3 months ago
- Lightning Attention-2: A Free Lunch for Handling Unlimited Sequence Lengths in Large Language Models☆325Updated 6 months ago