nlpodyssey / rwkv
RWKV (Receptance Weighted Key Value) is a RNN with Transformer-level performance
☆39Updated 2 years ago
Alternatives and similar repositories for rwkv:
Users that are interested in rwkv are comparing it to the libraries listed below
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- Let us make Psychohistory (as in Asimov) a reality, and accessible to everyone. Useful for LLM grounding and games / fiction / business /…☆40Updated last year
- Here we collect trick questions and failed tasks for open source LLMs to improve them.☆32Updated last year
- Prepare for DeekSeek R1 inference: Benchmark CPU, DRAM, SSD, iGPU, GPU, ... with efficient code.☆70Updated last month
- A converter and basic tester for rwkv onnx☆42Updated last year
- SparseGPT + GPTQ Compression of LLMs like LLaMa, OPT, Pythia☆41Updated 2 years ago
- RWKV-7: Surpassing GPT☆82Updated 4 months ago
- RWKV-v2-RNN trained on the Pile. See https://github.com/BlinkDL/RWKV-LM for details.☆67Updated 2 years ago
- ☆26Updated 2 years ago
- RWKV, in easy to read code☆71Updated this week
- RWKV infctx trainer, for training arbitary context sizes, to 10k and beyond!☆148Updated 7 months ago
- RWKV in nanoGPT style☆187Updated 9 months ago
- GoldFinch and other hybrid transformer components☆45Updated 8 months ago
- tinygrad port of the RWKV large language model.☆44Updated 2 weeks ago
- Trying to deconstruct RWKV in understandable terms☆14Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated 2 years ago
- Official repository for the paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers"☆36Updated last year
- Evaluating LLMs with Dynamic Data☆78Updated last month
- RWKV model implementation☆37Updated last year
- https://x.com/BlinkDL_AI/status/1884768989743882276☆27Updated last month
- Run ONNX RWKV-v4 models with GPU acceleration using DirectML [Windows], or just on CPU [Windows AND Linux]; Limited to 430M model at this…☆20Updated 2 years ago
- ☆32Updated this week
- A byte-level decoder architecture that matches the performance of tokenized Transformers.☆63Updated 11 months ago
- RWKV centralised docs for the community☆21Updated last week
- Inference code for LLaMA 2 models☆30Updated 8 months ago
- Official implementation for 'Extending LLMs’ Context Window with 100 Samples'☆75Updated last year
- Framework agnostic python runtime for RWKV models☆145Updated last year
- 32 times longer context window than vanilla Transformers and up to 4 times longer than memory efficient Transformers.☆46Updated last year
- Griffin MQA + Hawk Linear RNN Hybrid☆85Updated 11 months ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆117Updated last year