BBuf / RWKV-World-HF-Tokenizer
☆33Updated 4 months ago
Related projects ⓘ
Alternatives and complementary repositories for RWKV-World-HF-Tokenizer
- A repository for research on medium sized language models.☆74Updated 5 months ago
- A fast RWKV Tokenizer written in Rust☆36Updated 2 months ago
- QuIP quantization☆46Updated 8 months ago
- Official repository for the paper "SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention"☆92Updated last month
- ☆40Updated this week
- Official implementation for 'Extending LLMs’ Context Window with 100 Samples'☆74Updated 10 months ago
- ☆27Updated 5 months ago
- Implementation of the paper: "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" from Google in pyTO…☆52Updated last week
- DPO, but faster 🚀☆23Updated 3 weeks ago
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- RWKV infctx trainer, for training arbitary context sizes, to 10k and beyond!☆133Updated 3 months ago
- Evaluating LLMs with Dynamic Data☆72Updated last week
- My Implementation of Q-Sparse: All Large Language Models can be Fully Sparsely-Activated