BlinkDL / RWKV-v2-RNN-Pile
RWKV-v2-RNN trained on the Pile. See https://github.com/BlinkDL/RWKV-LM for details.
☆66Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for RWKV-v2-RNN-Pile
- This project aims to make RWKV Accessible to everyone using a Hugging Face like interface, while keeping it close to the R and D RWKV bra…☆64Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated last year
- ☆42Updated last year
- Let us make Psychohistory (as in Asimov) a reality, and accessible to everyone. Useful for LLM grounding and games / fiction / business /…☆40Updated last year
- Framework agnostic python runtime for RWKV models☆145Updated last year
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- ☆128Updated 2 years ago
- Experiments with generating opensource language model assistants☆97Updated last year
- Hidden Engrams: Long Term Memory for Transformer Model Inference☆34Updated 3 years ago
- Code repository for the c-BTM paper☆105Updated last year
- Code for the paper "The Impact of Positional Encoding on Length Generalization in Transformers", NeurIPS 2023☆127Updated 6 months ago
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆35Updated 11 months ago
- Implementation of Token Shift GPT - An autoregressive model that solely relies on shifting the sequence space for mixing☆47Updated 2 years ago
- SparseGPT + GPTQ Compression of LLMs like LLaMa, OPT, Pythia☆41Updated last year
- Latent Diffusion Language Models☆67Updated last year
- An experimental implementation of the retrieval-enhanced language model☆75Updated last year
- Inference script for Meta's LLaMA models using Hugging Face wrapper☆111Updated last year
- Text-writing denoising diffusion (and much more)☆30Updated last year
- LayerNorm(SmallInit(Embedding)) in a Transformer to improve convergence☆45Updated 2 years ago
- Transformers at any scale☆41Updated 10 months ago
- Multipack distributed sampler for fast padding-free training of LLMs☆178Updated 3 months ago
- Implementation of the conditionally routed attention in the CoLT5 architecture, in Pytorch☆225Updated 2 months ago
- ☆64Updated 2 years ago
- One stop shop for all things carp☆59Updated 2 years ago
- RWKV-7: Surpassing GPT☆44Updated this week
- GoldFinch and other hybrid transformer components☆39Updated 4 months ago
- Experiments around a simple idea for inducing multiple hierarchical predictive model within a GPT☆205Updated 3 months ago
- Techniques used to run BLOOM at inference in parallel☆37Updated 2 years ago