BlinkDL / WorldModel
Let us make Psychohistory (as in Asimov) a reality, and accessible to everyone. Useful for LLM grounding and games / fiction / business / finance / governance, and can align agents with human too.
☆40Updated last year
Related projects ⓘ
Alternatives and complementary repositories for WorldModel
- ☆42Updated last year
- This project aims to make RWKV Accessible to everyone using a Hugging Face like interface, while keeping it close to the R and D RWKV bra…☆64Updated last year
- RWKV-7: Surpassing GPT☆45Updated this week
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- Here we collect trick questions and failed tasks for open source LLMs to improve them.☆32Updated last year
- RWKV, in easy to read code☆55Updated this week
- ☆15Updated 8 months ago
- RWKV-v2-RNN trained on the Pile. See https://github.com/BlinkDL/RWKV-LM for details.☆66Updated 2 years ago
- GoldFinch and other hybrid transformer components☆39Updated 4 months ago
- SparseGPT + GPTQ Compression of LLMs like LLaMa, OPT, Pythia☆41Updated last year
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆35Updated 11 months ago
- Evaluating LLMs with Dynamic Data☆72Updated last week
- ☆33Updated 4 months ago
- Zeta implementation of a reusable and plug in and play feedforward from the paper "Exponentially Faster Language Modeling"☆15Updated last week
- Fast modular code to create and train cutting edge LLMs☆65Updated 6 months ago
- tinygrad port of the RWKV large language model.☆43Updated 5 months ago
- Enhancing LangChain prompts to work better with RWKV models☆34Updated last year
- Run ONNX RWKV-v4 models with GPU acceleration using DirectML [Windows], or just on CPU [Windows AND Linux]; Limited to 430M model at this…☆20Updated last year
- ☆49Updated 8 months ago
- Script and instruction how to fine-tune large RWKV model on your data for Alpaca dataset☆31Updated last year
- BigKnow2022: Bringing Language Models Up to Speed☆14Updated last year
- Here we will test various linear attention designs.☆56Updated 6 months ago
- QuIP quantization☆46Updated 8 months ago
- RWKV model implementation☆38Updated last year
- My Implementation of Q-Sparse: All Large Language Models can be Fully Sparsely-Activated☆30Updated 3 months ago
- 📖 — Notebooks related to RWKV☆59Updated last year
- Trying to deconstruct RWKV in understandable terms☆14Updated last year
- Training a reward model for RLHF using RWKV.☆14Updated last year
- RWKV in nanoGPT style☆177Updated 5 months ago