SakanaAI / evo-memory
Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.
☆302Updated 5 months ago
Alternatives and similar repositories for evo-memory:
Users that are interested in evo-memory are comparing it to the libraries listed below
- PyTorch implementation of models from the Zamba2 series.☆179Updated 2 months ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆314Updated 4 months ago
- Alice in Wonderland code base for experiments and raw experiments data☆129Updated 2 months ago
- A Self-adaptation Framework🐙 that adapts LLMs for unseen tasks in real-time!☆1,037Updated 2 months ago
- ☆107Updated 3 months ago
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆170Updated 2 months ago
- Pretraining code for a large-scale depth-recurrent language model☆734Updated this week
- smolLM with Entropix sampler on pytorch☆151Updated 5 months ago
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024☆284Updated last month
- [ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling☆860Updated last month
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"