SakanaAI / evo-memoryLinks
Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.
☆344Updated last year
Alternatives and similar repositories for evo-memory
Users that are interested in evo-memory are comparing it to the libraries listed below
Sorting:
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆362Updated last year
- PyTorch implementation of models from the Zamba2 series.☆186Updated 11 months ago
- GRadient-INformed MoE☆265Updated last year
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆174Updated 11 months ago
- This repo contains the source code for the paper "Evolution Strategies at Scale: LLM Fine-Tuning Beyond Reinforcement Learning"☆277Updated last month
- Train your own SOTA deductive reasoning model☆107Updated 9 months ago
- ☆131Updated last year
- Code for ExploreTom☆89Updated 6 months ago
- Super basic implementation (gist-like) of RLMs with REPL environments.☆290Updated 2 months ago
- smolLM with Entropix sampler on pytorch☆149Updated last year
- Simple & Scalable Pretraining for Neural Architecture Research☆305Updated 3 weeks ago
- Long context evaluation for large language models☆224Updated 9 months ago
- A compact LLM pretrained in 9 days by using high quality data☆337Updated 8 months ago
- Library for text-to-text regression, applicable to any input string representation and allows pretraining and fine-tuning over multiple r…☆301Updated last week
- DeMo: Decoupled Momentum Optimization☆198Updated last year
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆202Updated last year
- Build your own visual reasoning model☆415Updated last month
- ☆137Updated last year
- code for training & evaluating Contextual Document Embedding models☆201Updated 7 months ago
- Hypernetworks that adapt LLMs for specific benchmark tasks using only textual task description as the input☆933Updated 6 months ago
- Pretraining and inference code for a large-scale depth-recurrent language model☆856Updated 2 months ago
- EvaByte: Efficient Byte-level Language Models at Scale☆111Updated 8 months ago
- ☆205Updated last year
- Public repository for "The Surprising Effectiveness of Test-Time Training for Abstract Reasoning"☆341Updated last month
- Training teachers with reinforcement learning able to make LLMs learn how to reason for test time scaling.☆353Updated 6 months ago
- ☆185Updated last month
- ☆138Updated 4 months ago
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆249Updated 10 months ago
- Reverse Engineering Gemma 3n: Google's New Edge-Optimized Language Model☆255Updated 7 months ago
- ☆107Updated 5 months ago