SakanaAI / continuous-thought-machinesLinks
Continuous Thought Machines, because thought takes time and reasoning is a process.
☆1,687Updated last week
Alternatives and similar repositories for continuous-thought-machines
Users that are interested in continuous-thought-machines are comparing it to the libraries listed below
Sorting:
- A Self-adaptation Framework🐙 that adapts LLMs for unseen tasks in real-time!☆1,180Updated 11 months ago
- Unofficial implementation of Titans, SOTA memory for transformers, in Pytorch☆1,829Updated 2 weeks ago
- Pretraining and inference code for a large-scale depth-recurrent language model☆857Updated this week
- Code for BLT research paper☆2,018Updated 2 months ago
- AlphaGo Moment for Model Architecture Discovery.☆1,127Updated last month
- Darwin Gödel Machine: Open-Ended Evolution of Self-Improving Agents☆1,774Updated 4 months ago
- Self-Adapting Language Models☆1,629Updated 5 months ago
- Official Repository of Absolute Zero Reasoner☆1,782Updated 4 months ago
- Dream 7B, a large diffusion language model☆1,126Updated last month
- Muon is an optimizer for hidden layers in neural networks☆2,154Updated last month
- H-Net: Hierarchical Network with Dynamic Chunking☆798Updated last month
- dLLM: Simple Diffusion Language Modeling☆1,526Updated last week
- PyTorch Code for Energy-Based Transformers paper -- generalizable reasoning and scalable learning☆569Updated last month
- Training Large Language Model to Reason in a Continuous Latent Space☆1,419Updated 4 months ago
- A minimal implementation of DeepMind's Genie world model☆1,081Updated last month
- ☆596Updated 7 months ago
- Implementing DeepSeek R1's GRPO algorithm from scratch☆1,724Updated 8 months ago
- A Reproduction of GDM's Nested Learning Paper☆524Updated last month
- PyTorch code and models for VJEPA2 self-supervised learning from video.☆2,618Updated 4 months ago
- ShinkaEvolve: Towards Open-Ended and Sample-Efficient Program Evolution☆761Updated this week
- Automating the Search for Artificial Life with Foundation Models!☆448Updated 2 months ago
- [ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling☆936Updated last month
- Official PyTorch implementation for "Large Language Diffusion Models"☆3,440Updated last month
- Async RL Training at Scale☆976Updated this week
- Hypernetworks that adapt LLMs for specific benchmark tasks using only textual task description as the input☆933Updated 6 months ago
- Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.☆345Updated last year
- Frontier Models playing the board game Diplomacy.☆611Updated this week
- Official repository for the paper "Grokfast: Accelerated Grokking by Amplifying Slow Gradients"☆569Updated last year
- [ICLR2025 Spotlight🔥] Official Implementation of TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters☆580Updated 10 months ago
- Post-training with Tinker☆2,640Updated this week