test-time-training / e2eLinks
Official JAX implementation of End-to-End Test-Time Training for Long Context
☆511Updated last week
Alternatives and similar repositories for e2e
Users that are interested in e2e are comparing it to the libraries listed below
Sorting:
- ☆394Updated last week
- The official github repo for "Diffusion Language Models are Super Data Learners".☆220Updated 3 months ago
- ☆388Updated 3 months ago
- GPU-optimized framework for training diffusion language models at any scale. The backend of Quokka, Super Data Learners, and OpenMoE 2 tr…☆321Updated 2 months ago
- Training teachers with reinforcement learning able to make LLMs learn how to reason for test time scaling.☆358Updated 7 months ago
- [ICLR 2026] Official PyTorch Implementation of RLP: Reinforcement as a Pretraining Objective☆231Updated last week
- QeRL enables RL for 32B LLMs on a single H100 GPU.☆481Updated 2 months ago
- Chain of Experts (CoE) enables communication between experts within Mixture-of-Experts (MoE) models☆228Updated 3 months ago
- Official implementation of paper: SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training☆315Updated 9 months ago
- OpenTinker is an RL-as-a-Service infrastructure for foundation models☆625Updated last week
- PaCoRe: Learning to Scale Test-Time Compute with Parallel Coordinated Reasoning☆296Updated 3 weeks ago
- ☆236Updated this week
- Open-source release accompanying Gao et al. 2025☆501Updated last month
- Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models. TMLR 2025.☆148Updated 4 months ago
- LLaDA2.0 is the diffusion language model series developed by InclusionAI team, Ant Group.☆240Updated last month
- This repo contains the source code for the paper "Evolution Strategies at Scale: LLM Fine-Tuning Beyond Reinforcement Learning"☆288Updated 2 months ago
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).☆344Updated last month
- Training API and CLI☆325Updated last week
- Esoteric Language Models☆111Updated this week
- Physics of Language Models: Part 4.2, Canon Layers at Scale where Synthetic Pretraining Resonates in Reality☆317Updated last month
- Official PyTorch implementation for Hogwild! Inference: Parallel LLM Generation with a Concurrent Attention Cache☆140Updated 5 months ago
- [NeurIPS 2025] Reinforcement Learning for Reasoning in Large Language Models with One Training Example☆405Updated 2 months ago
- [ICLR2025] DiffuGPT and DiffuLLaMA: Scaling Diffusion Language Models via Adaptation from Autoregressive Models☆362Updated 8 months ago
- An AI benchmark for creative, human-like problem solving using Sudoku variants☆158Updated last month
- Extending the Context of Pretrained LLMs by Dropping Their Positional Embedding☆200Updated 3 weeks ago
- ☆208Updated last year
- [ICLR 2026] Tina: Tiny Reasoning Models via LoRA☆319Updated 4 months ago
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆252Updated last year
- [ICLR 2026] TraceRL & TraDo-8B: Revolutionizing Reinforcement Learning Framework for Diffusion Large Language Models☆423Updated last week
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆371Updated last year