test-time-training / e2eLinks
Official JAX implementation of End-to-End Test-Time Training for Long Context
☆445Updated this week
Alternatives and similar repositories for e2e
Users that are interested in e2e are comparing it to the libraries listed below
Sorting:
- The official github repo for "Diffusion Language Models are Super Data Learners".☆218Updated 2 months ago
- QeRL enables RL for 32B LLMs on a single H100 GPU.☆473Updated last month
- RLP: Reinforcement as a Pretraining Objective☆223Updated 3 months ago
- ☆373Updated 2 months ago
- Training teachers with reinforcement learning able to make LLMs learn how to reason for test time scaling.☆358Updated 6 months ago
- GPU-optimized framework for training diffusion language models at any scale. The backend of Quokka, Super Data Learners, and OpenMoE 2 tr…☆312Updated 2 months ago
- OpenTinker is an RL-as-a-Service infrastructure for foundation models☆577Updated this week
- Esoteric Language Models☆108Updated last month
- Simple & Scalable Pretraining for Neural Architecture Research☆306Updated last month
- Official implementation of paper: SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training☆315Updated 8 months ago
- This repo contains the source code for the paper "Evolution Strategies at Scale: LLM Fine-Tuning Beyond Reinforcement Learning"☆285Updated last month
- Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models. TMLR 2025.☆138Updated 4 months ago
- Chain of Experts (CoE) enables communication between experts within Mixture-of-Experts (MoE) models