lmgame-org / GRLLinks
Multi-Turn RL Training System with AgentTrainer for Language Model Game Reinforcement Learning
☆56Updated 3 weeks ago
Alternatives and similar repositories for GRL
Users that are interested in GRL are comparing it to the libraries listed below
Sorting:
- Defeating the Training-Inference Mismatch via FP16☆170Updated last month
- The evaluation framework for training-free sparse attention in LLMs☆108Updated 2 months ago
- ☆102Updated 10 months ago
- Memory optimized Mixture of Experts☆72Updated 5 months ago
- ☆109Updated 3 months ago
- ☆61Updated 6 months ago
- The official repository for SkyLadder: Better and Faster Pretraining via Context Window Scheduling☆41Updated last week
- DPO, but faster 🚀☆46Updated last year
- Esoteric Language Models☆108Updated last month
- Official JAX implementation of End-to-End Test-Time Training for Long Context☆214Updated last week
- Kinetics: Rethinking Test-Time Scaling Laws☆85Updated 5 months ago
- ☆85Updated last month
- An efficient implementation of the NSA (Native Sparse Attention) kernel☆128Updated 6 months ago
- Official implementation of Regularized Policy Gradient (RPG) (https://arxiv.org/abs/2505.17508)☆63Updated this week
- 🔥 LLM-powered GPU kernel synthesis: Train models to convert PyTorch ops into optimized Triton kernels via SFT+RL. Multi-turn compilation…☆110Updated last month
- [Preprint] RLVE: Scaling Up Reinforcement Learning for Language Models with Adaptive Verifiable Environments☆165Updated last month
- Flash-Muon: An Efficient Implementation of Muon Optimizer☆225Updated 6 months ago
- QeRL enables RL for 32B LLMs on a single H100 GPU.☆469Updated last month
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆131Updated last year
- ☆54Updated last year
- ☆112Updated last year
- ☆91Updated last year
- [ICML 2025] From Low Rank Gradient Subspace Stabilization to Low-Rank Weights: Observations, Theories and Applications☆52Updated 2 months ago
- Official implementation for Training LLMs with MXFP4☆116Updated 8 months ago
- Fast and memory-efficient exact attention☆75Updated 10 months ago
- ☆133Updated 7 months ago
- Using FlexAttention to compute attention with different masking patterns☆47Updated last year
- ☆213Updated last month
- Linear Attention Sequence Parallelism (LASP)☆88Updated last year
- [ICLR2025] Codebase for "ReMoE: Fully Differentiable Mixture-of-Experts with ReLU Routing", built on Megatron-LM.☆104Updated last year