rlite-project / RLiteLinks
A lightweight reinforcement learning framework that integrates seamlessly into your codebase, empowering developers to focus on algorithms with minimal intrusion.
☆29Updated 2 weeks ago
Alternatives and similar repositories for RLite
Users that are interested in RLite are comparing it to the libraries listed below
Sorting:
- Odysseus: Playground of LLM Sequence Parallelism☆69Updated 11 months ago
- Accelerate LLM preference tuning via prefix sharing with a single line of code☆41Updated last month
- LongSpec: Long-Context Speculative Decoding with Efficient Drafting and Verification☆53Updated 3 months ago
- ☆49Updated 3 weeks ago
- Async pipelined version of Verl☆91Updated last month
- Reproducing R1 for Code with Reliable Rewards☆201Updated last month
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆179Updated 2 months ago
- ☆144Updated this week
- ☆93Updated last week
- Official Implementation of SAM-Decoding: Speculative Decoding via Suffix Automaton☆28Updated 3 months ago
- ☆45Updated 3 months ago
- Official PyTorch implementation of the paper "dLLM-Cache: Accelerating Diffusion Large Language Models with Adaptive Caching" (dLLM-Cache…☆72Updated this week
- The code and data for the paper JiuZhang3.0☆45Updated last year
- Official Implementation of ARPO: End-to-End Policy Optimization for GUI Agents with Experience Replay☆71Updated last week
- This is the official repo of "QuickLLaMA: Query-aware Inference Acceleration for Large Language Models"☆51Updated 10 months ago
- The official implementation for Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free☆40Updated 3 weeks ago
- A scalable asynchronous reinforcement learning implementation with in-flight weight updates.☆119Updated this week
- qwen-nsa☆66Updated last month
- Code accompanying the paper "Noise Contrastive Alignment of Language Models with Explicit Rewards" (NeurIPS 2024)☆53Updated 6 months ago
- Code for Paper: Learning Adaptive Parallel Reasoning with Language Models☆94Updated last month
- [NeurIPS-2024] 📈 Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623☆84Updated 8 months ago
- Homepage for ProLong (Princeton long-context language models) and paper "How to Train Long-Context Language Models (Effectively)"☆186Updated 2 months ago
- Revisiting Mid-training in the Era of RL Scaling☆48Updated last month
- Official repository for "RLVR-World: Training World Models with Reinforcement Learning", https://arxiv.org/abs/2505.13934☆36Updated this week
- ☆93Updated 2 weeks ago
- Nano repo for RL training of LLMs☆60Updated this week
- ICML2025: Forest-of-Thought: Scaling Test-Time Compute for Enhancing LLM Reasoning☆39Updated last month
- ☆93Updated 8 months ago
- "what, how, where, and how well? a survey on test-time scaling in large language models" repository☆41Updated last week
- ☆33Updated 8 months ago