agentica-project / rllmLinks
☆16Updated this week
Alternatives and similar repositories for rllm
Users that are interested in rllm are comparing it to the libraries listed below
Sorting:
- Code for the paper "VinePPO: Unlocking RL Potential For LLM Reasoning Through Refined Credit Assignment"☆167Updated 2 months ago
- Async pipelined version of Verl☆110Updated 3 months ago
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆123Updated 10 months ago
- Homepage for ProLong (Princeton long-context language models) and paper "How to Train Long-Context Language Models (Effectively)"☆218Updated 4 months ago
- Code and data used in the paper: "Training on Incorrect Synthetic Data via RL Scales LLM Math Reasoning Eight-Fold"☆30Updated last year
- ☆68Updated last year
- Self-Alignment with Principle-Following Reward Models☆162Updated 2 months ago
- Official github repo for the paper "Compression Represents Intelligence Linearly" [COLM 2024]☆139Updated 10 months ago
- GenRM-CoT: Data release for verification rationales☆63Updated 9 months ago
- Official repository for R2E-Gym: Procedural Environment Generation and Hybrid Verifiers for Scaling Open-Weights SWE Agents☆136Updated 3 weeks ago
- Code for ACL2024 paper - Adversarial Preference Optimization (APO).☆56Updated last year
- ☆147Updated 8 months ago
- Repo of paper "Free Process Rewards without Process Labels"☆160Updated 4 months ago
- Official Repo for ICLR 2024 paper MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback by Xingyao Wang*, Ziha…☆128Updated last year
- ☆71Updated 4 months ago
- The code for creating the iGSM datasets in papers "Physics of Language Models Part 2.1, Grade-School Math and the Hidden Reasoning Proces…☆60Updated 6 months ago
- Official repository for ACL 2025 paper "ProcessBench: Identifying Process Errors in Mathematical Reasoning"☆166Updated 2 months ago
- [NeurIPS'24 Spotlight] Observational Scaling Laws☆56Updated 10 months ago
- (ICML 2024) Alphazero-like Tree-Search can guide large language model decoding and training☆278Updated last year
- Code and models for EMNLP 2024 paper "WPO: Enhancing RLHF with Weighted Preference Optimization"☆41Updated 10 months ago
- Code and example data for the paper: Rule Based Rewards for Language Model Safety☆190Updated last year
- ☆99Updated last year
- Research Code for "ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL"☆184Updated 3 months ago
- Domain-specific preference (DSP) data and customized RM fine-tuning.☆25Updated last year
- This is the repository that contains the source code for the Self-Evaluation Guided MCTS for online DPO.☆319Updated 11 months ago
- ☆49Updated 2 months ago
- ☆33Updated 10 months ago
- Code for the preprint "Cache Me If You Can: How Many KVs Do You Need for Effective Long-Context LMs?"☆41Updated 3 weeks ago
- Implementations of online merging optimizers proposed by Online Merging Optimizers for Boosting Rewards and Mitigating Tax in Alignment☆75Updated last year
- RLHF implementation details of OAI's 2019 codebase☆187Updated last year