DistRL-lab / distrl-openLinks
DistRL: An Asynchronous Distributed Reinforcement Learning Framework for On-Device Control Agents
☆25Updated last month
Alternatives and similar repositories for distrl-open
Users that are interested in distrl-open are comparing it to the libraries listed below
Sorting:
- ☆18Updated 3 months ago
- SPA-Bench: A Comprehensive Benchmark for SmartPhone Agent Evaluation☆43Updated last month
- Improving Math reasoning through Direct Preference Optimization with Verifiable Pairs☆15Updated 5 months ago
- Official implementation of the NeurIPS 2024 paper CORY☆20Updated 5 months ago
- Code release for "Generating Code World Models with Large Language Models Guided by Monte Carlo Tree Search" published at NeurIPS '24.☆11Updated 6 months ago
- [ICML 2025] "From Debate to Equilibrium: Belief-Driven Multi-Agent LLM Reasoning via Bayesian Nash Equilibrium"☆22Updated last month
- Benchmarking LLMs' Gaming Ability in Multi-Agent Environments☆87Updated 4 months ago
- [NeurIPS 2023] Large Language Models Are Semi-Parametric Reinforcement Learning Agents☆34Updated last year
- Code for NeurIPS 2024 paper "Regularizing Hidden States Enables Learning Generalizable Reward Model for LLMs"☆38Updated 6 months ago
- [ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization☆86Updated last year
- ☆25Updated 2 months ago
- Research Code for "ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL"☆189Updated 4 months ago
- A Framework for LLM-based Multi-Agent Reinforced Training and Inference☆224Updated 2 weeks ago
- ☆21Updated last month
- ☆280Updated 3 months ago
- The Entropy Mechanism of Reinforcement Learning for Large Language Model Reasoning.☆315Updated last month
- ☆209Updated 2 weeks ago
- [ACL'24, Outstanding Paper] Emulated Disalignment: Safety Alignment for Large Language Models May Backfire!☆38Updated last year
- Rewarded soups official implementation☆60Updated last year
- An index of algorithms for reinforcement learning from human feedback (rlhf))☆93Updated last year
- Official Repository of "Learning to Reason under Off-Policy Guidance"☆288Updated last month
- ☆21Updated last month
- Reinforced Multi-LLM Agents training☆40Updated 2 months ago
- ☆361Updated 3 weeks ago
- ☆116Updated 7 months ago
- This is the official implementation of paper "Leveraging Dual Process Theory in Language Agent Framework for Simultaneous Human-AI Collab…☆40Updated 3 months ago
- Official repo for paper DigiRL: Training In-The-Wild Device-Control Agents with Autonomous Reinforcement Learning.☆373Updated 6 months ago
- This my attempt to create Self-Correcting-LLM based on the paper Training Language Models to Self-Correct via Reinforcement Learning by g…☆35Updated last month
- AdaRFT: Efficient Reinforcement Finetuning via Adaptive Curriculum Learning☆41Updated 2 months ago
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆101Updated last month