rhyang2021 / ARIALinks
Source code for our paper: "ARIA: Training Language Agents with Intention-Driven Reward Aggregation".
☆25Updated 5 months ago
Alternatives and similar repositories for ARIA
Users that are interested in ARIA are comparing it to the libraries listed below
Sorting:
- R1-Searcher++: Incentivizing the Dynamic Knowledge Acquisition of LLMs via Reinforcement Learning☆69Updated 7 months ago
- RAG-RewardBench: Benchmarking Reward Models in Retrieval Augmented Generation for Preference Alignment☆16Updated last year
- ☆70Updated 7 months ago
- ☆55Updated 3 months ago
- ☆21Updated 8 months ago
- Official code for paper "SPA-RL: Reinforcing LLM Agent via Stepwise Progress Attribution"☆62Updated 4 months ago
- ☆51Updated last year
- [ICML 2025] M-STAR (Multimodal Self-Evolving TrAining for Reasoning) Project. Diving into Self-Evolving Training for Multimodal Reasoning☆69Updated 6 months ago
- This is the official implementation of the paper "S²R: Teaching LLMs to Self-verify and Self-correct via Reinforcement Learning"☆72Updated 8 months ago
- [ICML 2025] Teaching Language Models to Critique via Reinforcement Learning☆119Updated 8 months ago
- From Accuracy to Robustness: A Study of Rule- and Model-based Verifiers in Mathematical Reasoning.☆24Updated 3 months ago
- ☆70Updated 6 months ago
- Resources and paper list for 'Scaling Environments for Agents'. This repository accompanies our survey on how environments contribute to …☆53Updated 3 weeks ago
- [arxiv: 2505.02156] Adaptive Thinking via Mode Policy Optimization for Social Language Agents☆46Updated 6 months ago
- [AAAI 2026] Official codebase for "GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning".☆94Updated 2 months ago
- The rule-based evaluation subset and code implementation of Omni-MATH☆26Updated last year
- ☆47Updated 3 months ago
- The official repository of the Omni-MATH benchmark.☆91Updated last year
- The official code repository for the paper "Mirage or Method? How Model–Task Alignment Induces Divergent RL Conclusions".☆15Updated 4 months ago
- ☆50Updated 11 months ago
- Code for "Language Models Can Learn from Verbal Feedback Without Scalar Rewards"☆55Updated last week
- ☆58Updated last year
- [ICLR 25 Oral] RM-Bench: Benchmarking Reward Models of Language Models with Subtlety and Style☆73Updated 5 months ago
- This the implementation of LeCo☆31Updated 11 months ago
- [NeurIPS'25 Spotlight] ARM: Adaptive Reasoning Model☆63Updated 2 months ago
- ☆41Updated 4 months ago
- Code for Evolving Language Models without Labels: Majority Drives Selection, Novelty Promotes Variation (EVOL-RL).☆41Updated 2 months ago
- Revisiting Mid-training in the Era of Reinforcement Learning Scaling☆182Updated 5 months ago
- ☆13Updated last year
- ☆35Updated 3 weeks ago