McGill-NLP / agent-reward-benchLinks
AgentRewardBench: Evaluating Automatic Evaluations of Web Agent Trajectories
☆37Updated last month
Alternatives and similar repositories for agent-reward-bench
Users that are interested in agent-reward-bench are comparing it to the libraries listed below
Sorting:
- WideSearch: Benchmarking Agentic Broad Info-Seeking☆94Updated last month
- [NeurIPS 2025 Spotlight] ReasonFlux-Coder: Open-Source LLM Coders with Co-Evolving Reinforcement Learning☆122Updated 2 weeks ago
- ☆96Updated last month
- ☆103Updated 9 months ago
- Efficient Agent Training for Computer Use☆132Updated 3 weeks ago
- ☆46Updated 3 months ago
- ☆62Updated 3 months ago
- PreAct: Prediction Enhances Agent's Planning Ability (Coling2025)☆29Updated 9 months ago
- DeepDive: Advancing Deep Search Agents with Knowledge Graphs and Multi-Turn RL☆159Updated 2 weeks ago
- SSRL: Self-Search Reinforcement Learning☆145Updated last month
- [EMNLP 2025 Industry] Repo for "Z1: Efficient Test-time Scaling with Code"☆64Updated 5 months ago
- ☆37Updated last month
- ☆48Updated 7 months ago
- ☆77Updated last month
- ☆53Updated 7 months ago
- Revisiting Mid-training in the Era of Reinforcement Learning Scaling☆176Updated 2 months ago
- ☆19Updated 6 months ago
- [ICLR 2025] LongPO: Long Context Self-Evolution of Large Language Models through Short-to-Long Preference Optimization☆40Updated 7 months ago
- Klear-Reasoner: Advancing Reasoning Capability via Gradient-Preserving Clipping Policy Optimization☆71Updated last week
- [ACL'25] We propose a novel fine-tuning method, Separate Memory and Reasoning, which combines prompt tuning with LoRA.☆76Updated 2 weeks ago
- RL Scaling and Test-Time Scaling (ICML'25)☆111Updated 8 months ago
- SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning. COLM 2024 Accepted Paper☆33Updated last year
- The official repo for "AceCoder: Acing Coder RL via Automated Test-Case Synthesis" [ACL25]☆88Updated 5 months ago
- [ACL 2025] Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆106Updated 3 months ago
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆82Updated 6 months ago
- o1 Chain of Thought Examples☆33Updated last year
- RM-R1: Unleashing the Reasoning Potential of Reward Models☆137Updated 3 months ago
- ☆74Updated last month
- Process Reward Models That Think☆53Updated 3 months ago
- MPO: Boosting LLM Agents with Meta Plan Optimization (EMNLP 2025 Findings)☆71Updated last month