Yu-Fangxu / FoRLinks
[ICML 2025] Flow of Reasoning: Training LLMs for Divergent Reasoning with Minimal Examples
☆100Updated last month
Alternatives and similar repositories for FoR
Users that are interested in FoR are comparing it to the libraries listed below
Sorting:
- ☆114Updated 5 months ago
- "Improving Mathematical Reasoning with Process Supervision" by OPENAI☆110Updated 2 weeks ago
- [ACL'24] Code and data of paper "When is Tree Search Useful for LLM Planning? It Depends on the Discriminator"☆54Updated last year
- RL Scaling and Test-Time Scaling (ICML'25)☆108Updated 5 months ago
- augmented LLM with self reflection☆129Updated last year
- ☆54Updated 2 weeks ago
- Code for Paper: Autonomous Evaluation and Refinement of Digital Agents [COLM 2024]☆138Updated 7 months ago
- B-STAR: Monitoring and Balancing Exploration and Exploitation in Self-Taught Reasoners☆82Updated last month
- [ICML 2025] Teaching Language Models to Critique via Reinforcement Learning☆102Updated 2 months ago
- Code release for "Debating with More Persuasive LLMs Leads to More Truthful Answers"☆112Updated last year
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆74Updated 3 months ago
- Revisiting Mid-training in the Era of Reinforcement Learning Scaling☆137Updated this week
- ☆117Updated 4 months ago
- Interpretable Contrastive Monte Carlo Tree Search Reasoning☆49Updated 8 months ago
- official implementation of paper "Process Reward Model with Q-value Rankings"☆60Updated 5 months ago
- Implementation of the Quiet-STAR paper (https://arxiv.org/pdf/2403.09629.pdf)☆54Updated 11 months ago
- ☆19Updated 4 months ago
- ☆66Updated last year
- Code for paper "Optima: Optimizing Effectiveness and Efficiency for LLM-Based Multi-Agent System"☆59Updated 7 months ago
- Process Reward Models That Think☆45Updated last week
- Resources for our paper: "EvoAgent: Towards Automatic Multi-Agent Generation via Evolutionary Algorithms"☆113Updated 8 months ago
- [ACL 2025] Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆95Updated last month
- Critique-out-Loud Reward Models☆67Updated 8 months ago
- ☆98Updated last year
- Replicating O1 inference-time scaling laws☆89Updated 7 months ago
- ☆83Updated 2 months ago
- ☆47Updated 5 months ago
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆60Updated 7 months ago
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate" [COLM 2025]☆163Updated this week
- MPO: Boosting LLM Agents with Meta Plan Optimization☆62Updated 4 months ago