sail-sg / feedback-conditional-policyLinks
Code for "Language Models Can Learn from Verbal Feedback Without Scalar Rewards"
☆33Updated this week
Alternatives and similar repositories for feedback-conditional-policy
Users that are interested in feedback-conditional-policy are comparing it to the libraries listed below
Sorting:
- A Recipe for Building LLM Reasoners to Solve Complex Instructions☆24Updated 2 months ago
- Reinforcing General Reasoning without Verifiers☆86Updated 3 months ago
- The official repository for SkyLadder: Better and Faster Pretraining via Context Window Scheduling☆34Updated last month
- ☆18Updated 2 months ago
- ☆62Updated 3 months ago
- Optimizing Anytime Reasoning via Budget Relative Policy Optimization☆47Updated 2 months ago
- ☆48Updated 7 months ago
- Emergent Hierarchical Reasoning in LLMs/VLMs through Reinforcement Learning☆30Updated 3 weeks ago
- JudgeLRM: Large Reasoning Models as a Judge☆39Updated 2 weeks ago
- ☆34Updated last month
- Code for "Reasoning to Learn from Latent Thoughts"☆119Updated 6 months ago
- Official implementation of Bootstrapping Language Models via DPO Implicit Rewards☆44Updated 5 months ago
- ☆43Updated this week
- Process Reward Models That Think☆53Updated 2 months ago
- ☆18Updated 5 months ago
- RAG-RewardBench: Benchmarking Reward Models in Retrieval Augmented Generation for Preference Alignment☆16Updated 9 months ago
- Source code for our paper: "ARIA: Training Language Agents with Intention-Driven Reward Aggregation".☆22Updated last month
- [EMNLP 2025] LightThinker: Thinking Step-by-Step Compression☆104Updated 5 months ago
- [ACL'25 Oral] What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective☆74Updated 3 months ago
- [ICLR 2025] LongPO: Long Context Self-Evolution of Large Language Models through Short-to-Long Preference Optimization☆40Updated 7 months ago
- Exploration of automated dataset selection approaches at large scales.☆47Updated 7 months ago
- [ICML 2025] Teaching Language Models to Critique via Reinforcement Learning☆111Updated 4 months ago
- [ICLR 2025] When Attention Sink Emerges in Language Models: An Empirical View (Spotlight)☆125Updated 2 months ago
- The official code repository for the paper "Mirage or Method? How Model–Task Alignment Induces Divergent RL Conclusions".☆15Updated last month
- [ACL 2025] Are Your LLMs Capable of Stable Reasoning?☆30Updated last month
- [EMNLP 2025 Industry] Repo for "Z1: Efficient Test-time Scaling with Code"☆64Updated 5 months ago
- [NeurIPS-2024] 📈 Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623☆86Updated last year
- ☆18Updated 9 months ago
- B-STAR: Monitoring and Balancing Exploration and Exploitation in Self-Taught Reasoners☆85Updated 4 months ago
- Unofficial Implementation of Chain-of-Thought Reasoning Without Prompting☆33Updated last year