sail-sg / oatLinks
🌾 OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.
☆456Updated 2 weeks ago
Alternatives and similar repositories for oat
Users that are interested in oat are comparing it to the libraries listed below
Sorting:
- Reproducible, flexible LLM evaluations☆242Updated 2 months ago
- ☆315Updated 3 months ago
- RewardBench: the first evaluation tool for reward models.☆630Updated 3 months ago
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"☆257Updated 3 months ago
- Code for the paper: "Learning to Reason without External Rewards"