PKU-Alignment / safe-rlhf

Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
1,394Updated 7 months ago

Alternatives and similar repositories for safe-rlhf:

Users that are interested in safe-rlhf are comparing it to the libraries listed below