PKU-Alignment / safe-rlhf

Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
1,462Updated 11 months ago

Alternatives and similar repositories for safe-rlhf

Users that are interested in safe-rlhf are comparing it to the libraries listed below

Sorting: