PKU-Alignment / safe-rlhfLinks

Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
1,556Updated 2 months ago

Alternatives and similar repositories for safe-rlhf

Users that are interested in safe-rlhf are comparing it to the libraries listed below

Sorting: