PKU-Alignment / safe-rlhfLinks

Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
1,491Updated last year

Alternatives and similar repositories for safe-rlhf

Users that are interested in safe-rlhf are comparing it to the libraries listed below

Sorting: