junkangwu / beta-DPO

[NeurIPS 2024] Official code of $\beta$-DPO: Direct Preference Optimization with Dynamic $\beta$
39Updated 3 months ago

Alternatives and similar repositories for beta-DPO:

Users that are interested in beta-DPO are comparing it to the libraries listed below