junkangwu / beta-DPOView on GitHub
[NeurIPS 2024] Official code of $\beta$-DPO: Direct Preference Optimization with Dynamic $\beta$
50Oct 23, 2024Updated last year

Alternatives and similar repositories for beta-DPO

Users that are interested in beta-DPO are comparing it to the libraries listed below

Sorting:

Are these results useful?