andersonbcdefg / dpo-lora

direct preference optimization with only 1 model copy :)
12Updated 11 months ago

Related projects: