GitHub topics: preference-alignment
Shentao-YANG/Dense_Reward_T2I
Source code for "A Dense Reward View on Aligning Text-to-Image Diffusion with Preference" (ICML'24).
Language: Python - Size: 11.3 MB - Last synced at: 3 days ago - Pushed at: 12 months ago - Stars: 38 - Forks: 0

Meaquadddd/DPO-Shift
DPO-Shift: Shifting the Distribution of Direct Preference Optimization
Language: Python - Size: 216 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 4 - Forks: 1

Video-Bench/Video-Bench
Video Generation Benchmark
Language: Python - Size: 8.92 MB - Last synced at: about 1 month ago - Pushed at: 3 months ago - Stars: 7 - Forks: 2

zjukg/KnowPAT
[Paper][ACL 2024 Findings] Knowledgeable Preference Alignment for LLMs in Domain-specific Question Answering
Language: Python - Size: 9.03 MB - Last synced at: 5 months ago - Pushed at: 11 months ago - Stars: 187 - Forks: 17

junkangwu/beta-DPO
$\beta$-DPO: Direct Preference Optimization with Dynamic $\beta$
Language: Python - Size: 22.5 KB - Last synced at: 10 months ago - Pushed at: 10 months ago - Stars: 0 - Forks: 0

junkangwu/Dr_DPO
Towards Robust Alignment of Language Models: Distributionally Robustifying Direct Preference Optimization
Language: Python - Size: 24.4 KB - Last synced at: 10 months ago - Pushed at: 11 months ago - Stars: 0 - Forks: 0

BARUDA-AI/Awesome-Preference-Optimization
Survey of preference alignment algorithms
Size: 0 Bytes - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0
