DPO datasets
updated
Viewer
• Updated • 7.5k • 1.53k
• 171
argilla/distilabel-capybara-dpo-7k-binarized
Viewer
• Updated • 7.56k • 2.22k
• 182
llamafactory/DPO-En-Zh-20k
Viewer
• Updated • 20k • 324
• 99
argilla/distilabel-intel-orca-dpo-pairs
Viewer
• Updated • 12.9k • 5.18k
• 183
argilla/ultrafeedback-binarized-preferences-cleaned
Viewer
• Updated • 60.9k • 8.29k
• 162
argilla/distilabel-math-preference-dpo
Viewer
• Updated • 2.42k • 352
• 88
M4-ai/prm_dpo_pairs_cleaned
Viewer
• Updated • 7.99k • 38
• 11
jondurbin/truthy-dpo-v0.1
Viewer
• Updated • 1.02k • 346
• 136
YeungNLP/ultrafeedback_binarized
Viewer
• Updated • 63.1k • 42
• 1
shibing624/DPO-En-Zh-20k-Preference
Viewer
• Updated • 20k • 228
• 18
Preview
• Updated • 29
• 6
mlabonne/orpo-dpo-mix-40k
Viewer
• Updated • 44.2k • 659
• 301
Viewer
• Updated • 15.3k • 26
• 19
jondurbin/gutenberg-dpo-v0.1
Viewer
• Updated • 918 • 463
• 161
CyberNative/Code_Vulnerability_Security_DPO
Viewer
• Updated • 4.66k • 1.08k
• 154
mlabonne/orpo-dpo-mix-40k-flat
Viewer
• Updated • 44.2k • 19
• 14
selimc/orpo-dpo-mix-TR-20k
Viewer
• Updated • 19.9k • 22
• 7
efederici/alpaca-vs-alpaca-orpo-dpo
Viewer
• Updated • 49.2k • 80
• 7
Viewer
• Updated • 2.42k • 38
• 10
allenai/llama-3.1-tulu-3-8b-preference-mixture
Viewer
• Updated • 273k • 1.98k
• 26
allenai/llama-3.1-tulu-3-70b-preference-mixture
Viewer
• Updated • 337k • 80
• 19
HuggingFaceH4/ultrafeedback_binarized
Viewer
• Updated • 187k • 11.5k
• 329
Preview
• Updated • 1.27k
• 211
allenai/llama-3.1-tulu-3-405b-preference-mixture
Viewer
• Updated • 361k • 56
• 6
Viewer
• Updated • 450k • 14.2k
• 729
qihoo360/Light-R1-DPOData
Viewer
• Updated • 2.97k • 114
• 29