Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
nbeerbower
's Collections
DPO
Qwen3
Dumplings
abliteration loras
bruphin
flammen
llama 3 experiments
Nemo
DPO
updated
Jun 9
Various useful datasets with preference optimization
Upvote
5
jondurbin/gutenberg-dpo-v0.1
Viewer
•
Updated
Jan 12, 2024
•
918
•
411
•
151
nbeerbower/gutenberg2-dpo
Viewer
•
Updated
Nov 16, 2024
•
293
•
37
•
20
jondurbin/truthy-dpo-v0.1
Viewer
•
Updated
Jan 11, 2024
•
1.02k
•
202
•
134
kyujinpy/orca_math_dpo
Viewer
•
Updated
Apr 12, 2024
•
15.3k
•
73
•
19
antiven0m/physical-reasoning-dpo
Viewer
•
Updated
Feb 12
•
899
•
24
•
10
flammenai/MahouMix-v1
Viewer
•
Updated
May 30, 2024
•
267
•
8
•
4
flammenai/Date-DPO-NoAsterisks
Viewer
•
Updated
Sep 18, 2024
•
330
•
10
•
4
nbeerbower/Arkhaios-DPO
Viewer
•
Updated
Nov 12, 2024
•
222
•
8
•
8
nbeerbower/Purpura-DPO
Viewer
•
Updated
Nov 12, 2024
•
230
•
14
•
8
nbeerbower/Schule-DPO
Viewer
•
Updated
Nov 16, 2024
•
34
•
10
•
1
HumanLLMs/Human-Like-DPO-Dataset
Viewer
•
Updated
Jan 12
•
10.9k
•
949
•
229
nbeerbower/gutenberg-moderne-dpo
Viewer
•
Updated
Nov 17, 2024
•
346
•
19
•
3
nbeerbower/reddit-dpo
Viewer
•
Updated
Feb 1
•
76.9k
•
9
•
1
Atsunori/HelpSteer2-DPO
Viewer
•
Updated
Jul 11, 2024
•
7.59k
•
59
•
8
abacusai/MetaMath_DPO_FewShot
Viewer
•
Updated
Feb 26, 2024
•
395k
•
117
•
26
nbeerbower/GreatFirewall-DPO
Viewer
•
Updated
Mar 2
•
492
•
44
•
10
nbeerbower/synthetic-fiction-dpo
Viewer
•
Updated
May 31
•
550
•
12
•
1
nbeerbower/human-writing-dpo
Updated
Jun 8
•
23
•
7
Upvote
5
+1
Share collection
View history
Collection guide
Browse collections