Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up
alignment-handbook 's Collections
Handbook v0.1 models and datasets
DPO vs KTO vs IPO
Constitutional AI

DPO vs KTO vs IPO

updated Jan 16, 2024

A collection of datasets and models used for the Aligning LLMs with Direct Preference Optimization Methods blogpost

Upvote
12

  • HuggingFaceH4/orca_dpo_pairs

    Viewer • Updated Apr 14, 2024 • 12.9k • 419 • 29

  • HuggingFaceH4/ultrafeedback_binarized

    Viewer • Updated Oct 16, 2024 • 187k • 13k • 293
Upvote
12
  • Collection guide
  • Browse collections
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs