AI & ML interests

Preference Alignment, Superalignment

⚙️ Preference Agents

Preference Agents is an organization focused on developing and releasing small language models ("preference agents") that enable efficient personalization of larger language models (LLMs). Our agents learn user preferences and generate natural language rules that guide LLMs to produce tailored content, without requiring extensive fine-tuning of the larger models.

🎯 Our Approach

We train small, locally deployable language models to act as "steering wheels" for larger, pre-trained LLMs. These agents learn user preferences from small, personalized datasets and encode these preferences into concise natural language rules. These rules are then provided as context to the larger LLM, guiding its output towards the desired personalized style and content.

📦 Resources

Datasets

We release three datasets for research on personalized language modeling:

  • Enron-42k: A curated subset of the Enron email corpus, focused on original content creation. It contains approximately 40,240 emails from 191 unique senders.
  • The New Yorker: A curated subset of the All The News 2.0 corpus, containing 4000 articles from the New Yorker.
  • LAMP 3U Subset: A subset of the LAMP 3U Amazon product reviews dataset, containing 22,500 reviews from 15 users.

Both datasets are licensed under the Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) license.

Models

We provide fine-tuned Llama-3-8B-Instruct preference agent models for the Enron-42k and LAMP 3U datasets. These agents can be used to generate personalized emails or product reviews based on user preferences.

Note: Due to licensing restrictions on the New Yorker dataset, we are unable to release models trained on that data.

🚀 How to Use Our Resources

Datasets

from datasets import load_dataset

enron_dataset = load_dataset("preference-agents/Enron-42k", split="train")
lamp_dataset = load_dataset("preference-agents/LAMP-3U-Subset", split="train")

📊 Evaluation

Our approach was evaluated using GPT-4o and human evaluations, demonstrating significant improvements over baselines like zero-shot generation, few-shot learning, and naive fine-tuning.

📜 Citation

If you use our resources in your research or applications, please cite our paper:

pending

🙏 Acknowledgements

We thank Meta AI for providing the Llama-3 models, Google AI for access to Gemini 1.5 Pro, and Anthropic for access to Claude 3.5 Sonnet. We also acknowledge the creators of the Enron email corpus and the LAMP 3U dataset for making their valuable resources available to the research community.

models

None public yet