Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Group Preference Optimization: Few-Shot Alignment of Large Language Models

About

Many applications of large language models (LLMs), ranging from chatbots to creative writing, require nuanced subjective judgments that can differ significantly across different groups. Existing alignment algorithms can be expensive to align for each group, requiring prohibitive amounts of group-specific preference data and computation for real-world use cases. We introduce Group Preference Optimization (GPO), an alignment framework that steers language models to preferences of individual groups in a few-shot manner. In GPO, we augment the base LLM with an independent transformer module trained to predict the preferences of a group for the LLM generations. For few-shot learning, we parameterize this module as an in-context autoregressive transformer and train it via meta-learning on several groups. We empirically validate the efficacy of GPO through rigorous evaluations using LLMs with varied sizes on three human opinion adaptation tasks. These tasks involve adapting to the preferences of US demographic groups, global countries, and individual users. Our results demonstrate that GPO not only aligns models more accurately but also requires fewer group-specific preferences, and less training and inference computing resources, outperforming existing strategies such as in-context steering and fine-tuning methods.

Siyan Zhao, John Dang, Aditya Grover• 2023

Related benchmarks

TaskDatasetResultRank
Preference PredictionPRISM (test)
Accuracy56.48
51
Human preference predictionChatbot Arena latest (test)
Accuracy58.1
51
Personalized Reward ModelingPRISM Personalized
Accuracy59.16
44
Personalized Reward ModelingChatbot Arena Personalized
Accuracy58.5
42
Personalized Reward ModelingBESPOKE-Meta OOD
Binary Preference Accuracy52.04
18
Personalized Reward ModelingReddit TLDR 150 examples Seen
User-level Accuracy68.5
11
Personalized Reward ModelingReddit TLDR 150 examples Overall
User-level Accuracy68.6
11
Personalized Reward ModelingReddit TLDR 100 examples Unseen
User-level Accuracy68
11
Personalized Reward ModelingReddit TLDR 150 examples Unseen
User-level Accuracy68.6
11
Personalized Reward ModelingPRISM Unseen
User-level Accuracy0.642
11
Showing 10 of 14 rows

Other info

Follow for update