Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SLiC-HF: Sequence Likelihood Calibration with Human Feedback

About

Learning from human feedback has been shown to be effective at aligning language models with human preferences. Past work has often relied on Reinforcement Learning from Human Feedback (RLHF), which optimizes the language model using reward scores assigned from a reward model trained on human preference data. In this work we show how the recently introduced Sequence Likelihood Calibration (SLiC), can also be used to effectively learn from human preferences (SLiC-HF). Furthermore, we demonstrate this can be done with human feedback data collected for a different model, similar to off-policy, offline RL data. Automatic and human evaluation experiments on the TL;DR summarization task show that SLiC-HF significantly improves supervised fine-tuning baselines. Furthermore, SLiC-HF presents a competitive alternative to the PPO RLHF implementation used in past work while being much simpler to implement, easier to tune and more computationally efficient in practice.

Yao Zhao, Rishabh Joshi, Tianqi Liu, Misha Khalman, Mohammad Saleh, Peter J. Liu• 2023

Related benchmarks

TaskDatasetResultRank
LLM Alignment EvaluationAlpacaEval 2
LC Win Rate36.7
72
Language Model Alignment EvaluationArena Hard v0.1
Win Rate (%)25.1
36
Dialogue GenerationAnthropic HH (test)
Average Preference Score61.62
16
SummarizationReddit TL;DR (test)
Preference vs SFT (%)68.61
8
Reasoning and Language UnderstandingOpen LLM Leaderboard MMLU-PRO, IFEval, BBH, GPQA, MATH, GSM8K, ARC v0.4.0 (test)
MMLU-PRO26.52
7
Showing 5 of 5 rows

Other info

Follow for update