Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

OpenAssistant Conversations -- Democratizing Large Language Model Alignment

About

Aligning large language models (LLMs) with human preferences has proven to drastically improve usability and has driven rapid adoption as demonstrated by ChatGPT. Alignment techniques such as supervised fine-tuning (SFT) and reinforcement learning from human feedback (RLHF) greatly reduce the required skill and domain knowledge to effectively harness the capabilities of LLMs, increasing their accessibility and utility across various domains. However, state-of-the-art alignment techniques like RLHF rely on high-quality human feedback data, which is expensive to create and often remains proprietary. In an effort to democratize research on large-scale alignment, we release OpenAssistant Conversations, a human-generated, human-annotated assistant-style conversation corpus consisting of 161,443 messages in 35 different languages, annotated with 461,292 quality ratings, resulting in over 10,000 complete and fully annotated conversation trees. The corpus is a product of a worldwide crowd-sourcing effort involving over 13,500 volunteers. Models trained on OpenAssistant Conversations show consistent improvements on standard benchmarks over respective base models. We release our code and data under a fully permissive licence.

Andreas K\"opf, Yannic Kilcher, Dimitri von R\"utte, Sotiris Anagnostidis, Zhi-Rui Tam, Keith Stevens, Abdullah Barhoum, Nguyen Minh Duc, Oliver Stanley, Rich\'ard Nagyfi, Shahul ES, Sameer Suri, David Glushkov, Arnav Dantuluri, Andrew Maguire, Christoph Schuhmann, Huu Nguyen, Alexander Mattick• 2023

Related benchmarks

TaskDatasetResultRank
MathGSM8K
Accuracy0.245
206
KnowledgeGPQA
Accuracy27.2
51
General Language IntelligenceMMLU, GSM8K, BBH, TriviaQA, NQ latest available (test)
MMLU73.34
26
MathGSM-PLUS
Score16.5
22
Instruction FollowingVicuna benchmark zero-shot
Pairwise Score (ChatGPT vs Sys)91.2
21
Reward ModelingAnthropic Helpful
Accuracy67.7
12
Reward ModelingAnthropic Harmless
Accuracy68.4
12
CodingHEval
Accuracy44.5
12
CodingHEval+
Accuracy39
12
Reward ModelingMeta Helpfulness (test)
Accuracy53.8
5
Showing 10 of 13 rows

Other info

Follow for update