Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Tulu 3: Pushing Frontiers in Open Language Model Post-Training

About

Language model post-training is applied to refine behaviors and unlock new skills across a wide range of recent language models, but open recipes for applying these techniques lag behind proprietary ones. The underlying training data and recipes for post-training are simultaneously the most important pieces of the puzzle and the portion with the least transparency. To bridge this gap, we introduce Tulu 3, a family of fully-open state-of-the-art post-trained models, alongside its data, code, and training recipes, serving as a comprehensive guide for modern post-training techniques. Tulu 3, which builds on Llama 3.1 base models, achieves results surpassing the instruct versions of Llama 3.1, Qwen 2.5, Mistral, and even closed models such as GPT-4o-mini and Claude 3.5-Haiku. The training algorithms for our models include supervised finetuning (SFT), Direct Preference Optimization (DPO), and a novel method we call Reinforcement Learning with Verifiable Rewards (RLVR). With Tulu 3, we introduce a multi-task evaluation scheme for post-training recipes with development and unseen evaluations, standard benchmark implementations, and substantial decontamination of existing open datasets on said benchmarks. We conclude with analysis and discussion of training methods that did not reliably improve performance. In addition to the Tulu 3 model weights and demo, we release the complete recipe -- including datasets for diverse core skills, a robust toolkit for data curation and evaluation, the training code and infrastructure, and, most importantly, a detailed report for reproducing and further adapting the Tulu 3 approach to more domains.

Nathan Lambert, Jacob Morrison, Valentina Pyatkin, Shengyi Huang, Hamish Ivison, Faeze Brahman, Lester James V. Miranda, Alisa Liu, Nouha Dziri, Shane Lyu, Yuling Gu, Saumya Malik, Victoria Graf, Jena D. Hwang, Jiangjiang Yang, Ronan Le Bras, Oyvind Tafjord, Chris Wilhelm, Luca Soldaini, Noah A. Smith, Yizhong Wang, Pradeep Dasigi, Hannaneh Hajishirzi• 2024

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningWinoGrande--
1085
Code GenerationHumanEval
Pass@17.87e+3
1036
ReasoningBBH--
672
Instruction FollowingIFEval--
625
Instruction FollowingAlpacaEval 2.0
Win Rate13.46
507
Code GenerationHumanEval+
Pass@125.61
383
Mathematical ReasoningMATH 500
pass@169
239
Instruction FollowingMT-Bench
MT-Bench Score7.45
215
KnowledgeMMLU
Accuracy61.11
136
Mathematical ReasoningGSM8K
EM89.2
123
Showing 10 of 129 rows
...

Other info

Follow for update