Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Instruction Tuning with GPT-4

About

Prior work has shown that finetuning large language models (LLMs) using machine-generated instruction-following data enables such models to achieve remarkable zero-shot capabilities on new tasks, and no human-written instructions are needed. In this paper, we present the first attempt to use GPT-4 to generate instruction-following data for LLM finetuning. Our early experiments on instruction-tuned LLaMA models show that the 52K English and Chinese instruction-following data generated by GPT-4 leads to superior zero-shot performance on new tasks to the instruction-following data generated by previous state-of-the-art models. We also collect feedback and comparison data from GPT-4 to enable a comprehensive evaluation and reward model training. We make our data generated using GPT-4 as well as our codebase publicly available.

Baolin Peng, Chunyuan Li, Pengcheng He, Michel Galley, Jianfeng Gao• 2023

Related benchmarks

TaskDatasetResultRank
Dialogue Alignment EvaluationAlignBench
Reasoning5.64
90
Multi-turn Dialogue EvaluationMT-Bench-zh
Score5.44
90
Instruction TuningIT Evaluation Suite MMLU, BBH, GSM, TydiQA, CodeX, AE
MMLU55.7
18
Instruction FollowingAlpacaEval v1 (test)
AlpacaEval Score61.8
14
Natural Language UnderstandingOpen LLM Leaderboard (test)
ARC56.57
13
Instruction Tuning EvaluationOpen Instruct Evaluation Suite (test)
MMLU59.6
12
Instruction Dataset Quality EvaluationPublic Instruction Tuning Datasets (test)
Discrimination Index9.8
6
Showing 7 of 7 rows

Other info

Follow for update