Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Instruction Mining: Instruction Data Selection for Tuning Large Language Models

About

Large language models (LLMs) are initially pretrained for broad capabilities and then finetuned with instruction-following datasets to improve their performance in interacting with humans. Despite advances in finetuning, a standardized guideline for selecting high-quality datasets to optimize this process remains elusive. In this paper, we first propose InstructMining, an innovative method designed for automatically selecting premium instruction-following data for finetuning LLMs. Specifically, InstructMining utilizes natural language indicators as a measure of data quality, applying them to evaluate unseen datasets. During experimentation, we discover that double descent phenomenon exists in large language model finetuning. Based on this observation, we further leverage BlendSearch to help find the best subset among the entire dataset (i.e., 2,532 out of 100,000). Experiment results show that InstructMining-7B achieves state-of-the-art performance on two of the most popular benchmarks: LLM-as-a-judge and Huggingface OpenLLM leaderboard.

Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun• 2023

Related benchmarks

TaskDatasetResultRank
Language ModelingWikiText2
Perplexity12.02
2839
Language ModelingWikiText-2 (test)
PPL17.83
1949
Commonsense ReasoningHellaSwag
Accuracy75.51
1891
Language ModelingWikiText-2
Perplexity (PPL)13.67
1624
Commonsense ReasoningWinoGrande
Accuracy68.56
1085
Language ModelingPTB
Perplexity21.51
1034
Commonsense ReasoningPIQA
Accuracy77.58
751
Language ModelingPTB (test)
Perplexity28.87
526
Question AnsweringARC-E
Accuracy66.96
416
Question AnsweringBoolQ
Accuracy68.31
317
Showing 10 of 29 rows

Other info

Follow for update