Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MAmmoTH2: Scaling Instructions from the Web

About

Instruction tuning improves the reasoning abilities of large language models (LLMs), with data quality and scalability being the crucial factors. Most instruction tuning data come from human crowd-sourcing or GPT-4 distillation. We propose a paradigm to efficiently harvest 10 million naturally existing instruction data from the pre-training web corpus to enhance LLM reasoning. Our approach involves (1) recalling relevant documents, (2) extracting instruction-response pairs, and (3) refining the extracted pairs using open-source LLMs. Fine-tuning base LLMs on this dataset, we build MAmmoTH2 models, which significantly boost performance on reasoning benchmarks. Notably, MAmmoTH2-7B's (Mistral) performance increases from 11% to 36.7% on MATH and from 36% to 68.4% on GSM8K without training on any in-domain data. Further training MAmmoTH2 on public instruction tuning datasets yields MAmmoTH2-Plus, achieving state-of-the-art performance on several reasoning and chatbot benchmarks. Our work demonstrates how to harvest large-scale, high-quality instruction data without costly human annotation or GPT-4 distillation, providing a new paradigm for building better instruction tuning data.

Xiang Yue, Tuney Zheng, Ge Zhang, Wenhu Chen• 2024

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Accuracy86.4
1362
Code GenerationHumanEval
Pass@117.68
1036
Mathematical ReasoningMATH
Accuracy47
882
Multi-task Language UnderstandingMMLU
Accuracy68.3
876
Language UnderstandingMMLU
Accuracy64.89
825
Mathematical ReasoningMATH
Accuracy34.1
535
Instruction FollowingAlpacaEval 2.0
Win Rate33.8
507
Mathematical ReasoningMATH (test)
Overall Accuracy47
433
Mathematical ReasoningSVAMP
Accuracy90.3
403
Mathematical ReasoningMATH
Accuracy36.7
338
Showing 10 of 35 rows

Other info

Code

Follow for update