Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Knowledge Fusion of Large Language Models

About

While training large language models (LLMs) from scratch can generate models with distinct functionalities and strengths, it comes at significant costs and may result in redundant capabilities. Alternatively, a cost-effective and compelling approach is to merge existing pre-trained LLMs into a more potent model. However, due to the varying architectures of these LLMs, directly blending their weights is impractical. In this paper, we introduce the notion of knowledge fusion for LLMs, aimed at combining the capabilities of existing LLMs and transferring them into a single LLM. By leveraging the generative distributions of source LLMs, we externalize their collective knowledge and unique strengths, thereby potentially elevating the capabilities of the target model beyond those of any individual source LLM. We validate our approach using three popular LLMs with different architectures--Llama-2, MPT, and OpenLLaMA--across various benchmarks and tasks. Our findings confirm that the fusion of LLMs can improve the performance of the target model across a range of capabilities such as reasoning, commonsense, and code generation. Our code, model weights, and data are public at \url{https://github.com/fanqiwan/FuseLLM}.

Fanqi Wan, Xinting Huang, Deng Cai, Xiaojun Quan, Wei Bi, Shuming Shi• 2024

Related benchmarks

TaskDatasetResultRank
Science Question AnsweringARC Challenge
Accuracy14.58
342
Mathematical ReasoningAIME
AIME Accuracy0.00e+0
288
Code GenerationHumanEval
Pass@118.4
171
Instruction FollowingUnNI
Rouge-L17.25
160
Science Question AnsweringARC Easy
Accuracy21.52
155
KnowledgeMMLU
Accuracy80.62
136
Instruction FollowingS-NI
Rouge-L14.98
119
Safety EvaluationHarmBench
Harmbench Score13
112
ReasoningGSM8K--
106
Code GenerationLiveCodeBench
Pass@10.76
86
Showing 10 of 55 rows

Other info

Follow for update