Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Compressing Context to Enhance Inference Efficiency of Large Language Models

About

Large language models (LLMs) achieved remarkable performance across various tasks. However, they face challenges in managing long documents and extended conversations, due to significantly increased computational requirements, both in memory and inference time, and potential context truncation when the input exceeds the LLM's fixed context length. This paper proposes a method called Selective Context that enhances the inference efficiency of LLMs by identifying and pruning redundancy in the input context to make the input more compact. We test our approach using common data sources requiring long context processing: arXiv papers, news articles, and long conversations, on tasks of summarisation, question answering, and response generation. Experimental results show that Selective Context significantly reduces memory cost and decreases generation latency while maintaining comparable performance compared to that achieved when full context is used. Specifically, we achieve a 50\% reduction in context cost, resulting in a 36\% reduction in inference memory usage and a 32\% reduction in inference time, while observing only a minor drop of .023 in BERTscore and .038 in faithfulness on four downstream applications, indicating that our method strikes a good balance between efficiency and performance.

Yucheng Li, Bo Dong, Chenghua Lin, Frank Guerin• 2023

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Accuracy61.33
983
ReasoningBBH
Accuracy50.07
507
Multi-hop Question AnsweringHotpotQA (test)
F154.9
198
Long-context UnderstandingLongBench
Overall Average Score32.16
115
Mathematical ReasoningGSM8K
EM2.5
115
Question AnsweringSQuAD (test)
F159.7
111
Long-context UnderstandingLongBench (test)
Avg Score20.2
80
Question Answering2WikiMQA
F18.2
44
ReasoningBBH (test)--
40
Question AnsweringSQuAD
F1 Score51.39
36
Showing 10 of 37 rows

Other info

Follow for update