Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ANALOGYKB: Unlocking Analogical Reasoning of Language Models with A Million-scale Knowledge Base

About

Analogical reasoning is a fundamental cognitive ability of humans. However, current language models (LMs) still struggle to achieve human-like performance in analogical reasoning tasks due to a lack of resources for model training. In this work, we address this gap by proposing ANALOGYKB, a million-scale analogy knowledge base (KB) derived from existing knowledge graphs (KGs). ANALOGYKB identifies two types of analogies from the KGs: 1) analogies of the same relations, which can be directly extracted from the KGs, and 2) analogies of analogous relations, which are identified with a selection and filtering pipeline enabled by large language models (LLMs), followed by minor human efforts for data quality control. Evaluations on a series of datasets of two analogical reasoning tasks (analogy recognition and generation) demonstrate that ANALOGYKB successfully enables both smaller LMs and LLMs to gain better analogical reasoning capabilities.

Siyu Yuan, Jiangjie Chen, Changzhi Sun, Jiaqing Liang, Yanghua Xiao, Deqing Yang• 2023

Related benchmarks

TaskDatasetResultRank
Analogy GenerationE-KAR
Accuracy75
26
Analogy GenerationUNIT 4
Accuracy92
22
Analogy GenerationSCAN (out-of-domain)
Accuracy15.3
15
Analogy recognitionBATS
Accuracy92.42
15
Analogy recognitionUNIT 2
Accuracy88.32
15
Analogy recognitionUNIT 4
Accuracy76.15
15
Analogy recognitionSAT
Accuracy60.78
15
Analogy recognitionGoogle
Accuracy98.8
15
Analogy GenerationSAT
Accuracy0.91
11
Analogy GenerationBATS
Accuracy96
11
Showing 10 of 14 rows

Other info

Code

Follow for update