Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Compositional Generalization for Primitive Substitutions

About

Compositional generalization is a basic mechanism in human language learning, but current neural networks lack such ability. In this paper, we conduct fundamental research for encoding compositionality in neural networks. Conventional methods use a single representation for the input sentence, making it hard to apply prior knowledge of compositionality. In contrast, our approach leverages such knowledge with two representations, one generating attention maps, and the other mapping attended input words to output symbols. We reduce the entropy in each representation to improve generalization. Our experiments demonstrate significant improvements over the conventional methods in five NLP tasks including instruction learning and machine translation. In the SCAN domain, it boosts accuracies from 14.0% to 98.8% in Jump task, and from 92.0% to 99.7% in TurnLeft task. It also beats human performance on a few-shot learning task. We hope the proposed approach can help ease future research towards human-level compositional language learning.

Yuanpeng Li, Liang Zhao, Jianyu Wang, Joel Hestness• 2019

Related benchmarks

TaskDatasetResultRank
Semantic ParsingCFQ (MCD1)
Accuracy4.81
33
Semantic ParsingCFQ (MCD2)
Accuracy1.04
33
Semantic ParsingCFQ MCD3
Accuracy1.82
33
Semantic ParsingSCAN around right
Exact-match Accuracy83.2
16
Semantic ParsingSCAN (MCD1)
Exact-match Accuracy0.012
12
Semantic ParsingSCAN (MCD2)
Exact Match Accuracy1.7
12
Semantic ParsingSCAN MCD3
Exact Match Accuracy0.6
12
Semantic ParsingSCAN jump
Exact-match Accuracy98.8
11
Semantic ParsingSCAN 2x augmented (ADDJUMP)
Accuracy98.8
5
Showing 9 of 9 rows

Other info

Follow for update