Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Revisiting Iterative Back-Translation from the Perspective of Compositional Generalization

About

Human intelligence exhibits compositional generalization (i.e., the capacity to understand and produce unseen combinations of seen components), but current neural seq2seq models lack such ability. In this paper, we revisit iterative back-translation, a simple yet effective semi-supervised method, to investigate whether and how it can improve compositional generalization. In this work: (1) We first empirically show that iterative back-translation substantially improves the performance on compositional generalization benchmarks (CFQ and SCAN). (2) To understand why iterative back-translation is useful, we carefully examine the performance gains and find that iterative back-translation can increasingly correct errors in pseudo-parallel data. (3) To further encourage this mechanism, we propose curriculum iterative back-translation, which better improves the quality of pseudo-parallel data, thus further improving the performance.

Yinuo Guo, Hualei Zhu, Zeqi Lin, Bei Chen, Jian-Guang Lou, Dongmei Zhang• 2020

Related benchmarks

TaskDatasetResultRank
Semantic ParsingSCAN around right
Exact-match Accuracy37.8
16
Semantic ParsingCFQ MCD3 (test)
Accuracy64.6
15
Semantic ParsingCFQ MCD2 (test)
Accuracy0.578
15
Semantic ParsingCFQ MCD1 (test)
Accuracy64.8
15
Semantic ParsingSCAN (MCD2)
Exact Match Accuracy80.8
12
Semantic ParsingSCAN (MCD1)
Exact-match Accuracy0.643
12
Semantic ParsingSCAN MCD3
Exact Match Accuracy52.2
12
Semantic ParsingSCAN jump
Exact-match Accuracy99.6
11
Showing 8 of 8 rows

Other info

Follow for update