Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

On Tree-Based Neural Sentence Modeling

About

Neural networks with tree-based sentence encoders have shown better results on many downstream tasks. Most of existing tree-based encoders adopt syntactic parsing trees as the explicit structure prior. To study the effectiveness of different tree structures, we replace the parsing trees with trivial trees (i.e., binary balanced tree, left-branching tree and right-branching tree) in the encoders. Though trivial trees contain no syntactic information, those encoders get competitive or even better results on all of the ten downstream tasks we investigated. This surprising result indicates that explicit syntax guidance may not be the main contributor to the superior performances of tree-based neural sentence modeling. Further analysis show that tree modeling gives better results when crucial words are closer to the final representation. Additional experiments give more clues on how to design an effective tree-based encoder. Our code is open-source and available at https://github.com/ExplorerFreda/TreeEnc.

Haoyue Shi, Hao Zhou, Jiaze Chen, Lei Li• 2018

Related benchmarks

TaskDatasetResultRank
Natural Language InferenceSNLI
Accuracy84.4
174
Long-range sequence modelingLong Range Arena (LRA) (test)--
158
Natural Language InferenceMNLI (matched)
Accuracy71.1
110
Natural Language InferenceMNLI (mismatched)
Accuracy71.4
68
Natural Language InferenceSNLI hard 1.0 (test)
Accuracy69.2
27
Long-range sequence modelingLRA 92 (test)
ListOps Accuracy55.15
26
Paraphrase DetectionPAWS QQP
Accuracy42.9
16
Logical Expression EvaluationListOps-O Argument Generalization (Arguments 15)
Accuracy44.5
11
Logical Expression EvaluationListOps-O Length Generalization (Lengths 500-600)
Accuracy40.4
11
Logical Expression EvaluationListOps-O Argument Generalization (Arguments 10)
Accuracy0.4535
11
Showing 10 of 32 rows

Other info

Code

Follow for update