Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Modeling Hierarchical Structures with Continuous Recursive Neural Networks

About

Recursive Neural Networks (RvNNs), which compose sequences according to their underlying hierarchical syntactic structure, have performed well in several natural language processing tasks compared to similar models without structural biases. However, traditional RvNNs are incapable of inducing the latent structure in a plain text sequence on their own. Several extensions have been proposed to overcome this limitation. Nevertheless, these extensions tend to rely on surrogate gradients or reinforcement learning at the cost of higher bias or variance. In this work, we propose Continuous Recursive Neural Network (CRvNN) as a backpropagation-friendly alternative to address the aforementioned limitations. This is done by incorporating a continuous relaxation to the induced structure. We demonstrate that CRvNN achieves strong performance in challenging synthetic tasks such as logical inference and ListOps. We also show that CRvNN performs comparably or better than prior latent structure models on real-world tasks such as sentiment analysis and natural language inference.

Jishnu Ray Chowdhury, Cornelia Caragea• 2021

Related benchmarks

TaskDatasetResultRank
Natural Language InferenceSNLI
Accuracy85.3
174
Natural Language InferenceMNLI (matched)
Accuracy72.2
110
Natural Language InferenceMNLI (mismatched)
Accuracy72.6
68
Natural Language InferenceSNLI hard 1.0 (test)
Accuracy70.6
27
Sentiment ClassificationSST2 phrase
Accuracy88.3
16
Paraphrase DetectionPAWS QQP
Accuracy34.8
16
Paraphrase DetectionQQP IID
Accuracy84.8
8
Sentiment ClassificationIMDB (Contrast)
Accuracy77.8
8
Sentiment ClassificationIMDB Counterfactual
Accuracy85.38
8
Natural Language InferenceSNLI Counterfactual
Accuracy59.8
8
Showing 10 of 18 rows

Other info

Follow for update