Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

First Train to Generate, then Generate to Train: UnitedSynT5 for Few-Shot NLI

About

Natural Language Inference (NLI) tasks require identifying the relationship between sentence pairs, typically classified as entailment, contradiction, or neutrality. While the current state-of-the-art (SOTA) model, Entailment Few-Shot Learning (EFL), achieves a 93.1% accuracy on the Stanford Natural Language Inference (SNLI) dataset, further advancements are constrained by the dataset's limitations. To address this, we propose a novel approach leveraging synthetic data augmentation to enhance dataset diversity and complexity. We present UnitedSynT5, an advanced extension of EFL that leverages a T5-based generator to synthesize additional premise-hypothesis pairs, which are rigorously cleaned and integrated into the training data. These augmented examples are processed within the EFL framework, embedding labels directly into hypotheses for consistency. We train a GTR-T5-XL model on this expanded dataset, achieving a new benchmark of 94.7% accuracy on the SNLI dataset, 94.0% accuracy on the E-SNLI dataset, and 92.6% accuracy on the MultiNLI dataset, surpassing the previous SOTA models. This research demonstrates the potential of synthetic data augmentation in improving NLI models, offering a path forward for further advancements in natural language understanding tasks.

Sourav Banerjee, Anush Mahajan, Ayushi Agarwal, Eishkaran Singh• 2024

Related benchmarks

TaskDatasetResultRank
Natural Language InferenceSNLI (test)
Accuracy94.7
681
Natural Language InferenceSNLI
Accuracy94.7
174
Natural Language InferenceMultiNLI matched/mismatched
Accuracy92.6
10
Natural Language Inferencee-SNLI (test)
Accuracy94
9
Showing 4 of 4 rows

Other info

Follow for update