Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Paragraph-based Transformer Pre-training for Multi-Sentence Inference

About

Inference tasks such as answer sentence selection (AS2) or fact verification are typically solved by fine-tuning transformer-based models as individual sentence-pair classifiers. Recent studies show that these tasks benefit from modeling dependencies across multiple candidate sentences jointly. In this paper, we first show that popular pre-trained transformers perform poorly when used for fine-tuning on multi-candidate inference tasks. We then propose a new pre-training objective that models the paragraph-level semantics across multiple input sentences. Our evaluation on three AS2 and one fact verification datasets demonstrates the superiority of our pre-training technique over the traditional ones for transformers used as joint models for multi-candidate inference tasks, as well as when used as cross-encoders for sentence-pair formulations of these tasks. Our code and pre-trained models are released at https://github.com/amazon-research/wqa-multi-sentence-inference .

Luca Di Liello, Siddhant Garg, Luca Soldaini, Alessandro Moschitti• 2022

Related benchmarks

TaskDatasetResultRank
Answer SelectionWikiQA (test)
MAP0.885
149
Answer Sentence SelectionTREC-QA (test)
MAP91.1
63
Fact VerificationFEVER (dev)
Label Accuracy81.21
57
Answer Sentence SelectionASNQ (test)
P@163
45
Answer Sentence SelectionWikiQA
P@182.7
36
Fact VerificationFEVER (test)
LA Score74.39
32
Answer Sentence SelectionTREC-QA
P@192.2
24
Answer Sentence SelectionASNQ
P@164.3
24
Showing 8 of 8 rows

Other info

Code

Follow for update