Paragraph-based Transformer Pre-training for Multi-Sentence Inference
About
Inference tasks such as answer sentence selection (AS2) or fact verification are typically solved by fine-tuning transformer-based models as individual sentence-pair classifiers. Recent studies show that these tasks benefit from modeling dependencies across multiple candidate sentences jointly. In this paper, we first show that popular pre-trained transformers perform poorly when used for fine-tuning on multi-candidate inference tasks. We then propose a new pre-training objective that models the paragraph-level semantics across multiple input sentences. Our evaluation on three AS2 and one fact verification datasets demonstrates the superiority of our pre-training technique over the traditional ones for transformers used as joint models for multi-candidate inference tasks, as well as when used as cross-encoders for sentence-pair formulations of these tasks. Our code and pre-trained models are released at https://github.com/amazon-research/wqa-multi-sentence-inference .
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Answer Selection | WikiQA (test) | MAP0.885 | 149 | |
| Answer Sentence Selection | TREC-QA (test) | MAP91.1 | 63 | |
| Fact Verification | FEVER (dev) | Label Accuracy81.21 | 57 | |
| Answer Sentence Selection | ASNQ (test) | P@163 | 45 | |
| Answer Sentence Selection | WikiQA | P@182.7 | 36 | |
| Fact Verification | FEVER (test) | LA Score74.39 | 32 | |
| Answer Sentence Selection | TREC-QA | P@192.2 | 24 | |
| Answer Sentence Selection | ASNQ | P@164.3 | 24 |