Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Using contradictions improves question answering systems

About

This work examines the use of contradiction in natural language inference (NLI) for question answering (QA). Typically, NLI systems help answer questions by determining if a potential answer is \emph{entailed} (supported) by some background context. But is it useful to also determine if an answer contradicts the context? We test this in two settings, multiple choice and extractive QA, and find that systems that incorporate contradiction can do slightly better than entailment-only systems on certain datasets. However, the best performances come from using contradiction, entailment, and QA model confidence scores together. This has implications for the deployment of QA systems in domains such as medicine and science where safety is an issue.

\'Etienne Fortier-Dubois, Domenic Rosati• 2022

Related benchmarks

TaskDatasetResultRank
Multiple-choice Question AnsweringSciQ
Accuracy98.21
74
Multiple-choice Question AnsweringRACE
Accuracy92.68
46
Commonsense Question AnsweringCosmosQA
Accuracy91.12
36
Multiple-choice Question AnsweringMCTest
Accuracy100
22
Multiple-choice Question AnsweringDREAM
Accuracy98.77
22
Multiple-choice Question AnsweringQASC
Accuracy98.6
22
Question AnsweringRACE-C
Accuracy69.8
19
Extractive Question AnsweringSQuAD 20% coverage
F198.76
14
Multiple-choice Question AnsweringMCScript
Accuracy99.82
11
Extractive Question AnsweringBioASQ 20% coverage
F1 Score85.06
7
Showing 10 of 23 rows

Other info

Follow for update