Self-supervised Knowledge Triplet Learning for Zero-shot Question Answering
About
The aim of all Question Answering (QA) systems is to be able to generalize to unseen questions. Current supervised methods are reliant on expensive data annotation. Moreover, such annotations can introduce unintended annotator bias which makes systems focus more on the bias than the actual task. In this work, we propose Knowledge Triplet Learning (KTL), a self-supervised task over knowledge graphs. We propose heuristics to create synthetic graphs for commonsense and scientific knowledge. We propose methods of how to use KTL to perform zero-shot QA and our experiments show considerable improvements over large pre-trained transformer models.
Pratyay Banerjee, Chitta Baral• 2020
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Physical Interaction Question Answering | PIQA | Accuracy48.5 | 323 | |
| Science Question Answering | ARC-E | Accuracy33.4 | 138 | |
| Science Question Answering | ARC-C | Accuracy28.4 | 127 | |
| Abductive Natural Language Inference | aNLI (leaderboard) | Accuracy65.3 | 47 | |
| Question Answering | QASC | Score26.6 | 36 | |
| Commonsense Question Answering | SocialIQA (SIQA) (val) | Accuracy48.5 | 24 | |
| Commonsense Question Answering | CommonsenseQA (CSQA) (val) | Accuracy38.8 | 23 | |
| Commonsense Question Answering | Abductive NLI (aNLI) (val) | Accuracy0.653 | 21 |
Showing 8 of 8 rows