Learning to Retrieve Reasoning Paths over Wikipedia Graph for Question Answering
About
Answering questions that require multi-hop reasoning at web-scale necessitates retrieving multiple evidence documents, one of which often has little lexical or semantic relationship to the question. This paper introduces a new graph-based recurrent retrieval approach that learns to retrieve reasoning paths over the Wikipedia graph to answer multi-hop open-domain questions. Our retriever model trains a recurrent neural network that learns to sequentially retrieve evidence paragraphs in the reasoning path by conditioning on the previously retrieved documents. Our reader model ranks the reasoning paths and extracts the answer span included in the best reasoning path. Experimental results show state-of-the-art results in three open-domain QA datasets, showcasing the effectiveness and robustness of our method. Notably, our method achieves significant improvement in HotpotQA, outperforming the previous best model by more than 14 points.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multi-hop Question Answering | HotpotQA fullwiki setting (test) | Answer F173 | 64 | |
| End-to-end Open-Domain Question Answering | NQ (test) | Exact Match (EM)32.6 | 50 | |
| Answer extraction and supporting sentence prediction | HotpotQA fullwiki (test) | Answer EM60.04 | 48 | |
| Open-domain Question Answering | Natural Questions (NQ) | Exact Match (EM)32.6 | 46 | |
| Question Answering | HotpotQA distractor (dev) | Answer F181.2 | 45 | |
| Question Answering | HotpotQA (dev) | Answer F173.3 | 43 | |
| Multi-hop Question Answering | HotpotQA (dev) | Answer F165.8 | 43 | |
| Open-domain Question Answering | SQUAD Open (test) | Exact Match56.5 | 39 | |
| Multi-hop Question Answering | HotpotQA fullwiki setting (dev) | Answer F173.3 | 38 | |
| Question Answering | HotpotQA (test) | Ans F173 | 37 |