Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Passage Re-ranking with BERT

About

Recently, neural models pretrained on a language modeling task, such as ELMo (Peters et al., 2017), OpenAI GPT (Radford et al., 2018), and BERT (Devlin et al., 2018), have achieved impressive results on various natural language processing tasks such as question-answering and natural language inference. In this paper, we describe a simple re-implementation of BERT for query-based passage re-ranking. Our system is the state of the art on the TREC-CAR dataset and the top entry in the leaderboard of the MS MARCO passage retrieval task, outperforming the previous state of the art by 27% (relative) in MRR@10. The code to reproduce our results is available at https://github.com/nyu-dl/dl4marco-bert

Rodrigo Nogueira, Kyunghyun Cho• 2019

Related benchmarks

TaskDatasetResultRank
RetrievalMS MARCO (dev)
MRR@100.365
84
Passage RankingMS MARCO (dev)
MRR@1036.5
73
Information RetrievalRobust04
P@2040.42
72
Passage RankingNQ
MRR47.31
29
Passage RankingTREC DL 2020
R@10100
28
Passage retrievalNatural Questions (NQ)
Top-10 Accuracy65.18
28
Passage RankingWebQuestions (WQ)
R@1062.94
28
Passage RankingTREC DL 2019
R@1096.66
28
Text RetrievalMS MARCO Document
MRR@1045.2
26
Text RetrievalMS MARCO Passage
MRR@100.416
26
Showing 10 of 30 rows

Other info

Code

Follow for update