Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Deeper Text Understanding for IR with Contextual Neural Language Modeling

About

Neural networks provide new possibilities to automatically learn complex language patterns and query-document relations. Neural IR models have achieved promising results in learning query-document relevance patterns, but few explorations have been done on understanding the text content of a query or a document. This paper studies leveraging a recently-proposed contextual neural language model, BERT, to provide deeper text understanding for IR. Experimental results demonstrate that the contextual text representations from BERT are more effective than traditional word embeddings. Compared to bag-of-words retrieval models, the contextual language model can better leverage language structures, bringing large improvements on queries written in natural languages. Combining the text understanding ability with search knowledge leads to an enhanced pre-trained BERT model that can benefit related search tasks where training data are limited.

Zhuyun Dai, Jamie Callan• 2019

Related benchmarks

TaskDatasetResultRank
Document RankingTREC DL Track 2019 (test)
nDCG@1065.94
133
Passage retrievalMsMARCO (dev)
MRR@1024.3
116
RetrievalMS MARCO (dev)
MRR@100.243
84
Document RankingTREC DL Track 2020 (test)
nDCG@100.616
63
Passage RankingTREC DL 2019 (test)
NDCG@1055.1
33
Document RerankingRobust04 Description
MAP0.3464
13
Document RerankingRobust04 Title
MAP31.83
12
Document RerankingGOV2 Title
MAP31.93
12
Document RerankingGOV2 Description
MAP28.57
12
Document RerankingTREC-DL 2021
NDCG@1067.56
11
Showing 10 of 16 rows

Other info

Code

Follow for update