Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

RankT5: Fine-Tuning T5 for Text Ranking with Ranking Losses

About

Recently, substantial progress has been made in text ranking based on pretrained language models such as BERT. However, there are limited studies on how to leverage more powerful sequence-to-sequence models such as T5. Existing attempts usually formulate text ranking as classification and rely on postprocessing to obtain a ranked list. In this paper, we propose RankT5 and study two T5-based ranking model structures, an encoder-decoder and an encoder-only one, so that they not only can directly output ranking scores for each query-document pair, but also can be fine-tuned with "pairwise" or "listwise" ranking losses to optimize ranking performances. Our experiments show that the proposed models with ranking losses can achieve substantial ranking performance gains on different public text ranking data sets. Moreover, when fine-tuned with listwise ranking losses, the ranking model appears to have better zero-shot ranking performance on out-of-domain data sets compared to the model fine-tuned with classification losses.

Honglei Zhuang, Zhen Qin, Rolf Jagerman, Kai Hui, Ji Ma, Jing Lu, Jianmo Ni, Xuanhui Wang, Michael Bendersky• 2022

Related benchmarks

TaskDatasetResultRank
Information RetrievalScientific QA Base setting
HitRate@146.9
38
Question AnsweringScientific QA Base setting
F1 Score40.61
38
RerankingBEIR
NQ NDCG@50.5097
35
RerankingTREC
NDCG@5 (DL19)71.66
35
RerankingSciRAG-SSLI easy 1.0 (test)
Hit Rate @ 155.4
19
Scientific Question AnsweringSciRAG-SSLI hard 1.0 (test)
F1 Score45.48
19
RerankingSciRAG-SSLI hard 1.0 (test)
Hit Rate @ 152.6
19
Scientific Question AnsweringSciRAG-SSLI easy 1.0 (test)
F1 Score44.81
19
Text RankingBEIR out-of-domain
Arguana Score33
9
Information RetrievalBEIR BM25 Top-100 initial retrieval
TREC-COVID Score81.7
7
Showing 10 of 12 rows

Other info

Follow for update