ListT5: Listwise Reranking with Fusion-in-Decoder Improves Zero-shot Retrieval
About
We propose ListT5, a novel reranking approach based on Fusion-in-Decoder (FiD) that handles multiple candidate passages at both train and inference time. We also introduce an efficient inference framework for listwise ranking based on m-ary tournament sort with output caching. We evaluate and compare our model on the BEIR benchmark for zero-shot retrieval task, demonstrating that ListT5 (1) outperforms the state-of-the-art RankT5 baseline with a notable +1.3 gain in the average NDCG@10 score, (2) has an efficiency comparable to pointwise ranking models and surpasses the efficiency of previous listwise ranking models, and (3) overcomes the lost-in-the-middle problem of previous listwise rerankers. Our code, model checkpoints, and the evaluation framework are fully open-sourced at \url{https://github.com/soyoung97/ListT5}.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Ranking | BEIR selected subset v1.0.0 (test) | TREC-COVID84.7 | 38 | |
| Information Retrieval | Scientific QA Base setting | HitRate@11.2 | 38 | |
| Question Answering | Scientific QA Base setting | F1 Score24.21 | 38 | |
| Reranking | SciRAG-SSLI easy 1.0 (test) | Hit Rate @ 10.2 | 19 | |
| Reranking | SciRAG-SSLI hard 1.0 (test) | Hit Rate @ 10.6 | 19 | |
| Scientific Question Answering | SciRAG-SSLI easy 1.0 (test) | F1 Score19.64 | 19 | |
| Scientific Question Answering | SciRAG-SSLI hard 1.0 (test) | F1 Score21.34 | 19 | |
| Information Retrieval | BEIR BM25 Top-100 initial retrieval | TREC-COVID Score84.7 | 7 | |
| Information Retrieval | BEIR BM25 Top-1000 initial retrieval | TREC-COVID82.1 | 3 |