Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

ReasonRank: Empowering Passage Ranking with Strong Reasoning Ability

About

Large Language Model (LLM) based listwise ranking has shown superior performance in many passage ranking tasks. With the development of Large Reasoning Models, many studies have demonstrated that step-by-step reasoning during test-time helps improve listwise ranking performance. However, due to the scarcity of reasoning-intensive training data, existing rerankers perform poorly in many complex ranking scenarios and the ranking ability of reasoning-intensive rerankers remains largely underdeveloped. In this paper, we first propose an automated reasoning-intensive training data synthesis framework, which sources training queries and passages from diverse domains and applies DeepSeek-R1 to generate high-quality training labels. A self-consistency data filtering mechanism is designed to ensure the data quality. To empower the listwise reranker with strong reasoning ability, we further propose a two-stage post-training approach, which includes a cold-start supervised fine-tuning (SFT) stage for reasoning pattern learning and a reinforcement learning (RL) stage for further ranking ability enhancement. During the RL stage, based on the nature of listwise ranking, we design a multi-view ranking reward, which is more effective than a ranking metric-based reward. Extensive experiments demonstrate that our trained reasoning-intensive reranker \textbf{ReasonRank} outperforms existing baselines significantly and also achieves much lower latency than pointwise reranker Rank1. \textbf{Through further experiments, our ReasonRank has achieved state-of-the-art (SOTA) performance 40.6 on the BRIGHT leaderboard\footnote{https://brightbenchmark.github.io/}.} Our codes are available at https://github.com/8421BCD/ReasonRank.

Wenhan Liu, Xinyu Ma, Weiwei Sun, Yutao Zhu, Yuchen Li, Dawei Yin, Zhicheng Dou• 2025

Related benchmarks

TaskDatasetResultRank
Information RetrievalBRIGHT 1.0 (test)
nDCG@10 (Avg)26.5
35
RecommendationHRT
ND@50.8046
26
Long-context Memory Retrieval and ReasoningLongMemEval 128K
F1 Score44.26
20
Long-context Memory Retrieval and ReasoningLoCoMo 32K
F1 Score41.04
20
Long-context Memory Retrieval and ReasoningLongMemEval 1M
F1 Score48.2
20
Long-context Memory Retrieval and ReasoningWebWalker 128K
F1 Score25.72
20
Long-context Memory Retrieval and ReasoningHotpotQA 128K
F1 Score22.37
20
Long-context Memory Retrieval and ReasoningPersonaMem 32K
F1 Score23.78
20
Long-context Memory Retrieval and ReasoningPerM 128K V2
F1 Score22.27
20
Long-context Memory Retrieval and ReasoningZH4O 128K
F1 Score47.16
20
Showing 10 of 35 rows

Other info

Follow for update