REARANK: Reasoning Re-ranking Agent via Reinforcement Learning
About
We present REARANK, a large language model (LLM)-based listwise reasoning reranking agent. REARANK explicitly reasons before reranking, significantly improving both performance and interpretability. Leveraging reinforcement learning and data augmentation, REARANK achieves substantial improvements over baseline models across popular information retrieval benchmarks, notably requiring only 179 annotated samples. Built on top of Qwen2.5-7B, our REARANK-7B demonstrates performance comparable to GPT-4 on both in-domain and out-of-domain benchmarks and even surpasses GPT-4 on reasoning-intensive BRIGHT benchmarks. These results underscore the effectiveness of our approach and highlight how reinforcement learning can enhance LLM reasoning capabilities in reranking.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | HotpotQA | F143.2 | 114 | |
| Question Answering | MuSiQue | EM5.6 | 84 | |
| Multi-hop Question Answering | HotpotQA | F142 | 79 | |
| Question Answering | 2Wiki | F129.4 | 75 | |
| Multi-hop Question Answering | 2Wiki | F1 Score30.1 | 41 | |
| Information Retrieval | BRIGHT 1.0 (test) | nDCG@10 (Avg)24.6 | 18 | |
| Rich-media Re-ranking | Rich-Media Search All Query 1.0 (test) | N@488.5 | 10 | |
| Rich-media Re-ranking | Rich-Media Search Complex Query 1.0 (test) | N@489.1 | 10 | |
| Rich-media Re-ranking | Rich-Media Search Broad-needs Query 1.0 (test) | N@485.1 | 10 | |
| Rich-media Re-ranking | Rich-Media Search Simple Query 1.0 (test) | N@493.1 | 10 |