Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

RaFe: Ranking Feedback Improves Query Rewriting for RAG

About

As Large Language Models (LLMs) and Retrieval Augmentation Generation (RAG) techniques have evolved, query rewriting has been widely incorporated into the RAG system for downstream tasks like open-domain QA. Many works have attempted to utilize small models with reinforcement learning rather than costly LLMs to improve query rewriting. However, current methods require annotations (e.g., labeled relevant documents or downstream answers) or predesigned rewards for feedback, which lack generalization, and fail to utilize signals tailored for query rewriting. In this paper, we propose ours, a framework for training query rewriting models free of annotations. By leveraging a publicly available reranker, ours~provides feedback aligned well with the rewriting objectives. Experimental results demonstrate that ours~can obtain better performance than baselines.

Shengyu Mao, Yong Jiang, Boli Chen, Xiao Li, Peng Wang, Xinyu Wang, Pengjun Xie, Fei Huang, Huajun Chen, Ningyu Zhang• 2024

Related benchmarks

TaskDatasetResultRank
Multi-hop Question AnsweringHotpotQA
F1 Score55.4
294
Multi-hop Question Answering2Wiki
Exact Match36.2
152
Medical Question AnsweringMedical QA Evaluation Suite (MedQA, MedMCQA, MMLU-Med, PubMedQA, BioASQ, SEER, DDXPlus, MIMIC-IV)
MedQA Score70.86
27
Multi-hop Question AnsweringMuSiQue
Exact Match (EM)11.8
24
Showing 4 of 4 rows

Other info

Follow for update