Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Conversational Question Reformulation via Sequence-to-Sequence Architectures and Pretrained Language Models

About

This paper presents an empirical study of conversational question reformulation (CQR) with sequence-to-sequence architectures and pretrained language models (PLMs). We leverage PLMs to address the strong token-to-token independence assumption made in the common objective, maximum likelihood estimation, for the CQR task. In CQR benchmarks of task-oriented dialogue systems, we evaluate fine-tuned PLMs on the recently-introduced CANARD dataset as an in-domain task and validate the models using data from the TREC 2019 CAsT Track as an out-domain task. Examining a variety of architectures with different numbers of parameters, we demonstrate that the recent text-to-text transfer transformer (T5) achieves the best results both on CANARD and CAsT with fewer parameters, compared to similar transformer architectures.

Sheng-Chieh Lin, Jheng-Hong Yang, Rodrigo Nogueira, Ming-Feng Tsai, Chuan-Ju Wang, Jimmy Lin• 2020

Related benchmarks

TaskDatasetResultRank
Conversational RetrievalQReCC (test)
Recall@1053.8
43
Conversational Search RetrievalTopiOCQA (test)
MRR23.4
21
Dense RetrievalCAsT 19
MRR70.1
7
Dense RetrievalCAsT 20
MRR42.3
7
Showing 4 of 4 rows

Other info

Follow for update